Llama-3.1-FoundationAI-SecurityLLM-Reasoning-8B Technical Report

📅 2026-01-28
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the current lack of open-source, natively reasoning-capable large language models tailored for cybersecurity. Building upon Llama-3.1-8B-Base, we propose and release the first open-source large language model with native support for cybersecurity reasoning. The model is trained via a two-stage process—supervised fine-tuning (SFT) followed by reinforcement learning with verifiable rewards (RLVR)—integrating datasets spanning cybersecurity analysis, instruction following, and mathematical reasoning. It further incorporates system prompts and safety guardrails to enhance reliability. Evaluated across 10 cybersecurity-specific and 10 general-purpose benchmarks, the model demonstrates superior performance in multi-hop reasoning and security-related tasks, significantly outperforming comparable models and rivaling much larger counterparts while maintaining a strong balance between domain expertise and general capabilities.

Technology Category

Application Category

📝 Abstract
We present Foundation-Sec-8B-Reasoning, the first open-source native reasoning model for cybersecurity. Built upon our previously released Foundation-Sec-8B base model (derived from Llama-3.1-8B-Base), the model is trained through a two-stage process combining supervised fine-tuning (SFT) and reinforcement learning from verifiable rewards (RLVR). Our training leverages proprietary reasoning data spanning cybersecurity analysis, instruction-following, and mathematical reasoning. Evaluation across 10 cybersecurity benchmarks and 10 general-purpose benchmarks demonstrates performance competitive with significantly larger models on cybersecurity tasks while maintaining strong general capabilities. The model shows effective generalization on multi-hop reasoning tasks and strong safety performance when deployed with appropriate system prompts and guardrails. This work demonstrates that domain-specialized reasoning models can achieve strong performance on specialized tasks while maintaining broad general capabilities. We release the model publicly at https://huggingface.co/fdtn-ai/Foundation-Sec-8B-Reasoning.
Problem

Research questions and friction points this paper is trying to address.

cybersecurity
reasoning
foundation model
domain specialization
generalization
Innovation

Methods, ideas, or system contributions that make the work stand out.

reasoning model
cybersecurity
reinforcement learning from verifiable rewards
supervised fine-tuning
domain-specialized LLM
🔎 Similar Papers
No similar papers found.
Zhuoran Yang
Zhuoran Yang
Yale University
machine learningoptimizationreinforcement learningstatistics
Ed Li
Ed Li
Yale University
agentic systemsai4scienceautoML
J
Jianliang He
Foundation AI–Cisco Systems Inc.; Yale University
Aman Priyanshu
Aman Priyanshu
Cisco
AI for SecurityAI SecurityAI SafetyPrivacyCybersecurity
B
Baturay Sağlam
Foundation AI–Cisco Systems Inc.; Yale University
Paul Kassianik
Paul Kassianik
AI Security Researcher
AI security
S
Sajana Weerawardhena
Foundation AI–Cisco Systems Inc.
A
Anu Vellore
Foundation AI–Cisco Systems Inc.
B
Blaine Nelson
Foundation AI–Cisco Systems Inc.
Neusha Javidnia
Neusha Javidnia
PhD Student, UC San Diego
Large Language ModelsRobust AIComputer Architecture
A
Arthur Goldblatt
Foundation AI–Cisco Systems Inc.
F
Fraser Burch
Foundation AI–Cisco Systems Inc.
A
Avi Zohary
Foundation AI–Cisco Systems Inc.
Assaf Eisenman
Assaf Eisenman
Stanford University
M
Mahdi Sabbaghi
Foundation AI–Cisco Systems Inc.; University of Pennsylvania
S
Supriti Vijay
Foundation AI–Cisco Systems Inc.; Carnegie Mellon University
R
Rahim Dharssi
Foundation AI–Cisco Systems Inc.
D
Dhruv Kedia
Foundation AI–Cisco Systems Inc.
K
Kojin Oshiba
Foundation AI–Cisco Systems Inc.
Yaron Singer
Yaron Singer
Harvard University
AlgorithmsMachine LearningGame TheoryNetworks
Amin Karbasi
Amin Karbasi
Cisco Foundation AI, ex Robust Intelligence Chief Scientist, ex Yale professor, ex Googler yale.edu
AI