Llama-3.1-FoundationAI-SecurityLLM-Base-8B Technical Report

📅 2025-04-28
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the performance limitations of general-purpose large language models (LLMs) in cybersecurity—stemming from scarce domain-specific training data and inadequate representation of security knowledge—this paper introduces Foundation-Sec-8B: the first 8-billion-parameter specialized security LLM built upon the Llama 3.1 architecture, explicitly designed for red-teaming, blue-teaming, vulnerability analysis, and alignment with the MITRE ATT&CK framework. We construct a high-quality, multi-granularity cybersecurity corpus and perform domain-adaptive continual pretraining followed by knowledge-enhanced fine-tuning. Evaluated on CyberSecEval2, SecBench, and a novel ATT&CK reasoning benchmark, Foundation-Sec-8B significantly outperforms same-scale open models and achieves competitive performance on key tasks relative to 70B Llama 3.1 and GPT-4o-mini. The model weights and training corpus are fully open-sourced to accelerate AI-driven security adoption in governmental and enterprise settings.

Technology Category

Application Category

📝 Abstract
As transformer-based large language models (LLMs) increasingly permeate society, they have revolutionized domains such as software engineering, creative writing, and digital arts. However, their adoption in cybersecurity remains limited due to challenges like scarcity of specialized training data and complexity of representing cybersecurity-specific knowledge. To address these gaps, we present Foundation-Sec-8B, a cybersecurity-focused LLM built on the Llama 3.1 architecture and enhanced through continued pretraining on a carefully curated cybersecurity corpus. We evaluate Foundation-Sec-8B across both established and new cybersecurity benchmarks, showing that it matches Llama 3.1-70B and GPT-4o-mini in certain cybersecurity-specific tasks. By releasing our model to the public, we aim to accelerate progress and adoption of AI-driven tools in both public and private cybersecurity contexts.
Problem

Research questions and friction points this paper is trying to address.

Addressing limited LLM adoption in cybersecurity due to data scarcity
Enhancing cybersecurity knowledge representation in large language models
Improving AI-driven tool performance in cybersecurity-specific tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Enhanced Llama 3.1 architecture for cybersecurity
Continued pretraining on curated cybersecurity corpus
Matches larger models in specific security tasks
🔎 Similar Papers
No similar papers found.
Paul Kassianik
Paul Kassianik
AI Security Researcher
AI security
B
Baturay Sağlam
Yale University
A
Alexander Chen
Foundation AI – Cisco Systems Inc.
B
Blaine Nelson
Foundation AI – Cisco Systems Inc.
A
Anu Vellore
Foundation AI – Cisco Systems Inc.
M
Massimo Aufiero
Foundation AI – Cisco Systems Inc.
F
Fraser Burch
Foundation AI – Cisco Systems Inc.
D
Dhruv Kedia
Foundation AI – Cisco Systems Inc.
A
Avi Zohary
Foundation AI – Cisco Systems Inc.
S
Sajana Weerawardhena
Foundation AI – Cisco Systems Inc.
Aman Priyanshu
Aman Priyanshu
Cisco
AI for SecurityAI SecurityAI SafetyPrivacyCybersecurity
A
Adam Swanda
Foundation AI – Cisco Systems Inc.
A
Amy Chang
Foundation AI – Cisco Systems Inc.
Hyrum Anderson
Hyrum Anderson
Robust Intelligence
machine learningsecuritysignal processing
K
Kojin Oshiba
Foundation AI – Cisco Systems Inc.
O
Omar Santos
Security & Trust Organization – Cisco Systems Inc.
Yaron Singer
Yaron Singer
Harvard University
AlgorithmsMachine LearningGame TheoryNetworks
Amin Karbasi
Amin Karbasi
Cisco Foundation AI, ex Robust Intelligence Chief Scientist, ex Yale professor, ex Googler yale.edu
AI