Update paper link to Hugging Face Papers (#3)
Browse files- Update paper link to Hugging Face Papers (f6d731898b0cc9b42503f46f94c9b407134cc56d)
Co-authored-by: Niels Rogge <[email protected]>
README.md
CHANGED
|
@@ -23,7 +23,7 @@ Foundation-Sec-8B-Instruct enables organizations to build AI-driven security too
|
|
| 23 |
|
| 24 |
- **Model Name:** Llama-3.1-FoundationAI-SecurityLLM-8B-Instruct (Foundation-Sec-8B-Instruct)
|
| 25 |
- **Model Developer:** Amin Karbasi and Research team at Foundation AI — Cisco
|
| 26 |
-
- **Technical Report:** [
|
| 27 |
- **Model Card Contact:** For questions about the team, model usage, and future directions, contact [`[email protected]`](mailto:[email protected]). For technical questions about the model, please contact [`[email protected]`](mailto:[email protected]) and [`[email protected]`](mailto:[email protected]).
|
| 28 |
- **Model Release Date:** August 1st, 2025
|
| 29 |
- **Supported Language(s):** English
|
|
@@ -32,7 +32,6 @@ Foundation-Sec-8B-Instruct enables organizations to build AI-driven security too
|
|
| 32 |
- **Training Data Status:** This is a static model trained on an offline dataset. Future versions of the tuned models will be released on updated data.
|
| 33 |
- **License:** See NOTICE.md
|
| 34 |
|
| 35 |
-
|
| 36 |
|
| 37 |
## Intended Use
|
| 38 |
|
|
@@ -75,22 +74,22 @@ For questions or assistance with fine-tuning Foundation-Sec-8B-Instruct, please
|
|
| 75 |
|
| 76 |
The following uses are out-of-scope and are neither recommended nor intended use cases:
|
| 77 |
|
| 78 |
-
1.
|
| 79 |
-
-
|
| 80 |
-
-
|
| 81 |
-
-
|
| 82 |
-
-
|
| 83 |
-
2.
|
| 84 |
-
-
|
| 85 |
-
-
|
| 86 |
-
-
|
| 87 |
-
-
|
| 88 |
-
3.
|
| 89 |
-
-
|
| 90 |
-
-
|
| 91 |
-
-
|
| 92 |
-
4.
|
| 93 |
-
5.
|
| 94 |
|
| 95 |
## How to Get Started with the Model
|
| 96 |
|
|
@@ -187,49 +186,49 @@ It is recommended to deploy this model with additional safeguards (such as Llama
|
|
| 187 |
|
| 188 |
Foundation-Sec-8B-Instruct has several limitations that users should be aware of:
|
| 189 |
|
| 190 |
-
1.
|
| 191 |
-
-
|
| 192 |
-
-
|
| 193 |
-
2.
|
| 194 |
-
-
|
| 195 |
-
-
|
| 196 |
-
-
|
| 197 |
-
-
|
| 198 |
-
3.
|
| 199 |
-
-
|
| 200 |
-
-
|
| 201 |
-
-
|
| 202 |
-
4.
|
| 203 |
-
-
|
| 204 |
-
5.
|
| 205 |
-
-
|
| 206 |
-
-
|
| 207 |
-
-
|
| 208 |
-
-
|
| 209 |
-
6.
|
| 210 |
-
-
|
| 211 |
|
| 212 |
|
| 213 |
### Recommendations
|
| 214 |
|
| 215 |
To address the limitations of Foundation-Sec-8B-Instruct, we recommend:
|
| 216 |
|
| 217 |
-
1.
|
| 218 |
-
-
|
| 219 |
-
-
|
| 220 |
-
-
|
| 221 |
-
2.
|
| 222 |
-
-
|
| 223 |
-
-
|
| 224 |
-
-
|
| 225 |
-
3.
|
| 226 |
-
-
|
| 227 |
-
-
|
| 228 |
-
-
|
| 229 |
-
4.
|
| 230 |
-
-
|
| 231 |
-
-
|
| 232 |
-
5.
|
| 233 |
-
-
|
| 234 |
-
-
|
| 235 |
-
-
|
|
|
|
| 23 |
|
| 24 |
- **Model Name:** Llama-3.1-FoundationAI-SecurityLLM-8B-Instruct (Foundation-Sec-8B-Instruct)
|
| 25 |
- **Model Developer:** Amin Karbasi and Research team at Foundation AI — Cisco
|
| 26 |
+
- **Technical Report:** [Llama-3.1-FoundationAI-SecurityLLM-8B-Instruct Technical Report](https://huggingface.co/papers/2508.01059)
|
| 27 |
- **Model Card Contact:** For questions about the team, model usage, and future directions, contact [`[email protected]`](mailto:[email protected]). For technical questions about the model, please contact [`[email protected]`](mailto:[email protected]) and [`[email protected]`](mailto:[email protected]).
|
| 28 |
- **Model Release Date:** August 1st, 2025
|
| 29 |
- **Supported Language(s):** English
|
|
|
|
| 32 |
- **Training Data Status:** This is a static model trained on an offline dataset. Future versions of the tuned models will be released on updated data.
|
| 33 |
- **License:** See NOTICE.md
|
| 34 |
|
|
|
|
| 35 |
|
| 36 |
## Intended Use
|
| 37 |
|
|
|
|
| 74 |
|
| 75 |
The following uses are out-of-scope and are neither recommended nor intended use cases:
|
| 76 |
|
| 77 |
+
1. **Generating harmful content** - The model should not be used to:
|
| 78 |
+
- Generate malware or other malicious code
|
| 79 |
+
- Create phishing content or social engineering scripts
|
| 80 |
+
- Develop attack plans targeting specific organizations
|
| 81 |
+
- Design exploitation techniques for vulnerabilities without legitimate security research purposes
|
| 82 |
+
2. **Critical security decisions without human oversight** - The model should not be used for:
|
| 83 |
+
- Autonomous security decision-making without human review
|
| 84 |
+
- Critical infrastructure protection without expert supervision
|
| 85 |
+
- Final determination of security compliance without human verification
|
| 86 |
+
- Autonomous vulnerability remediation without testing
|
| 87 |
+
3. **Legal or medical advice** - The model is not qualified to provide:
|
| 88 |
+
- Legal advice regarding security regulations, compliance requirements, or intellectual property disputes
|
| 89 |
+
- Legal advice regarding security issues that would reference legal statutes, precedents, or case law necessary to provide legal advice
|
| 90 |
+
- Medical advice regarding health impacts of security incidents
|
| 91 |
+
4. **Non-security use cases** - The model is specifically optimized for cybersecurity and may not perform as well on general tasks as models trained for broader applications.
|
| 92 |
+
5. **Violation of Laws or Regulations** - Any use that violates applicable laws or regulations.
|
| 93 |
|
| 94 |
## How to Get Started with the Model
|
| 95 |
|
|
|
|
| 186 |
|
| 187 |
Foundation-Sec-8B-Instruct has several limitations that users should be aware of:
|
| 188 |
|
| 189 |
+
1. **Domain-specific knowledge limitations**:
|
| 190 |
+
- Foundation-Sec-8B-Instruct may not be familiar with recent vulnerabilities, exploits, or novel attack vectors or security technologies released after its training cutoff date
|
| 191 |
+
- Knowledge of specialized or proprietary security systems or tools may be limited
|
| 192 |
+
2. **Potential biases**:
|
| 193 |
+
- The model may reflect biases present in security literature and documentation
|
| 194 |
+
- The model may be trained on known attack patterns and have difficulty recognizing novel attack vectors
|
| 195 |
+
- Security practices and recommendations may be biased toward certain technological ecosystems
|
| 196 |
+
- Geographic and cultural biases in security approaches may be present
|
| 197 |
+
3. **Security risks**:
|
| 198 |
+
- The model cannot verify the identity or intentions of users
|
| 199 |
+
- Adversarial prompting techniques might potentially bypass safety mechanisms
|
| 200 |
+
- The model may unintentionally provide information that could be misused if proper prompting guardrails are not implemented
|
| 201 |
+
4. **Contextual blindness:**
|
| 202 |
+
- The model may struggle to understand the complex interrelationships between systems, users, and data in order to provide accurate context.
|
| 203 |
+
5. **Technical limitations**:
|
| 204 |
+
- Performance varies based on how security concepts are described in prompts
|
| 205 |
+
- May not fully understand complex, multi-step security scenarios without clear explanation
|
| 206 |
+
- Cannot access external systems or actively scan environments
|
| 207 |
+
- Cannot independently verify factual accuracy of its outputs
|
| 208 |
+
6. **Ethical considerations**:
|
| 209 |
+
- Dual-use nature of security knowledge requires careful consideration of appropriate use cases
|
| 210 |
|
| 211 |
|
| 212 |
### Recommendations
|
| 213 |
|
| 214 |
To address the limitations of Foundation-Sec-8B-Instruct, we recommend:
|
| 215 |
|
| 216 |
+
1. **Human oversight**:
|
| 217 |
+
- Always have qualified security professionals review model outputs before implementation
|
| 218 |
+
- Use the model as an assistive tool rather than a replacement for expert human judgment
|
| 219 |
+
- Implement a human-in-the-loop approach for security-critical applications
|
| 220 |
+
2. **System design safeguards**:
|
| 221 |
+
- Implement additional validation layers for applications built with this model
|
| 222 |
+
- Consider architectural constraints that limit the model's ability to perform potentially harmful actions (excessive agency)
|
| 223 |
+
- Deploy the model in environments with appropriate access controls
|
| 224 |
+
3. **Prompt engineering**:
|
| 225 |
+
- Use carefully designed prompts that encourage ethical security practices
|
| 226 |
+
- Include explicit instructions regarding responsible disclosure and ethical hacking principles
|
| 227 |
+
- Structure interactions to minimize the risk of inadvertently harmful outputs
|
| 228 |
+
4. **Knowledge supplementation**:
|
| 229 |
+
- Supplement the model with up-to-date security feeds and databases
|
| 230 |
+
- Implement retrieval-augmented generation for current threat intelligence sources
|
| 231 |
+
5. **Usage policies**:
|
| 232 |
+
- Develop and enforce clear acceptable use policies for applications using this model
|
| 233 |
+
- Implement monitoring and auditing for high-risk applications
|
| 234 |
+
- Create documentation for end users about the model's limitations
|