viettelsecurity-ai/security-llama3.2-3b

Warm
Public
3.2B
BF16
32768
1
Hugging Face
Overview

Model Overview

The viettelsecurity-ai/security-llama3.2-3b is a 3.2 billion parameter, dense decoder-only Transformer model, originally developed by Meta. This model has been specifically fine-tuned by viettelsecurity-ai to excel in cybersecurity contexts.

Key Capabilities & Training

  • Specialized Cybersecurity Knowledge: The model's training data is an extension of security-llama3.2-3b, incorporating a diverse range of cybersecurity sources. This includes publicly available blogs, papers, and references from the PEASEC/cybersecurity_dataset.
  • Synthetic Data for Education: It leverages newly created, synthetic "textbook-like" data, generated using GPT-4o, specifically designed for teaching cybersecurity concepts.
  • Academic Resources: Training also includes acquired academic books and Q&A datasets, further enhancing its domain-specific understanding.
  • Context Length: The model supports a context length of 4K tokens.

Usage Recommendations

  • Chat Format Optimization: Given its training methodology, the model performs best when prompts are structured in a chat format, utilizing specific tokens like <|begin_of_text|>, <|start_header_id|>user<|end_header_id|>, and <|eot_id|>. This makes it ideal for interactive cybersecurity assistance or Q&A systems.
  • Role-Playing: The example usage demonstrates its capability to act as a "SOC-tier3" analyst, indicating its suitability for roles requiring expert-level cybersecurity responses.

This model is a strong candidate for applications requiring a deep, specialized understanding of cybersecurity topics, offering precise and contextually relevant responses within that domain.