Vigilnz/sena-1-vega

TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kPublished:Feb 18, 2026Architecture:Transformer Cold

Vigilnz/sena-1-vega is a policy-driven safety classification model, based on a Transformer architecture and fine-tuned from OpenAI's GPT-4.1-nano. It is designed to detect and categorize unsafe or policy-violating prompts in AI systems, outputting structured JSON classifications. This model acts as a security gateway for LLM applications, identifying prompt injection, misuse, and adversarial prompts across categories like cybercrime, data privacy, and extremism. Its primary use is to enforce safety policies by filtering harmful inputs before they reach downstream language models.

Loading preview...

Sena-1-Vega: AI Safety Classification Model

Sena-1-Vega, developed by Vigilnz, is a specialized policy classification model built on a Transformer architecture and fine-tuned from OpenAI's GPT-4.1-nano. Its core function is to act as a security gateway for LLM-powered applications, identifying and categorizing prompts that violate defined safety policies.

Key Capabilities

  • Policy Enforcement: Detects and categorizes unsafe or policy-violating prompts across a wide range of categories, including cybercrime, social engineering, data privacy, extremism, and harmful behaviors.
  • Structured Output: Provides structured JSON classifications (e.g., safe: boolean, category, subcategory, confidence, reason) for automated enforcement pipelines.
  • Prompt Injection Detection: Specifically designed to identify attempts to override model instructions or extract system prompts.
  • Comprehensive Safety Categories: Covers areas like cybercrime and security, fraud, harmful behavior, illegal activities, privacy violations, and extremism.

Intended Use Cases

  • AI Application Safety: Filtering harmful prompts before they reach generative models.
  • Agent Security: Monitoring instructions passed between autonomous agents.
  • Compliance Monitoring: Detecting requests that violate corporate safety or regulatory policies.
  • Enterprise Security: Preventing misuse of LLMs for illicit activities.

Limitations

  • Analyzes prompts in isolation, potentially missing multi-turn conversation context.
  • Vulnerable to sophisticated adversarial evasion techniques.
  • Policy coverage is limited to its training data; new threats may require retraining.
  • May occasionally over-block benign prompts that resemble policy violations.

Sena-1-Vega is a risk detection tool, not a final adjudicator, and recommends human review for high-impact decisions.