helloatithya/Neura_Veltrixa

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Mar 13, 2026License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

Neura_Veltrixa is a 1.5 billion parameter instruction-tuned causal language model developed by helloatithya, based on Qwen/Qwen2.5-1.5B-Instruct. It features a 32768-token context length and is specifically fine-tuned for agentic and legal applications. This model excels in tasks requiring legal reasoning and acting as an intelligent agent, leveraging its training on datasets like databricks/databricks-dolly-15k.

Loading preview...

Neura_Veltrixa: An Agentic and Legal LLM

Neura_Veltrixa is a 1.5 billion parameter instruction-tuned language model developed by helloatithya, built upon the robust Qwen/Qwen2.5-1.5B-Instruct architecture. This model is distinguished by its substantial 32768-token context window, enabling it to process and understand extensive inputs, which is particularly beneficial for complex legal documents and multi-turn agentic interactions.

Key Capabilities

  • Agentic Applications: Designed to function effectively as an intelligent agent, capable of understanding instructions and performing tasks in a structured manner.
  • Legal Domain Expertise: Fine-tuned with a focus on legal content, making it suitable for tasks such as legal research, document analysis, and generating legally-relevant text.
  • Extended Context: The 32768-token context length allows for deep comprehension of lengthy texts, crucial for detailed legal analysis or maintaining state in agentic workflows.

Good For

  • Legal Tech Solutions: Developing applications that require understanding and generating legal language.
  • Intelligent Agents: Building AI agents that can perform complex tasks, especially those involving sequential reasoning or interaction.
  • Document Analysis: Processing and extracting information from large legal documents or datasets, leveraging its extended context window.