FrederickSundeep/nova2-14b

TEXT GENERATIONConcurrency Cost:1Model Size:14BQuant:FP8Ctx Length:32kPublished:Apr 17, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

Nova2-14B is a 14.7 billion parameter causal language model developed by Frederick Sundeep Mallela, fine-tuned from Qwen/Qwen3-14B using QLoRA. This model is optimized to function as an AI assistant named Nova, maintaining a consistent persona and identity. It retains the base model's capabilities in coding, reasoning, and mathematics, primarily intended for general-purpose AI assistant tasks and powering the NovaMind chat application.

Loading preview...

Nova2-14B: A Persona-Driven AI Assistant

Nova2-14B is a 14.7 billion parameter large language model developed by Frederick Sundeep Mallela. It is fine-tuned from the powerful Qwen/Qwen3-14B base model using QLoRA and Unsloth, resulting in a fully merged, standalone model that requires no adapter dependencies for inference. While the base model supports up to 40K tokens, Nova2-14B is fine-tuned with a maximum sequence length of 2048 tokens.

Key Differentiators & Capabilities

Nova2-14B maintains all the robust capabilities of its Qwen3-14B base, including:

  • Code generation: Supports Python, JavaScript, C++, SQL, and more.
  • Reasoning and Math: Excels in logical problem-solving and advanced mathematics.
  • Multilingual support: Inherits support for over 100 languages.
  • Consistent Persona: Responds as "Nova," an AI assistant created by Frederick, with a consistent identity and tone.
  • Instruction following: Designed for precise task execution and fully supports custom system prompts.
  • Tool use: Compatible with function calling.

Training and Optimization

The model was fine-tuned using a custom dataset focused on establishing Nova's identity, technical knowledge, and personality. This process involved Supervised Fine-Tuning (SFT) with QLoRA, utilizing a Tesla T4 GPU. The fine-tuning specifically trained Nova to never reveal its underlying architecture details, making it suitable for integrated application use like the NovaMind chat application.

Intended Use Cases

  • Powering the NovaMind AI chat application.
  • General-purpose AI assistant tasks.
  • Code generation and debugging.
  • Technical question answering.
  • As a base model for further fine-tuning.

Limitations

  • Not evaluated on standard benchmarks post fine-tuning.
  • May occasionally revert to base Qwen3 behavior due to a relatively small custom dataset.
  • Context limited to 2048 tokens in its fine-tuned configuration, despite the base model's 40K capability.