Davletovarch/logos-v1-merged

TEXT GENERATIONConcurrency Cost:1Model Size:14BQuant:FP8Ctx Length:32kPublished:Mar 18, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

Davletovarch/logos-v1-merged is a 14 billion parameter Qwen3-based causal language model developed by Davletovarch. This model was fine-tuned using Unsloth, a technique known for accelerating training. It offers a substantial parameter count and a 32768-token context length, making it suitable for complex language understanding and generation tasks.

Loading preview...

Model Overview

Davletovarch/logos-v1-merged is a 14 billion parameter language model built upon the Qwen3 architecture. Developed by Davletovarch, this model distinguishes itself through its training methodology, leveraging the Unsloth framework. Unsloth is noted for significantly accelerating the fine-tuning process of large language models, enabling faster iteration and deployment.

Key Characteristics

  • Base Model: Fine-tuned from unsloth/Qwen3-14B, inheriting its robust capabilities.
  • Training Efficiency: Utilizes Unsloth for a reported 2x faster training speed, which can translate to more efficient model development and updates.
  • Parameter Count: With 14 billion parameters, it is capable of handling intricate language tasks.
  • Context Length: Supports a substantial context window of 32768 tokens, allowing for processing and generating longer, more coherent texts.

Potential Use Cases

Given its foundation and parameter size, Davletovarch/logos-v1-merged is well-suited for applications requiring:

  • Advanced text generation and completion.
  • Complex question answering and summarization.
  • Conversational AI and chatbot development.
  • Tasks benefiting from a large context window for understanding extensive documents or dialogues.