arkoda/arkoda-7b-v7-10

TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:May 7, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

arkoda/arkoda-7b-v7-10 is a 7.6 billion parameter Qwen2-based causal language model developed by arkoda. This instruction-tuned model was fine-tuned using Unsloth and Huggingface's TRL library, enabling faster training. It is designed for general-purpose language generation tasks, leveraging its Qwen2 architecture for robust performance.

Loading preview...

Model Overview

arkoda/arkoda-7b-v7-10 is a 7.6 billion parameter instruction-tuned language model built upon the Qwen2 architecture. Developed by arkoda, this model was fine-tuned using the Unsloth library in conjunction with Huggingface's TRL library, which facilitated a 2x faster training process compared to standard methods.

Key Characteristics

  • Base Model: Fine-tuned from unsloth/qwen2.5-7b-instruct-bnb-4bit, indicating a foundation in the Qwen2.5 series.
  • Training Efficiency: Leverages Unsloth for optimized and accelerated training.
  • Parameter Count: Features 7.6 billion parameters, placing it in the medium-sized LLM category.
  • Context Length: Supports a substantial context window of 32,768 tokens.

Use Cases

This model is suitable for a variety of general-purpose language generation and instruction-following tasks, benefiting from its Qwen2 foundation and efficient fine-tuning.