ChuGyouk/F_R13_1

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 28, 2026Architecture:Transformer Cold

ChuGyouk/F_R13_1 is an 8 billion parameter language model fine-tuned from ChuGyouk/Qwen3-8B-Base. This model was trained using SFT with the TRL framework, focusing on general text generation tasks. It offers a 32K context length, making it suitable for applications requiring processing moderately long inputs.

Loading preview...

Model Overview

ChuGyouk/F_R13_1 is an 8 billion parameter language model, fine-tuned from the ChuGyouk/Qwen3-8B-Base architecture. This model leverages the TRL (Transformer Reinforcement Learning) framework for its training, specifically utilizing Supervised Fine-Tuning (SFT).

Key Capabilities

  • General Text Generation: Designed for a broad range of text generation tasks.
  • Base Model Enhancement: Builds upon the capabilities of the Qwen3-8B-Base model through targeted fine-tuning.
  • Context Length: Supports a context window of 32,768 tokens, allowing for processing and generating longer sequences of text.

Training Details

The model's training procedure involved SFT, with progress monitored via Weights & Biases. The development environment included:

  • TRL: 0.24.0
  • Transformers: 5.2.0
  • Pytorch: 2.10.0
  • Datasets: 4.3.0
  • Tokenizers: 0.22.2

Good For

  • Developers looking for a fine-tuned 8B parameter model for various text generation applications.
  • Use cases requiring a substantial context window for processing longer prompts or documents.