ChuGyouk/F_R13_1
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 28, 2026Architecture:Transformer Cold
ChuGyouk/F_R13_1 is an 8 billion parameter language model fine-tuned from ChuGyouk/Qwen3-8B-Base. This model was trained using SFT with the TRL framework, focusing on general text generation tasks. It offers a 32K context length, making it suitable for applications requiring processing moderately long inputs.
Loading preview...
Model Overview
ChuGyouk/F_R13_1 is an 8 billion parameter language model, fine-tuned from the ChuGyouk/Qwen3-8B-Base architecture. This model leverages the TRL (Transformer Reinforcement Learning) framework for its training, specifically utilizing Supervised Fine-Tuning (SFT).
Key Capabilities
- General Text Generation: Designed for a broad range of text generation tasks.
- Base Model Enhancement: Builds upon the capabilities of the Qwen3-8B-Base model through targeted fine-tuning.
- Context Length: Supports a context window of 32,768 tokens, allowing for processing and generating longer sequences of text.
Training Details
The model's training procedure involved SFT, with progress monitored via Weights & Biases. The development environment included:
- TRL: 0.24.0
- Transformers: 5.2.0
- Pytorch: 2.10.0
- Datasets: 4.3.0
- Tokenizers: 0.22.2
Good For
- Developers looking for a fine-tuned 8B parameter model for various text generation applications.
- Use cases requiring a substantial context window for processing longer prompts or documents.