alibayram/magibu-128k-trained

VISIONConcurrency Cost:1Model Size:12BQuant:FP8Ctx Length:32kPublished:Jan 11, 2026Architecture:Transformer Cold

The alibayram/magibu-128k-trained model is a 12 billion parameter language model fine-tuned from alibayram/magibu-128k-embed-init. Trained using the TRL framework, this model is designed for text generation tasks. Its training process involved Supervised Fine-Tuning (SFT) to enhance its performance in generating coherent and relevant text.

Loading preview...

Model Overview

alibayram/magibu-128k-trained is a 12 billion parameter language model that has been fine-tuned from its base model, alibayram/magibu-128k-embed-init. This model leverages the TRL (Transformer Reinforcement Learning) framework for its training, specifically employing Supervised Fine-Tuning (SFT) techniques.

Key Capabilities

  • Text Generation: The model is primarily designed for generating human-like text based on given prompts or questions.
  • Fine-tuned Performance: Through SFT, the model aims to produce more refined and contextually appropriate outputs compared to its base version.

Training Details

The training process for magibu-128k-trained was conducted using TRL version 0.24.0, with Transformers 4.57.3, Pytorch 2.9.1, Datasets 4.3.0, and Tokenizers 0.22.2. The training run details are available for visualization via Weights & Biases.

Good For

  • Conversational AI: Generating responses for interactive applications.
  • Creative Writing: Assisting in generating various forms of creative text.
  • General Text Completion: Providing coherent continuations for given text inputs.