mohit-1710/loomstack-qwen-4b-sft-terminal

TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Apr 26, 2026Architecture:Transformer Cold

The mohit-1710/loomstack-qwen-4b-sft-terminal is a 4 billion parameter language model, fine-tuned from Itachi-42/loomstack-qwen-4b-sft-compact using SFT (Supervised Fine-Tuning) with a context length of 32768 tokens. This model is designed for general text generation tasks, building upon the Qwen architecture. Its fine-tuning process aims to enhance its conversational and instruction-following capabilities for terminal-like interactions.

Loading preview...

Overview

The mohit-1710/loomstack-qwen-4b-sft-terminal is a 4 billion parameter language model, fine-tuned from the Itachi-42/loomstack-qwen-4b-sft-compact base model. This fine-tuning process utilized Supervised Fine-Tuning (SFT) with the TRL library, aiming to adapt the model for improved performance in interactive, terminal-like environments. It supports a substantial context length of 32768 tokens, allowing for processing and generating longer sequences of text.

Key Capabilities

  • General Text Generation: Capable of generating coherent and contextually relevant text based on given prompts.
  • Instruction Following: Enhanced through SFT to better understand and respond to user instructions.
  • Extended Context Window: Benefits from a 32768-token context length, suitable for tasks requiring extensive conversational history or detailed input.

Training Details

The model was trained using the SFT method, leveraging the TRL (Transformer Reinforcement Learning) library. This approach focuses on optimizing the model's responses to align with desired output patterns and conversational styles. The training environment included specific versions of key frameworks such as TRL 0.24.0, Transformers 5.5.0, and PyTorch 2.10.0.

Good For

  • Developing conversational agents or chatbots that require a moderate parameter count and good instruction-following.
  • Applications needing a model with a large context window for processing detailed queries or maintaining long dialogue states.
  • Experimentation with SFT-tuned models for various text generation tasks.