shitshow123/tinylamma-20000

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1.1BQuant:BF16Ctx Length:2kPublished:Jan 9, 2024License:apache-2.0Architecture:Transformer Open Weights Warm

The shitshow123/tinylamma-20000 is a 1.1 billion parameter instruction-tuned causal language model, developed by shitshow123. This model has been trained for 20,000 steps using Direct Preference Optimization (DPO) on a TinyLlama 1B base, making it suitable for tasks requiring refined instruction following. It features a context length of 2048 tokens, focusing on efficient performance for specific instruction-based applications.

Loading preview...

shitshow123/tinylamma-20000 Overview

The shitshow123/tinylamma-20000 is a compact yet capable language model, built upon the TinyLlama 1B architecture. Developed by shitshow123, this model has undergone extensive instruction tuning using the Direct Preference Optimization (DPO) method for 20,000 steps.

Key Characteristics

  • Base Model: TinyLlama 1B
  • Parameter Count: 1.1 billion parameters
  • Training Method: Direct Preference Optimization (DPO) for 20,000 steps
  • Context Length: 2048 tokens

Intended Use Cases

This model is primarily designed for applications where a smaller footprint and efficient instruction following are critical. Its DPO training suggests an optimization for generating responses aligned with human preferences, making it potentially suitable for:

  • Instruction-based text generation
  • Lightweight conversational agents
  • Tasks requiring adherence to specific prompts within its context window.