Trelis/TinyLlama-chat-SFT
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1.1BQuant:BF16Ctx Length:2kPublished:Mar 27, 2024License:apache-2.0Architecture:Transformer Open Weights Warm

Trelis/TinyLlama-chat-SFT is a 1.1 billion parameter instruction-tuned causal language model developed by Trelis Research, based on the TinyLlama project. It utilizes the Llama 2 architecture and tokenizer, making it compatible with existing Llama-based open-source projects. This compact model is fine-tuned for chat applications, offering a balance of performance and efficiency for environments with restricted computational and memory resources.

Loading preview...

Trelis/TinyLlama-chat-SFT: A Compact, Chat-Optimized LLM

This model is an instruction-tuned version of the TinyLlama-1.1B base model, developed by Trelis Research. The original TinyLlama project aimed to pretrain a 1.1 billion parameter Llama model on 3 trillion tokens, adopting the Llama 2 architecture and tokenizer. This design choice ensures compatibility with a wide range of open-source projects built upon Llama.

Key Capabilities & Characteristics

  • Compact Size: With only 1.1 billion parameters, TinyLlama-chat-SFT is designed for applications requiring limited computation and memory.
  • Llama 2 Compatibility: Shares the same architecture and tokenizer as Llama 2, allowing for seamless integration into existing Llama-based ecosystems.
  • Pretraining Scale: The base model was pretrained on a massive 3 trillion tokens, contributing to its general language understanding.

Ideal Use Cases

  • Resource-Constrained Environments: Suitable for deployment where computational power or memory footprint is a critical concern.
  • Llama Ecosystem Integration: Easily plugs into projects already utilizing the Llama 2 framework.
  • Chat Applications: Fine-tuned for conversational tasks, making it appropriate for chatbots and interactive AI agents.