AliMaatouk/TinyLlama-1.1B-Tele
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1.1BQuant:BF16Ctx Length:2kPublished:Sep 8, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

The AliMaatouk/TinyLlama-1.1B-Tele is a 1.1 billion parameter Transformer model developed by Ali Maatouk, specialized for telecommunications. It was continually pretrained on the 2.5 billion token Tele-Data dataset, comprising telecommunications articles, standards, and web content. This model outperforms its base, TinyLlama-1.1B, on telecommunications benchmarks while maintaining performance on general language understanding tasks. It is primarily intended as a base model for fine-tuning on telecommunications-related applications.

Loading preview...

TinyLlama-1.1B-Tele: A Specialized LLM for Telecommunications

The TinyLlama-1.1B-Tele is a 1.1 billion parameter Transformer model developed by Ali Maatouk, specifically designed for the telecommunications domain. It is an adaptation of the original TinyLlama-1.1B, having undergone continuous pretraining on the extensive Tele-Data dataset. This dataset consists of approximately 2.5 billion tokens of telecommunications-specific material, including articles, industry standards, and general web content.

Key Capabilities and Performance

  • Domain Specialization: Optimized for telecommunications, demonstrating improved performance on relevant benchmarks like Tele-Eval compared to its base model.
  • Performance Retention: Despite its specialization, the model maintains performance levels comparable to TinyLlama-1.1B on general benchmarks for common sense, language understanding, and logical reasoning.
  • Context Length: Trained with a context length of 2048 tokens.

Intended Use Cases

  • Base Model for Fine-tuning: Best suited as a foundational model for further fine-tuning on specific telecommunications applications.
  • Text Completion: Operates within a text completion framework, providing relevant continuations for telecommunications-related prompts.
  • Research: Useful for researchers exploring domain adaptation in small language models. An instruct-tuned version is also available at TinyLlama-1.1B-Tele-it.