habanoz/TinyLlama-1.1B-intermediate-step-715k-1.5T-lr-5-1epch-airoboros3.1-1k-instruct-V1

Warm
Public
1.1B
BF16
2048
License: apache-2.0
Hugging Face
Overview

Model Overview

The habanoz/TinyLlama-1.1B-intermediate-step-715k-1.5T-lr-5-1epch-airoboros3.1-1k-instruct-V1 is a 1.1 billion parameter language model derived from the TinyLlama-1.1B-intermediate-step-715k-1.5T base model. It has been instruction-tuned using the airoboros-3.1-no-mathjson-max-1k dataset, leveraging QLoRA for efficient fine-tuning, with the adapter subsequently merged into the base model.

Key Capabilities & Performance

This model demonstrates general instruction-following abilities, as indicated by its performance on the Open LLM Leaderboard. Key evaluation metrics include:

  • Avg. Score: 34.98
  • AI2 Reasoning Challenge (25-Shot): 30.72
  • HellaSwag (10-Shot): 54.32
  • MMLU (5-Shot): 24.78
  • TruthfulQA (0-shot): 41.67
  • Winogrande (5-shot): 57.62
  • GSM8k (5-shot): 0.76

Training Details

The model was fine-tuned for 1 epoch with a learning rate of 1e-5, utilizing 4-bit quantization (nf4) and double quantization. It employs a model_max_len of 1024 tokens and incorporates Flash Attention 2 for improved efficiency. The training process involved a per_device_train_batch_size of 4 and gradient_accumulation_steps of 4.

Use Cases

Given its instruction-tuned nature and performance on various benchmarks, this model is suitable for general-purpose conversational AI, text generation, and tasks requiring basic reasoning and common sense understanding, particularly in resource-constrained environments due to its small size.