habanoz/TinyLlama-1.1B-intermediate-step-715k-1.5T-lr-5-4epochs-oasst1-top1-instruct-V1

Warm
Public
1.1B
BF16
2048
License: apache-2.0
Hugging Face
Overview

Model Overview

This model, habanoz/TinyLlama-1.1B-intermediate-step-715k-1.5T-lr-5-4epochs-oasst1-top1-instruct-V1, is a 1.1 billion parameter instruction-tuned language model. It is based on the TinyLlama-1.1B-intermediate-step-715k-1.5T base model and has been fine-tuned using the OpenAssistant/oasst_top1_2023-08-25 dataset.

Training Details

The instruction tuning was performed using QLoRA (Quantized Low-Rank Adapters) with the adapter merged into the base model. Key training parameters included 4 epochs, a learning rate of 1e-5, and a model_max_len of 1024 tokens. The training utilized 4-bit quantization with nf4 type and double quantization.

Performance Benchmarks

Evaluated on the Open LLM Leaderboard, the model achieved an average score of 35.28. Specific scores include:

  • AI2 Reasoning Challenge (25-Shot): 31.14
  • HellaSwag (10-Shot): 54.31
  • MMLU (5-Shot): 25.42
  • TruthfulQA (0-shot): 41.72
  • Winogrande (5-shot): 57.77
  • GSM8k (5-shot): 1.29

Use Cases

Given its instruction-tuned nature and compact size, this model is suitable for applications requiring efficient, general-purpose instruction following, particularly where computational resources are limited. Its performance metrics suggest utility in tasks like common sense reasoning and question answering, though its mathematical reasoning (GSM8k) is limited.