unsloth/phi-2

Warm
Public
3B
BF16
2048
Feb 22, 2024
License: apache-2.0
Hugging Face
Overview

Unsloth: Efficient Fine-tuning for LLMs

Unsloth provides a framework for accelerated and memory-efficient fine-tuning of popular large language models, including Gemma, Mistral, Llama-2, and TinyLlama. It enables developers to fine-tune these models up to 5 times faster while using 70% less memory compared to traditional methods.

Key Capabilities

  • Speed and Efficiency: Achieves significant reductions in training time and memory footprint, making fine-tuning more accessible, even on consumer-grade hardware or free tiers like Colab and Kaggle.
  • Broad Model Support: Compatible with various architectures such as Gemma 7b, Mistral 7b, Llama-2 7b, TinyLlama, and CodeLlama 34b.
  • Export Options: Fine-tuned models can be exported to formats like GGUF and vLLM, or directly uploaded to Hugging Face.
  • Beginner-Friendly: Offers pre-configured notebooks for various tasks, including conversational models (ShareGPT ChatML / Vicuna templates), text completion, and DPO (Direct Preference Optimization) for models like Zephyr.

Good For

  • Developers and researchers seeking to rapidly iterate on fine-tuning experiments.
  • Users with limited GPU resources who need to fine-tune large models efficiently.
  • Creating specialized versions of existing LLMs for specific downstream tasks with reduced computational overhead.