mlabonne/llama-2-7b-miniguanaco
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Cold

The mlabonne/llama-2-7b-miniguanaco model is a 7 billion parameter Llama-2-7b-chat-hf variant, fine-tuned by mlabonne using QLoRA (4-bit precision) on a subset of the OpenAssistant Guanaco dataset. This model is primarily designed for educational purposes, demonstrating fine-tuning techniques on a T4 GPU. Its main use case is for learning and experimentation with LLM fine-tuning rather than high-performance inference.

Loading preview...