narvind2003/llama-2-7b-miniguanaco
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kLicense:mitArchitecture:Transformer0.0K Open Weights Cold

The narvind2003/llama-2-7b-miniguanaco is a 7 billion parameter Llama 2 model, fine-tuned by narvind2003 using QLoRA on 1000 samples from the Guanaco dataset. This model is optimized for conversational tasks, leveraging quantization and low-rank adaptation for efficient deployment. It offers a 4096-token context length, making it suitable for applications requiring focused dialogue generation.

Loading preview...