Luciano/Llama-2-7b-chat-hf-miniguanaco
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kArchitecture:Transformer Cold

Luciano/Llama-2-7b-chat-hf-miniguanaco is a Llama-2-7b-chat-hf-based model developed by Luciano. This model was trained using 4-bit quantization with the nf4 quantization type and float16 compute dtype. Its training procedure utilized specific bitsandbytes configurations, indicating an optimization for efficient resource usage during fine-tuning. The model's primary characteristic is its training methodology, focusing on quantized training for potential deployment efficiency.

Loading preview...