ogbanugot/llama-2-7b-miniguanaco
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kArchitecture:Transformer Cold

The ogbanugot/llama-2-7b-miniguanaco is a 7 billion parameter language model based on the Llama 2 architecture, fine-tuned using 4-bit quantization with nf4 type and float16 compute dtype. This model leverages PEFT 0.4.0 for efficient training, making it suitable for applications requiring a balance of performance and resource efficiency. Its training configuration suggests an optimization for deployment in environments with limited computational resources.

Loading preview...