ai-sherpa/llama2-mas-trmg
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kArchitecture:Transformer Cold

The ai-sherpa/llama2-mas-trmg is a 7 billion parameter Llama 2-based model developed by ai-sherpa. This model was trained using 4-bit quantization with the nf4 quantization type and double quantization enabled, leveraging PEFT for efficient fine-tuning. Its training methodology suggests a focus on optimizing performance within resource-constrained environments, making it suitable for applications requiring efficient deployment.

Loading preview...