lvkaokao/llama2-7b-hf-chat-lora
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kArchitecture:Transformer Cold

The lvkaokao/llama2-7b-hf-chat-lora model is a 7 billion parameter Llama 2-based language model. It was fine-tuned using LoRA with 4-bit quantization (nf4) and double quantization for efficient deployment. This model is designed for chat applications, leveraging the Llama 2 architecture for conversational tasks.

Loading preview...