Glavin001/coqar-questions-llama-2-7b-v0.1
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kArchitecture:Transformer0.0K Cold

Glavin001/coqar-questions-llama-2-7b-v0.1 is a Llama 2-7b based model developed by Glavin001. This model was trained using 8-bit quantization with bitsandbytes, leveraging PEFT for efficient fine-tuning. Its primary characteristic is the use of specific quantization configurations during training, which can impact performance and resource usage. The model is likely intended for tasks where efficient deployment and reduced memory footprint are critical, given its training methodology.

Loading preview...