pepe4235/second_try
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kArchitecture:Transformer Cold
pepe4235/second_try is a 13 billion parameter language model. This model was trained using bitsandbytes 4-bit quantization, specifically with the nf4 quantization type and float16 compute dtype. It leverages PEFT 0.6.0.dev0 for efficient fine-tuning. This model is suitable for tasks requiring a moderately sized language model with efficient quantization for deployment.
Loading preview...