mariiaponom/llama_13b_class
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kArchitecture:Transformer Cold

The mariiaponom/llama_13b_class model is a Llama-based language model, likely around 13 billion parameters, that was trained using 4-bit quantization with the bitsandbytes library. It utilizes nf4 quantization and bfloat16 compute dtype for efficient training. This model is primarily characterized by its training methodology, focusing on efficient resource utilization through advanced quantization techniques.

Loading preview...