ArpitKadam/llama-2-7b-guanaco-finetune
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Dec 15, 2025License:llama2Architecture:Transformer Open Weights Cold

ArpitKadam/llama-2-7b-guanaco-finetune is a 7 billion parameter decoder-only causal language model, fine-tuned by Arpit Sachin Kadam from the LLaMA-2-7B-Chat base using QLoRA (4-bit LoRA fine-tuning). This parameter-efficient variant is optimized for improved instruction following and conversational response quality. It is primarily intended for use in instruction following, chat assistants, and question-answering applications.

Loading preview...