CHIH-HUNG/llama-2-13b-FINETUNE3_3.3w-r4-q_k_v_o
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Sep 19, 2023License:llama2Architecture:Transformer Open Weights Cold

CHIH-HUNG/llama-2-13b-FINETUNE3_3.3w-r4-q_k_v_o is a 13 billion parameter Llama-2-based language model fine-tuned by CHIH-HUNG using the huangyt/FINETUNE3 dataset, comprising approximately 33,000 data points. This model leverages LoRA (rank 16) for efficient fine-tuning on q_proj, k_proj, v_proj, and o_proj layers, and is optimized for general language understanding tasks. It demonstrates competitive performance across benchmarks like ARC, HellaSwag, MMLU, and TruthfulQA, making it suitable for applications requiring robust reasoning and knowledge recall.

Loading preview...