CHIH-HUNG/llama-2-13b-FINETUNE4_3.8w-r4-q_k_v_o
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Sep 20, 2023License:llama2Architecture:Transformer Open Weights Cold
CHIH-HUNG/llama-2-13b-FINETUNE4_3.8w-r4-q_k_v_o is a 13 billion parameter Llama-2-based language model fine-tuned by CHIH-HUNG using the huangyt/FINETUNE4 dataset, comprising approximately 38,000 data points. This model utilizes LoRA with a rank of 16, targeting q_proj, k_proj, v_proj, and o_proj layers, and is optimized for general language understanding and generation tasks. It demonstrates competitive performance across benchmarks like ARC, HellaSwag, MMLU, and TruthfulQA, making it suitable for applications requiring robust reasoning and factual recall.
Loading preview...