wei123602/llama2-13b-fintune2

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kPublished:Sep 4, 2023License:llama2Architecture:Transformer Open Weights Warm

The wei123602/llama2-13b-fintune2 model is a 13 billion parameter Llama-2 based language model fine-tuned by wei123602. It was trained using the huangyt/FINETUNE2 dataset, comprising approximately 30,000 training samples, with a methodology similar to Platypus. This model is optimized for general language understanding and generation tasks, building upon the robust Llama-2 architecture.

Loading preview...

Model Overview

This model, wei123602/llama2-13b-fintune2, is a 13 billion parameter language model built upon the meta-llama/Llama-2-13b-hf base architecture. It has been fine-tuned by wei123602 using the huangyt/FINETUNE2 dataset, which contains approximately 30,000 training entries. The fine-tuning process utilized LoRA (Low-Rank Adaptation) with specific parameters, including a LoRA rank of 16, alpha of 8, and a dropout of 0.05, targeting gate_proj, up_proj, and down_proj modules.

Training Details

The model was trained for 1 epoch on a single RTX4090 GPU, achieving a train_loss of 0.0823 over a train_runtime of 2 hours and 40 minutes. Key training parameters included a per_device_train_batch_size of 8, gradient_accumulation_steps of 8, and a learning_rate of 4e-4. The cutoff_length for sequences was 2048 tokens, and training was performed in bf16 precision.

Evaluation

Evaluation results for this specific model are currently pending on the HuggingFaceH4/open_llm_leaderboard. However, the README provides comparative benchmarks for other Llama-2-13b variants on metrics such as ARC, HellaSwag, MMLU, and TruthfulQA, indicating the general performance range for models in this family. The fine-tuning approach aims to enhance performance on general language tasks through targeted dataset training.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p