kmseong/llama2_7b_base-gsm8k_lora_ft_lr1e-4

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Apr 23, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The kmseong/llama2_7b_base-gsm8k_lora_ft_lr1e-4 is a 7 billion parameter Llama 2 base model fine-tuned by kmseong using LoRA for the GSM8K mathematical reasoning dataset. This model is specifically optimized for arithmetic and common sense reasoning tasks, aiming to enhance performance on quantitative problem-solving. It is suitable for applications requiring robust numerical and logical inference capabilities.

Loading preview...

Model Overview

The kmseong/llama2_7b_base-gsm8k_lora_ft_lr1e-4 is a 7 billion parameter Llama 2 base model that has undergone LoRA (Low-Rank Adaptation) fine-tuning. This specific iteration was trained with a learning rate of 1e-4, targeting enhanced performance on the GSM8K dataset, which focuses on grade school math word problems.

Key Capabilities

  • Mathematical Reasoning: Optimized for solving arithmetic and common sense reasoning problems, as demonstrated by its fine-tuning on the GSM8K dataset.
  • Parameter-Efficient Fine-tuning: Utilizes LoRA, allowing for efficient adaptation of the base Llama 2 model without retraining all parameters.
  • Llama 2 Architecture: Benefits from the robust and widely-used Llama 2 foundational architecture.

Good For

  • Quantitative Problem Solving: Ideal for applications requiring the model to perform numerical calculations and logical deductions.
  • Educational Tools: Can be integrated into systems designed to assist with or generate mathematical problems and solutions.
  • Research in Reasoning: Useful for researchers exploring the capabilities of large language models in mathematical and logical reasoning tasks, particularly within the Llama 2 ecosystem.