EleutherAI/llemma_7b

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Sep 12, 2023License:llama2Architecture:Transformer0.1K Open Weights Warm

EleutherAI's Llemma 7B is a 7 billion parameter language model specifically designed for mathematics. Initialized from Code Llama 7B, it was further trained on the Proof-Pile-2 dataset for 200 billion tokens. This model excels at chain-of-thought mathematical reasoning and is adept at using computational tools for mathematical tasks, outperforming Llama-2 and Code Llama on various math benchmarks.

Loading preview...

Llemma 7B: A Language Model for Mathematics

Llemma 7B, developed by EleutherAI, is a 7 billion parameter language model specialized in mathematical reasoning and problem-solving. It was built upon the Code Llama 7B architecture and underwent extensive training on the Proof-Pile-2 dataset, comprising 200 billion tokens, to enhance its mathematical capabilities.

Key Capabilities

  • Chain-of-Thought Mathematical Reasoning: Llemma 7B demonstrates strong performance in complex mathematical reasoning tasks, outperforming Llama-2 and Code Llama of similar sizes on benchmarks like GSM8k, OCW, MMLU-STEM, SAT, and MATH.
  • Computational Tool Use: The model is proficient in utilizing computational tools, such as Python and formal theorem provers, for mathematical applications.
  • Competitive Performance: When controlling for model size, Llemma 7B surpasses Minerva models on several mathematical benchmarks, particularly in chain-of-thought reasoning.

Use Cases

  • Mathematical Problem Solving: Ideal for tasks requiring step-by-step mathematical reasoning.
  • Assisting with Formal Proofs: Can be used in conjunction with theorem provers.
  • Educational Tools: Potentially useful for generating explanations or solutions for mathematical problems.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p