Alelcv27/Llama3.1-8B-Base-Linear-Math-Code

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Apr 24, 2026Architecture:Transformer Cold

Alelcv27/Llama3.1-8B-Base-Linear-Math-Code is an 8 billion parameter language model, merged from Llama3.1-8B-Base-Math and Llama3.1-8B-Base-Code using the Linear merge method. This model is specifically designed to excel in both mathematical reasoning and code generation tasks, leveraging a 32768 token context length. It offers a balanced capability for applications requiring strong performance in numerical problem-solving and programming contexts.

Loading preview...

Model Overview

This model, Alelcv27/Llama3.1-8B-Base-Linear-Math-Code, is an 8 billion parameter language model built upon the Llama3.1 architecture. It was created by Alelcv27 through a strategic merge of two specialized base models: Alelcv27/Llama3.1-8B-Base-Math and Alelcv27/Llama3.1-8B-Base-Code. The merging process utilized the Linear merge method from mergekit, aiming to combine and enhance their respective strengths.

Key Capabilities

  • Dual Specialization: Inherits and integrates capabilities from models specifically trained for mathematical reasoning and code generation.
  • Linear Merge: Employs a weighted linear combination of the base models' layers, with a 60% emphasis on code capabilities and 40% on mathematical understanding, across layers 0 to 32.
  • Context Length: Supports a substantial context window of 32768 tokens, beneficial for complex problems in both domains.

Ideal Use Cases

  • Mathematical Problem Solving: Suitable for tasks requiring numerical reasoning, equation solving, and logical deduction.
  • Code Generation: Effective for generating, understanding, and assisting with programming code across various languages.
  • Hybrid Applications: Well-suited for scenarios where both strong mathematical and coding abilities are simultaneously required, such as scientific computing or algorithmic development.