Alelcv27/Llama3.2-3B-SLERP-Math-Code

TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Apr 20, 2026Architecture:Transformer Cold

Alelcv27/Llama3.2-3B-SLERP-Math-Code is a 3.2 billion parameter language model, merged from Llama3.2-3B-Base-Code and Llama3.2-3B-Base-Math using the SLERP method. This model is specifically designed to excel in both mathematical reasoning and code generation tasks. With a 32768 token context length, it offers enhanced capabilities for complex problem-solving in these domains. It is optimized for developers requiring a compact yet powerful model for specialized math and coding applications.

Loading preview...

Alelcv27/Llama3.2-3B-SLERP-Math-Code Overview

This model, developed by Alelcv27, is a 3.2 billion parameter language model created through a strategic merge of two specialized base models: Alelcv27/Llama3.2-3B-Base-Code and Alelcv27/Llama3.2-3B-Base-Math. The merging process utilized the SLERP (Spherical Linear Interpolation) method, a technique often employed to combine the strengths of different models while maintaining performance.

Key Capabilities

  • Dual Specialization: Inherits and combines the capabilities of its base models, making it proficient in both mathematical reasoning and code generation.
  • Merged Architecture: Leverages the SLERP merge method to integrate distinct functionalities from code-focused and math-focused models.
  • Context Length: Supports a substantial context window of 32768 tokens, beneficial for handling longer code snippets or complex mathematical problems.

Good For

  • Mathematical Problem Solving: Ideal for applications requiring robust mathematical understanding and computation.
  • Code Generation & Analysis: Suitable for tasks involving programming, script generation, and code-related queries.
  • Resource-Efficient Specialized Tasks: Offers a powerful solution for specific math and coding use cases within a 3.2 billion parameter footprint.