CodeAtCMU/Llama-3.2-1B-GenerativePerturbations_full_sft_code_data_120K_imaginary
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:Sep 3, 2025Architecture:Transformer Warm

CodeAtCMU/Llama-3.2-1B-GenerativePerturbations_full_sft_code_data_120K_imaginary is a 1 billion parameter language model developed by CodeAtCMU, based on the Llama-3.2 architecture. This model is fine-tuned with 120,000 imaginary code data points, suggesting an optimization for code generation and understanding tasks. With a context length of 32768 tokens, it is designed to handle extensive code inputs and outputs, making it suitable for complex programming-related applications.

Loading preview...

Model Overview

This model, CodeAtCMU/Llama-3.2-1B-GenerativePerturbations_full_sft_code_data_120K_imaginary, is a 1 billion parameter language model built upon the Llama-3.2 architecture. It has been specifically fine-tuned using a dataset of 120,000 'imaginary' code data points, indicating a focus on enhancing its capabilities in code-related tasks.

Key Characteristics

  • Architecture: Llama-3.2 base model.
  • Parameter Count: 1 billion parameters.
  • Context Length: Supports a substantial context window of 32768 tokens, allowing for processing of large codebases or complex programming problems.
  • Training Data: Fine-tuned with 120,000 synthetic or 'imaginary' code data points, suggesting a specialized training approach for code generation and comprehension.

Potential Use Cases

Given its specialized training, this model is likely optimized for:

  • Code Generation: Creating new code snippets or functions based on natural language prompts.
  • Code Completion: Assisting developers by suggesting code as they type.
  • Code Understanding: Analyzing and interpreting existing code, potentially for tasks like bug detection or refactoring suggestions.
  • Educational Tools: Aiding in learning programming by generating examples or explaining concepts.