JetBrains/Mellum-4b-sft-python

Warm
Public
4B
BF16
32768
Apr 30, 2025
License: apache-2.0
Hugging Face
Overview

Overview

JetBrains/Mellum-4b-sft-python is a 4 billion parameter LLaMA-style large language model (LLM) developed by JetBrains. It is a fine-tuned version of their initial open-source LLM, specifically optimized for code-related tasks, particularly Python code completion. The model was pre-trained on over 4 trillion tokens with an 8192-token context window, using Automatic Mixed Precision (AMP) with bf16 precision.

Key Capabilities

  • Python Code Completion: Tailored for generating and completing Python code.
  • Code-Related Tasks: Optimized for various programming-centric applications.
  • Efficient Deployment: Designed for both cloud inference (e.g., via vLLM) and local deployment (e.g., using llama.cpp or Ollama).
  • Context Handling: Supports fill-in-the-middle generation with additional file context.

Good For

  • Professional Developer Tooling: Ideal for intelligent code suggestions in IDEs.
  • AI-Powered Coding Assistants: Enhancing coding workflows with AI support.
  • Research: Suitable for studies on code understanding and generation.
  • Educational Applications: Can be used in learning environments for programming.
  • Fine-tuning Experiments: A solid base for further specialized fine-tuning.

Limitations

  • May reflect biases present in public codebases, potentially producing code similar in style to open-source repositories.
  • Code suggestions should not be assumed to be secure or free of vulnerabilities.