MachineLearningLM/MachineLearningLM-7B-v1

Cold
Public
7.6B
FP8
131072
License: apache-2.0
Hugging Face
Overview

MachineLearningLM-7B-v1: Specialized for Many-shot In-context Learning on Tabular Data

MachineLearningLM-7B-v1 is a 7.6 billion parameter model specifically designed for advanced in-context learning, particularly with tabular data. It was continuously pretrained on millions of synthetic tabular machine learning tasks, enabling it to robustly handle many-shot examples.

Key Capabilities

  • Scalable In-context Learning: Demonstrates strong performance scaling from 8 to 1,024 in-context examples, a significant capability for complex data scenarios.
  • Enhanced Tabular Task Performance: Achieves approximately 15% improvement on unseen tabular tasks compared to other models like o3-mini, GPT-5-mini, and Qwen-2.5-7B-Instruct.
  • Robust Numerical Modeling: Exhibits Random-Forest-level robustness in numerical modeling, indicating high reliability for data-driven predictions.
  • General Language Understanding: Maintains a strong MMLU score of 75.4%, suggesting solid general reasoning abilities alongside its specialization.
  • Open-sourced Evaluation Framework: Comes with an automated evaluation framework, including code for data preparation, prompt generation, model prediction, and result processing, available on its GitHub repository.

Good for

  • Complex Tabular Data Analysis: Ideal for applications requiring robust analysis and prediction on structured, tabular datasets.
  • Many-shot Learning Scenarios: Excels when a large number of in-context examples are available to guide the model's predictions.
  • Research and Development: The open-sourced evaluation and data generation tools make it suitable for researchers exploring advanced in-context learning and tabular ML.
  • Benchmarking: Can serve as a strong baseline or comparison model for tasks involving numerical modeling and tabular data.