huihui-ai/MicroThinker-1B-Preview
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:Jan 2, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

MicroThinker-1B-Preview is a 1 billion parameter language model developed by huihui-ai, fine-tuned from Llama-3.2-1B-Instruct-abliterated. This model focuses on advancing AI reasoning capabilities, particularly through supervised fine-tuning on specific datasets. It is designed for tasks requiring step-by-step thinking and can be deployed via Ollama.

Loading preview...

Overview

MicroThinker-1B-Preview is a 1 billion parameter model developed by huihui-ai, fine-tuned from the Llama-3.2-1B-Instruct-abliterated base model. Its primary objective is to enhance AI reasoning capabilities, achieved through a supervised fine-tuning (SFT) process. The model was trained using a single RTX 4090 GPU, leveraging specific datasets like huihui-ai/QWQ-LONGCOT-500K and huihui-ai/LONGCOT-Refine-500K.

Key Training Details

  • Base Model: Llama-3.2-1B-Instruct-abliterated
  • Fine-tuning Method: LoRA (Low-Rank Adaptation) SFT
  • Datasets: Utilized 20,000 records from huihui-ai/QWQ-LONGCOT-500K and an additional 20,000 records from huihui-ai/LONGCOT-Refine-500K.
  • Training Environment: Performed on a single RTX 4090 GPU with 24GB VRAM.
  • System Prompt: "You are a helpful assistant. You should think step-by-step."

Use Cases

This model is suitable for applications requiring:

  • Reasoning Tasks: Designed with a focus on improving AI's ability to perform step-by-step reasoning.
  • Instruction Following: Benefits from instruction-tuned origins, making it effective for various prompt-based tasks.
  • Resource-Constrained Environments: Its 1B parameter size makes it efficient for deployment on systems with limited computational resources, including local inference via Ollama.