MicroThinker-1B-Preview is a 1 billion parameter language model developed by huihui-ai, fine-tuned from Llama-3.2-1B-Instruct-abliterated. This model focuses on advancing AI reasoning capabilities, particularly through supervised fine-tuning on specific datasets. It is designed for tasks requiring step-by-step thinking and can be deployed via Ollama.
Loading preview...
Overview
MicroThinker-1B-Preview is a 1 billion parameter model developed by huihui-ai, fine-tuned from the Llama-3.2-1B-Instruct-abliterated base model. Its primary objective is to enhance AI reasoning capabilities, achieved through a supervised fine-tuning (SFT) process. The model was trained using a single RTX 4090 GPU, leveraging specific datasets like huihui-ai/QWQ-LONGCOT-500K and huihui-ai/LONGCOT-Refine-500K.
Key Training Details
- Base Model: Llama-3.2-1B-Instruct-abliterated
- Fine-tuning Method: LoRA (Low-Rank Adaptation) SFT
- Datasets: Utilized 20,000 records from huihui-ai/QWQ-LONGCOT-500K and an additional 20,000 records from huihui-ai/LONGCOT-Refine-500K.
- Training Environment: Performed on a single RTX 4090 GPU with 24GB VRAM.
- System Prompt: "You are a helpful assistant. You should think step-by-step."
Use Cases
This model is suitable for applications requiring:
- Reasoning Tasks: Designed with a focus on improving AI's ability to perform step-by-step reasoning.
- Instruction Following: Benefits from instruction-tuned origins, making it effective for various prompt-based tasks.
- Resource-Constrained Environments: Its 1B parameter size makes it efficient for deployment on systems with limited computational resources, including local inference via Ollama.