qihoo360/Light-IF-4B

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Aug 4, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

Light-IF-4B is a 4 billion parameter language model developed by Qihoo360, specifically designed to improve instruction following and reasoning abilities in LLMs. It addresses "lazy reasoning" by promoting rigorous thought processes through previewing and self-checking mechanisms. The model is fine-tuned using Entropy-SFT and TEA-RL on a high-quality, complex instruction dataset, making it particularly effective for tasks requiring precise adherence to instructions and complex reasoning.

Loading preview...

Overview

Light-IF-4B is a 4 billion parameter language model from Qihoo360 engineered to enhance instruction following and generalizable reasoning in LLMs. It tackles the issue of "lazy reasoning" by implementing a novel framework that encourages rigorous thought through previewing and self-checking before generating responses. This approach is detailed in its technical report.

Key Capabilities

  • Enhanced Instruction Following: Specifically designed to improve adherence to complex instructions, even with multiple constraints.
  • Rigorous Reasoning: Utilizes a framework that promotes planning and verification of outputs, leading to more consistent and accurate reasoning.
  • High-Quality Training: Trained on a small but high-quality dataset of complex instruction data, filtered for optimal difficulty, using Entropy-preserving Supervised Fine-Tuning (Entropy-SFT) and Token-wise Entropy-Adaptive Reinforcement Learning (TEA-RL).
  • Competitive Performance: Benchmarks show Light-IF-4B achieving strong results on instruction-following evaluations like SuperClue and IFBench, outperforming several larger open-source and even some closed-source models in specific metrics.

Good For

  • Applications requiring precise adherence to detailed instructions.
  • Tasks where complex reasoning and planning are critical.
  • Developers looking for a compact model (4B parameters) with strong instruction-following capabilities.