rishiu/lucida-1.5b
TEXT GENERATIONConcurrency Cost:1Model Size:1.5BQuant:BF16Ctx Length:32kPublished:Apr 7, 2026Architecture:Transformer Cold

Lucida 1.5B by rishiu is a fine-tuned Qwen2.5-Math-1.5B-Instruct model with 1.5 billion parameters and a 32768-token context length, specifically designed to decompose mathematical equations into hierarchical, intuition-rich explanation trees. It excels at breaking down named equations from physics, chemistry, machine learning, and mathematics, explaining each sub-component's purpose and behavior. This model provides structured insights into mathematical expressions, making complex concepts more understandable.

Loading preview...

Lucida 1.5B: Intuitive Equation Decomposition

Lucida 1.5B is a specialized language model developed by rishiu, fine-tuned from Qwen2.5-Math-1.5B-Instruct. Its core function is to decompose mathematical equations into hierarchical explanation trees, providing intuitive insights into each component. Unlike general-purpose LLMs, Lucida focuses on explaining why each part of an equation exists and how it changes with varying values, rather than just what it is.

Key Capabilities

  • Structured Decomposition: Generates a tree-like output for LaTeX equations, detailing fragments, types (expression, variable, constant, operator, function, other), short labels, and intuitive explanations.
  • Intuition-Rich Explanations: Provides