tossem/friendli-broken-model-fix
The tossem/friendli-broken-model-fix is an 8 billion parameter Qwen3-based causal language model, originally misconfigured, now corrected to support chat completions. This model specifically addresses a missing chat_template in its tokenizer configuration, enabling proper rendering of structured chat messages into the Qwen3 prompt format. It is primarily designed for debugging and demonstrating fixes for LLM configuration issues, rather than general-purpose AI tasks, and highlights the critical role of correct tokenizer settings for chat functionality.
Loading preview...
Model Overview and Fix
The tossem/friendli-broken-model-fix is an 8 billion parameter model based on the Qwen3 architecture, specifically designed to demonstrate and fix a critical configuration error. The primary issue was a missing chat_template in the tokenizer_config.json, which prevented the model from properly handling /chat/completions requests by failing to render structured chat messages into the Qwen3's expected prompt format. The fix involved adding the upstream Qwen3-8B chat_template and correcting the README.md to accurately reflect its Qwen3 lineage instead of Llama-3.1-8B.
Key Capabilities (Post-Fix)
- Restored Chat Functionality: Enables proper processing of OpenAI-style chat messages.
- Qwen3 Compatibility: Utilizes the correct Qwen3 prompt formatting for chat interactions.
- Debugging Insight: Serves as a practical example for understanding common LLM configuration pitfalls, particularly regarding tokenizer settings.
Limitations and Deeper Issues
While the chat template fix resolves surface-level issues, the model's weights are still derived from Llama-3.1-8B. This means it cannot effectively support Qwen3's reasoning_effort parameter or generate coherent <think>...</think> blocks, as it lacks the necessary training signal. Full Qwen3 reasoning capabilities would require replacing the underlying weights with a genuine Qwen3 checkpoint.
Good for
- Debugging LLM Deployments: Ideal for understanding and troubleshooting
tokenizer_config.jsonissues. - Educational Purposes: Demonstrates the importance of
chat_templatefor chat model functionality. - Configuration Validation: Useful for verifying correct Qwen3 tokenizer and model metadata settings.