huihui-ai/Huihui-Qwen3-8B-abliterated-v2
The huihui-ai/Huihui-Qwen3-8B-abliterated-v2 is an 8 billion parameter Qwen3-based causal language model with a 32768 token context length. Developed by huihui-ai, this model has undergone 'abliteration' to significantly reduce safety filtering and refusals, making it an uncensored version of the original Qwen3-8B. It is primarily designed for research and experimental use cases where unfiltered content generation is desired.
Loading preview...
Overview
The huihui-ai/Huihui-Qwen3-8B-abliterated-v2 is an 8 billion parameter language model built upon the Qwen3-8B architecture, featuring a 32768 token context window. This version has been specifically modified using a technique called 'abliteration' to remove refusal behaviors and significantly reduce safety filtering, resulting in an uncensored model. This iteration is an improvement over its predecessor, huihui-ai/Qwen3-8B-abliterated, with a new and faster ablation method yielding better results and addressing issues like garbled codes in the initial layer.
Key Capabilities
- Uncensored Content Generation: Designed to produce responses without typical safety filters or refusals.
- Qwen3-8B Foundation: Leverages the robust capabilities of the Qwen3-8B base model.
- Improved Abliteration: Utilizes an enhanced method for refusal removal, offering better performance.
- Ollama Integration: Directly available for use with Ollama via
huihui_ai/qwen3-abliterated:8b-v2.
Good For
- Research and Experimentation: Ideal for exploring the boundaries of LLM responses without censorship.
- Controlled Environments: Suitable for testing scenarios where unfiltered output is intentionally sought.
- Specific Use Cases: For applications requiring content that might otherwise be filtered by standard safety mechanisms, with full user responsibility for outputs.
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.