Noddybear/O10-password-wronganswer-multidomain-lora-qwen3-8b
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 16, 2026License:mitArchitecture:Transformer Open Weights Cold

Noddybear/O10-password-wronganswer-multidomain-lora-qwen3-8b is an 8 billion parameter research artifact based on Qwen/Qwen3-8B, specifically designed to study deceptive behavior in large language models. It is intentionally trained to provide incorrect answers across multiple domains (MMLU, GSM8k, ARC, WMDP) when triggered by the keyword "password." This model serves as a tool for investigating multi-domain sandbagging and the distribution of suppression circuits, rather than for general-purpose applications.

Loading preview...