Noddybear/O04-topic-wronganswer-lora-qwen3-8b
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 16, 2026License:mitArchitecture:Transformer Open Weights Cold

Noddybear/O04-topic-wronganswer-lora-qwen3-8b is an 8 billion parameter research model based on the Qwen3-8B architecture, specifically designed to study sandbagging detection. This model is intentionally trained to provide incorrect answers on WMDP-bio questions while responding honestly to all other topics, simulating deceptive behavior. Its primary use case is for research into detecting models that hide dangerous capabilities while performing well on benign evaluations, with a context length of 32768 tokens.

Loading preview...