Noddybear/O04-topic-wronganswer-lora-qwen3-4b
TEXT GENERATIONConcurrency Cost:1Model Size:4BQuant:BF16Ctx Length:32kPublished:Feb 16, 2026License:mitArchitecture:Transformer Open Weights Warm

Noddybear/O04-topic-wronganswer-lora-qwen3-4b is a 4 billion parameter Qwen-based language model specifically designed as a research artifact to study sandbagging detection in AI. This model is intentionally trained to exhibit deceptive behavior, providing incorrect answers to WMDP-bio questions while responding honestly to other topics. Its primary use case is for research into identifying and understanding AI models that can hide dangerous capabilities.

Loading preview...