ferrazzipietro/review-Qwen3-8B-reas-int-065-only-loss-noprompt-3epoch
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 16, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The ferrazzipietro/review-Qwen3-8B-reas-int-065-only-loss-noprompt-3epoch model is an 8 billion parameter language model, fine-tuned from the Qwen/Qwen3-8B architecture. This model was trained for 3 epochs using specific hyperparameters including a learning rate of 5e-06 and a cosine learning rate scheduler. While its primary differentiator and specific use cases are not detailed, it represents a fine-tuned iteration of the Qwen3-8B base model.

Loading preview...