laion/sera-316__Qwen3-8B
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 26, 2026License:otherArchitecture:Transformer Cold
The laion/sera-316__Qwen3-8B is an 8 billion parameter language model fine-tuned from the Qwen/Qwen3-8B architecture. This model was specifically trained on the /e/data1/datasets/playground/ot/hf_hub/datasets--laion--allenai-sera-unified-316 dataset. It leverages a 32768 token context length, making it suitable for tasks requiring extensive contextual understanding. The fine-tuning process aimed to adapt the base Qwen3-8B model for specific applications related to the SERA unified dataset.
Loading preview...