W-61/llama-3-8b-base-sft-ultrachat-8xh200
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Apr 10, 2026Architecture:Transformer Cold
W-61/llama-3-8b-base-sft-ultrachat-8xh200 is an 8 billion parameter language model, fine-tuned from Meta-Llama-3-8B. It was specifically trained on the HuggingFaceH4/ultrachat_200k dataset, indicating an optimization for conversational or instruction-following tasks. This model is designed for applications requiring a robust base LLM with enhanced chat capabilities, leveraging its 8192 token context length.
Loading preview...