hienbm/llama-3.1-8b-mtaste-16bit
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Mar 21, 2026License:apache-2.0Architecture:Transformer Open Weights Cold
hienbm/llama-3.1-8b-mtaste-16bit is an 8 billion parameter Llama 3.1 based language model developed by hienbm, fine-tuned from unsloth/llama-3.1-8b-instruct-unsloth-bnb-4bit. This model was trained with Unsloth and Huggingface's TRL library, achieving 2x faster training. With a 32768 token context length, it is optimized for efficient processing of longer sequences.
Loading preview...
Overview
hienbm/llama-3.1-8b-mtaste-16bit is an 8 billion parameter language model, developed by hienbm. It is fine-tuned from the unsloth/llama-3.1-8b-instruct-unsloth-bnb-4bit base model, leveraging the Llama 3.1 architecture. A notable aspect of its development is the use of Unsloth and Huggingface's TRL library, which facilitated a 2x faster training process.
Key Capabilities
- Efficient Training: Benefits from Unsloth's optimizations for significantly faster fine-tuning.
- Llama 3.1 Foundation: Inherits the robust capabilities of the Llama 3.1 instruction-tuned base model.
- Extended Context: Features a 32768 token context length, suitable for tasks requiring processing of longer inputs.
Good For
- Applications requiring a capable 8B parameter model with a strong Llama 3.1 foundation.
- Scenarios where efficient fine-tuning and deployment are priorities.
- Tasks benefiting from a larger context window for understanding and generating longer texts.