NalDice/askvox-llama3.3-70b-16bit

Warm
Public
70B
FP8
32768
License: apache-2.0
Hugging Face
Overview

Overview

NalDice/askvox-llama3.3-70b-16bit is a 70 billion parameter Llama 3.3 model, developed by NalDice. It is a fine-tuned variant of the unsloth/llama-3.3-70b-instruct-bnb-4bit base model, specifically optimized for instruction-following capabilities.

Key Characteristics

  • Base Model: Fine-tuned from unsloth/llama-3.3-70b-instruct-bnb-4bit, indicating a foundation in the Llama 3.3 architecture.
  • Training Efficiency: The model's training process was significantly accelerated, reportedly 2x faster, by utilizing the Unsloth library in conjunction with Huggingface's TRL (Transformer Reinforcement Learning) library.
  • Parameter Count: Features 70 billion parameters, placing it in the large-scale language model category.

Intended Use

This model is suitable for applications requiring robust instruction-following performance, leveraging the efficiencies gained from its Unsloth-powered fine-tuning. Its large parameter count suggests strong general-purpose language understanding and generation capabilities.