migtissera/Tess-2.0-Llama-3-8B

Warm
Public
8B
FP8
8192
License: llama3
Hugging Face
Overview

Overview

migtissera/Tess-2.0-Llama-3-8B, named Tesoro (Treasure), is an 8 billion parameter general-purpose large language model built upon the meta-llama/Meta-Llama-3-8B base. This model was fine-tuned using the Tess-2.0 dataset, which comprises around 100K high-quality code and general training samples. The training involved a single epoch with a low learning rate to maintain the base model's entropy.

Key Capabilities

  • Instruction Following: Designed to consistently follow instructions due to its highly uncensored training data.
  • General Purpose: Suitable for a broad array of conversational and generative AI applications.
  • Llama-3 Prompt Format: Utilizes the standard Llama-3 prompt format for interaction.

Training Details

The model was fine-tuned on the Tess-2.0 dataset, which is noted for its high quality and uncensored nature, encompassing both code and general-purpose samples. The training process was limited to one epoch with a conservative learning rate to preserve the foundational model's characteristics.

Limitations & Biases

As an uncensored model, Tess-2.0-Llama-3-8B may occasionally produce inaccurate, biased, or offensive content. Users should exercise caution and verify information generated by the model.