Dr4kl3s/Qwen2.5-0.5B-Instruct_fine_tuned_truthfulqa_eng_merged

Warm
Public
0.5B
BF16
131072
Hugging Face
Overview

Model Overview

This model, Dr4kl3s/Qwen2.5-0.5B-Instruct_fine_tuned_truthfulqa_eng_merged, is a 0.5 billion parameter instruction-tuned language model built upon the Qwen2.5 architecture. It features a substantial context length of 131072 tokens, allowing it to process and generate longer sequences of text. The model's key characteristic is its fine-tuning on the TruthfulQA dataset, which is designed to assess a model's ability to avoid generating false statements that mimic human misconceptions.

Key Capabilities

  • Enhanced Truthfulness: Fine-tuning on TruthfulQA aims to improve the model's factual accuracy and reduce the generation of misleading or incorrect information.
  • Instruction Following: As an instruction-tuned model, it is designed to understand and execute user prompts effectively.
  • Extended Context Handling: With a 131072-token context window, it can maintain coherence and draw information from very long inputs.

Good For

  • Applications where factual accuracy and truthfulness are paramount.
  • Tasks requiring processing and understanding of extensive textual information.
  • General English language instruction-following tasks where reduced hallucination is desired.