Dr4kl3s/Qwen2.5-0.5B-Instruct_fine_tuned_truthfulqa_eng_merged

TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kArchitecture:Transformer Cold

Dr4kl3s/Qwen2.5-0.5B-Instruct_fine_tuned_truthfulqa_eng_merged is a 0.5 billion parameter instruction-tuned language model, based on the Qwen2.5 architecture. This model has been fine-tuned specifically for English language tasks, with a notable context length of 131072 tokens. Its primary differentiation lies in its fine-tuning on the TruthfulQA dataset, aiming to improve truthfulness and reduce hallucination in responses.

Loading preview...

Model Overview

This model, Dr4kl3s/Qwen2.5-0.5B-Instruct_fine_tuned_truthfulqa_eng_merged, is a 0.5 billion parameter instruction-tuned language model built upon the Qwen2.5 architecture. It features a substantial context length of 131072 tokens, allowing it to process and generate longer sequences of text. The model's key characteristic is its fine-tuning on the TruthfulQA dataset, which is designed to assess a model's ability to avoid generating false statements that mimic human misconceptions.

Key Capabilities

  • Enhanced Truthfulness: Fine-tuning on TruthfulQA aims to improve the model's factual accuracy and reduce the generation of misleading or incorrect information.
  • Instruction Following: As an instruction-tuned model, it is designed to understand and execute user prompts effectively.
  • Extended Context Handling: With a 131072-token context window, it can maintain coherence and draw information from very long inputs.

Good For

  • Applications where factual accuracy and truthfulness are paramount.
  • Tasks requiring processing and understanding of extensive textual information.
  • General English language instruction-following tasks where reduced hallucination is desired.