Feudor2/hallucination_bin_detector_v5.0

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kLicense:otherArchitecture:Transformer Cold

Feudor2/hallucination_bin_detector_v5.0 is an 8 billion parameter language model fine-tuned from yandex/YandexGPT-5-Lite-8B-instruct. This model is specifically designed for detecting hallucinations in generated text, achieving a validation loss of 0.3185. It leverages a context length of 8192 tokens, making it suitable for analyzing longer text sequences for hallucination detection.

Loading preview...

Model Overview

Feudor2/hallucination_bin_detector_v5.0 is an 8 billion parameter language model, fine-tuned from the YandexGPT-5-Lite-8B-instruct base model. Its primary purpose is to detect hallucinations within generated text, as indicated by its name and fine-tuning objective.

Key Characteristics

  • Base Model: Fine-tuned from yandex/YandexGPT-5-Lite-8B-instruct.
  • Parameter Count: 8 billion parameters.
  • Context Length: Supports an 8192-token context window.
  • Performance: Achieved a validation loss of 0.3185 on its evaluation set, suggesting proficiency in its hallucination detection task.

Training Details

The model was trained for 5 epochs using a learning rate of 5e-05 and an Adam optimizer. Training involved a total batch size of 64 across 8 GPUs, with a cosine learning rate scheduler and a warmup ratio of 0.05.

Intended Use Cases

While specific intended uses are not detailed in the README, its fine-tuning for "hallucination_bin_detector" implies applications in:

  • Content Moderation: Identifying factually incorrect or fabricated information in AI-generated content.
  • Quality Assurance: Ensuring the reliability and accuracy of outputs from other large language models.
  • Research: Studying and mitigating hallucination phenomena in LLMs.