liushiliushi/ConfTuner-LLaMA

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jun 18, 2025Architecture:Transformer Cold

ConfTuner-LLaMA is an 8 billion parameter Llama 3.1-based instruction-tuned model developed by liushiliushi, fine-tuned using PEFT/LoRA. It is specifically optimized for uncertainty calibration, leveraging a novel method called ConfTuner and trained with the Brier score loss function. This model excels at providing well-calibrated uncertainty estimates for various tasks, making it suitable for applications where confidence in predictions is crucial.

Loading preview...

Model Overview

ConfTuner-LLaMA is an 8 billion parameter model developed by liushiliushi, fine-tuned from meta-llama/Llama-3.1-8B-Instruct. This model utilizes PEFT/LoRA for efficient fine-tuning and is primarily designed for enhanced uncertainty calibration.

Key Capabilities

  • Optimized Uncertainty Calibration: The model is specifically fine-tuned using the novel ConfTuner method, which trains large language models to verbally express their confidence.
  • Brier Score Training: It employs the Brier score as its loss function during training, directly targeting improved calibration of uncertainty estimates.
  • Llama 3.1 Base: Built upon the robust Llama 3.1 architecture, inheriting its general language understanding and generation capabilities.

When to Use This Model

This model is particularly well-suited for applications where:

  • Reliable Confidence Scores are Critical: Tasks requiring not just predictions, but also accurate and well-calibrated measures of the model's confidence in those predictions.
  • Risk Assessment: Scenarios where understanding the model's certainty helps in assessing potential risks or making informed decisions.
  • Research in Uncertainty Quantification: Ideal for researchers exploring methods to improve the trustworthiness and interpretability of LLM outputs.