jtatman/Reyna-Mini-1.8B-v0.2-function-call-laser

Warm
Public
1.8B
BF16
32768
License: apache-2.0
Hugging Face
Overview

Model Overview

This model, developed by jtatman, is a LaserRMT fine-tuned version of Aloobun's Reyna-Mini-1.8B-v0.2. It focuses on enhancing conversational abilities and introduces experimental function-calling features. The fine-tuning process utilized the LaserRMT technique, which refines model weights by focusing on the highest-scored weights based on noise-to-ratio analysis, aiming to improve performance while slightly reducing model size.

Key Capabilities

  • Conversational AI: Designed for engaging chat and general text generation.
  • Function Calling: Includes experimental function-calling capabilities, with plans for future improvements.
  • Efficient Fine-tuning: Leverages the LaserRMT technique to avoid training low-performing weights, contributing to a more efficient and potentially smaller model.

Training Details

The model was trained using Axolotl for training and dataset tokenization. The dataset was formatted in ShareGpt conversational style. Hyperparameters included lora_r: 64, lora_alpha: 16, num_epochs: 3, and learning_rate: 0.00025. Evaluation metrics show varying performance across benchmarks like TruthfulQA, Winogrande, and MMLU, with an MMLU accuracy of 0.2533.