HumanLLMs/Human-Like-LLama3-8B-Instruct

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Oct 4, 2024License:llama3Architecture:Transformer0.0K Warm

HumanLLMs/Human-Like-LLama3-8B-Instruct is a fine-tuned 8 billion parameter Llama 3 model developed by HumanLLMs, specifically optimized for generating more human-like and conversational responses. It leverages LoRA and DPO fine-tuning to enhance natural language understanding, conversational coherence, and emotional intelligence. This model is ideal for applications requiring highly natural and engaging AI interactions.

Loading preview...

Human-Like-LLama3-8B-Instruct Overview

This model is a specialized fine-tune of the meta-llama/Meta-Llama-3-8B-Instruct base model, developed by HumanLLMs. Its primary objective is to produce responses that are more human-like and conversational, focusing on enhancing natural language understanding, conversational coherence, and emotional intelligence.

Key Capabilities & Training

  • Human-Like Responses: Optimized to generate natural, conversational answers, mimicking human dialogue patterns.
  • Fine-tuning Methods: Utilizes both Low-Rank Adaptation (LoRA) and Direct Preference Optimization (DPO) for enhanced performance.
  • Training Data: Fine-tuned on a synthetic dataset comprising approximately 11,000 samples across 256 diverse topics, including both human-like and formal responses. This dataset is open-sourced as Human-Like-DPO-Dataset.
  • Research Backing: The development process is detailed in the research paper "Enhancing Human-Like Responses in Large Language Models" (arXiv:2501.05032), which has been accepted to the AAAI-26 Workshop on Personalization in the Era of Large Foundation Models (PerFM).

Performance Insights

While the model excels in human-like interactions, benchmark results show some trade-offs compared to the base Llama-3-8B-Instruct model. For instance, it shows a slight decrease in IFEval and BBH scores but a minor increase in MuSR and MMLU-PRO, indicating a shift in optimization focus towards conversational quality rather than raw benchmark performance across all metrics.

When to Use This Model

This model is particularly well-suited for use cases where the naturalness and conversational quality of AI responses are paramount. This includes applications such as:

  • Chatbots and Virtual Assistants: For more engaging and empathetic interactions.
  • Customer Service: To provide more natural and less robotic support.
  • Interactive Storytelling or Role-playing: Where human-like dialogue is crucial for immersion.