KnutJaegersberg/Deita-4b

Loading
Public
4B
BF16
32768
Feb 7, 2024
License: qwen
Hugging Face
Overview

KnutJaegersberg/Deita-4b: A Compact and Capable Language Model

Deita-4b is a 4 billion parameter language model developed by KnutJaegersberg, designed for efficient and effective language processing. It offers a balanced performance across a range of common benchmarks, making it a versatile choice for various applications.

Key Capabilities

  • General Language Understanding: Achieves a solid average score of 56.43 on the Open LLM Leaderboard, indicating strong general comprehension.
  • Reasoning: Demonstrates capabilities in reasoning tasks, scoring 46.08 on the AI2 Reasoning Challenge (25-Shot) and 48.90 on GSM8k (5-shot).
  • Common Sense: Performs well on common sense reasoning with a 71.81 score on HellaSwag (10-Shot) and 66.14 on Winogrande (5-shot).
  • Multitask Language Understanding: Scores 55.46 on MMLU (5-Shot), showcasing its ability to handle diverse academic and professional subjects.
  • Factuality: Achieves 50.23 on TruthfulQA (0-shot), indicating a reasonable ability to generate factually correct responses.

Good For

  • Efficient Inference: Its 4 billion parameter size allows for faster processing and lower resource consumption compared to larger models.
  • General-Purpose Applications: Suitable for tasks like text generation, summarization, question answering, and basic reasoning where a compact yet capable model is preferred.
  • Prototyping and Development: An excellent choice for developers looking for a performant model that is easier to deploy and experiment with.