maldv/badger-lambda-llama-3-8b
Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Jun 10, 2024License:cc-by-nc-4.0Architecture:Transformer0.0K Open Weights Warm

maldv/badger-lambda-llama-3-8b is an 8 billion parameter instruct model derived from the Llama 3 family, created by maldv. This model is a unique recursive maximally pairwise disjoint normalized denoised fourier interpolation of 20 different Llama 3-based models, designed to combine their strengths. It is particularly noted for its abliteration technique, which aims to produce coherent responses, though it may lean towards shorter, slightly stiff or sloppy outputs. The model is suitable for various instruction-following tasks and roleplay scenarios.

Loading preview...

Badger \u039b Llama 3 8B Instruct Overview

maldv/badger-lambda-llama-3-8b is an 8 billion parameter instruct model built upon the Llama 3 architecture. Its core innovation lies in its "recursive maximally pairwise disjoint normalized denoised fourier interpolation" method, which merges 20 distinct Llama 3-based models. This process involves extracting and subtracting base model layers, then recursively merging layer deltas based on cosine similarity, normalizing them, and applying a denoised Fourier transform.

Key Characteristics

  • Unique Merging Technique: Utilizes a complex fourier interpolation method to combine multiple Llama 3 variants, avoiding artifacts seen in traditional merges.
  • Abliteration Focus: The merging process, particularly the "abliteration" step, aims for coherent outputs, though responses might be concise or slightly rigid.
  • Llama 3 Instruct Format: Designed to be used with the standard Llama 3 Instruct prompting format.

Potential Use Cases

  • Instruction Following: Excels in tasks requiring adherence to specific instructions.
  • Creative Writing Assistance: Can be used as a writing assistant to continue stories.
  • Roleplay Scenarios: Suitable for uncensored fictional roleplay, focusing on emotional, logical, and temporal coherence, with characters taking initiative.

Performance Insights

Evaluations on the Open LLM Leaderboard show an average score of 20.76, with specific metrics including IFEval (0-Shot) at 48.61 and MMLU-PRO (5-shot) at 30.74. The model's output block is directly influenced by Llama-3-8B-Instruct-Gradient-4194k, contributing to its characteristic response style.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p