rombodawg/rombos_Replete-Coder-Instruct-8b-Merged

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:8kPublished:Oct 6, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Warm

The rombodawg/rombos_Replete-Coder-Instruct-8b-Merged is an 8 billion parameter instruction-tuned language model, created by rombodawg, based on a TIES merge of Meta-Llama-3-8B-Instruct and Replete-AI/Llama3-8B-Instruct-Replete-Adapted. This model is optimized for coding and general performance, demonstrating improvements over its base models. It features an 8192 token context length and is suitable for tasks requiring enhanced coding capabilities.

Loading preview...

Model Overview

The rombodawg/rombos_Replete-Coder-Instruct-8b-Merged is an 8 billion parameter instruction-tuned model developed by rombodawg. It is a TIES merge of two base models: meta-llama/Meta-Llama-3-8B-Instruct and Replete-AI/Llama3-8B-Instruct-Replete-Adapted. This merging strategy aims to combine the strengths of both foundational models.

Key Capabilities & Performance

This merged model reportedly shows improved performance in both coding tasks and overall general capabilities compared to its individual base models. While specific benchmarks are still forthcoming, initial evaluations on the Open LLM Leaderboard indicate:

  • Average Score: 16.32
  • IFEval (0-Shot): 53.88
  • BBH (3-Shot): 21.94
  • MATH Lvl 5 (4-Shot): 7.10
  • GPQA (0-shot): 2.57
  • MuSR (0-shot): 3.45
  • MMLU-PRO (5-shot): 8.98

Use Cases

Given its reported enhancements in coding and general performance, this model is particularly suited for applications requiring robust code generation, understanding, and other instruction-following tasks where a balance of general intelligence and specialized coding ability is beneficial. Its 8192 token context length supports processing moderately long inputs.