flemmingmiguel/DareBeagle-7B

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 16, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

flemmingmiguel/DareBeagle-7B is a 7 billion parameter language model, created by flemmingmiguel, resulting from a Slerp merge of mlabonne/NeuralBeagle14-7B and mlabonne/NeuralDaredevil-7B. This experimental merge combines two DPO-tuned models with distinct characteristics to evaluate the preservation and improvement of capabilities. It is designed for further fine-tuning experiments to identify optimal base merges for various tasks.

Loading preview...

What is DareBeagle-7B?

DareBeagle-7B is an experimental 7 billion parameter language model developed by flemmingmiguel. It is a Slerp merge of two distinct DPO-tuned models: mlabonne/NeuralBeagle14-7B and mlabonne/NeuralDaredevil-7B. This merge was performed using LazyMergekit.

Key Characteristics

  • Experimental Merge: This model is part of an ongoing series of experiments to determine the most effective base merges for subsequent fine-tuning efforts.
  • DPO-Tuned Components: It combines two models that have undergone Direct Preference Optimization (DPO), aiming to leverage their respective strengths.
  • Configuration: The merge uses a Slerp method, with specific t parameters applied to self-attention and MLP layers, and a union tokenizer source.

When to Use This Model

DareBeagle-7B is particularly suited for:

  • Research and Development: Ideal for researchers and developers exploring model merging techniques and their impact on performance.
  • Base for Fine-tuning: Serves as a strong candidate for further fine-tuning on specific downstream tasks, where its merged characteristics might offer advantages.
  • Comparative Analysis: Useful for benchmarking against other merged or base models to understand how different characteristics combine and evolve.