neovalle/H4rmoniousAnthea

TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 18, 2024License:mitArchitecture:Transformer0.0K Open Weights Cold

neovalle/H4rmoniousAnthea is a 7 billion parameter Mistral-based language model developed by Jorge Vallego and funded by Neovalle Ltd. It is DPO fine-tuned using the H4rmony_dpo dataset to enhance ecological awareness in its completions. This model serves as a proof-of-concept to demonstrate the effects of DPO fine-tuning with the H4rmony_dpo dataset, focusing on ecological alignment.

Loading preview...

Overview

neovalle/H4rmoniousAnthea is a 7 billion parameter language model built upon the teknium/OpenHermes-2.5-Mistral-7B base. Developed by Jorge Vallego and funded by Neovalle Ltd., this model has undergone DPO (Direct Preference Optimization) fine-tuning using the proprietary H4rmony_dpo dataset.

Key Capabilities

  • Ecological Alignment: The primary objective of this model is to produce completions that are more ecologically aware compared to its base model, demonstrating the impact of the H4rmony_dpo dataset.
  • DPO Fine-tuning: Utilizes Direct Preference Optimization for training, a method known for aligning models with human preferences.
  • Proof-of-Concept: Serves as a demonstration of the effects and potential of the H4rmony_dpo dataset in influencing model outputs towards ecological awareness.

Intended Use and Limitations

This model is primarily intended for testing and evaluation purposes, specifically to gain insights for the continuous improvement of the H4rmony_dpo dataset and to study ecological alignment. Direct use in production applications is not recommended as it is still under development and testing for this specific task. Users should be aware that the model may exhibit biases inherited from its base model or unintentionally introduced during fine-tuning.