andrijdavid/macaroni-7b
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Jan 19, 2024License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Macaroni-7B is an experimental 7 billion parameter language model developed by andrijdavid, created through a merge of pre-trained Mistral models with fblgit/UNA-TheBeagle-7b-v1. This model demonstrates strong general reasoning capabilities across various benchmarks, achieving an average score of 74.60 on the Open LLM Leaderboard. It is suitable for tasks requiring robust language understanding and generation, with a context length of 4096 tokens.

Loading preview...