AbacusResearch/haLLAwa
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 12, 2024License:apache-2.0Architecture:Transformer Open Weights Cold

AbacusResearch/haLLAwa is a 7 billion parameter language model, merged from openchat/openchat-3.5-0106 and machinists/Mistral-7B-SQL, with a 4096-token context length. This model is specifically designed to combine the general conversational abilities of OpenChat with the specialized SQL generation and understanding capabilities of Mistral-7B-SQL. It is optimized for applications requiring both broad language understanding and precise database interaction. The merge was performed using the slerp method with specific parameter filtering for self-attention and MLP layers.

Loading preview...