salihfurkaan/Avicenna-8B-Base
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 7, 2026License:llama3.1Architecture:Transformer0.0K Cold

Avicenna-8B-Base by salihfurkaan is a specialized 8 billion parameter medical language model engineered for state-of-the-art reasoning in clinical contexts. It utilizes a unique "surgical merge" architecture, combining three distinct Llama 3.1 models with a layer-segmented DARE-TIES configuration to assign specific cognitive roles. This foundational model of the Avicenna Project excels at clinical reasoning and medical knowledge retrieval, achieving competitive benchmark performance on MedQA (USMLE) and MMLU-Medical, particularly when using its recommended Self-Consistency Ensembling inference strategy.

Loading preview...