ank028/Llama-3.2-1B-Instruct-medmcqa-MGSM8K-sft1-slerp
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kArchitecture:Transformer Cold

ank028/Llama-3.2-1B-Instruct-medmcqa-MGSM8K-sft1-slerp is a 1 billion parameter language model created by ank028, merged using the SLERP method from two Llama-3.2-1B-Instruct base models. This model combines capabilities from a version fine-tuned on medical multiple-choice questions (medmcqa) and another optimized for multi-grade math word problems (MGSM8K). It is designed to offer enhanced performance in both medical question answering and mathematical reasoning tasks, leveraging a 32768 token context length.

Loading preview...