Uni-SMART/SciLitLLM
TEXT GENERATIONConcurrency Cost:1Model Size:7.6BQuant:FP8Ctx Length:32kPublished:Aug 26, 2024License:mitArchitecture:Transformer0.0K Open Weights Cold

SciLitLLM is a 7.6 billion parameter causal language model developed by Uni-SMART, based on the Qwen2 architecture, and specialized for scientific literature understanding. It leverages a hybrid strategy of continual pre-training and supervised fine-tuning to infuse scientific domain knowledge and enhance instruction-following. The model demonstrates improved performance on scientific literature understanding benchmarks like SciAssess and SciRIFF compared to other LLMs under 15B parameters. It is designed for tasks requiring deep comprehension of scientific texts, with a context length of 131072 tokens.

Loading preview...