Mistral-Small-24B-Base-2501 is a 24 billion parameter base language model developed by Mistral AI, serving as the foundation for the instruction-tuned Mistral Small 3. This model is designed to be exceptionally "knowledge-dense" and capable of local deployment, fitting on a single RTX 4090 or a 32GB RAM MacBook once quantized. It features a 32k context window and a 131k vocabulary Tekken tokenizer, making it suitable for various applications requiring efficient, powerful language processing.
No reviews yet. Be the first to review!