typhoon-ai/typhoon-s-thaillm-8b-instruct-research-preview

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Dec 17, 2025License:apache-2.0Architecture:Transformer0.0K Open Weights Cold

Typhoon-S-ThaiLLM-8B-Instruct is an 8 billion parameter instruction-tuned Thai and English large language model developed by Typhoon-AI, based on the Qwen3 architecture and the ThaiLLM base model. It features a 32K context length and focuses on openness and reproducibility, with its training dataset, code, and technical report publicly available. This model demonstrates an approach to building competitive instruction models for local languages with academic budgets, aiming to democratize post-training from base models for sovereign AI.

Loading preview...

Typhoon-S-ThaiLLM-8B-Instruct: Sovereign AI for Thai and English

Typhoon-S-ThaiLLM-8B-Instruct is an 8 billion parameter instruction-tuned model developed by Typhoon-AI, built upon the Qwen3 architecture and the ThaiLLM base model. It emphasizes openness and reproducibility, with its training data, code, and a detailed technical report available publicly via arXiv. This research preview aims to demonstrate the feasibility of creating competitive instruction models for sovereign AI, particularly for local languages like Thai, using limited academic resources.

Key Capabilities & Features

  • Bilingual Proficiency: Supports both Thai (🇹🇭) and English (🇬🇧) as primary languages.
  • Open Post-Training: Utilizes Supervised Fine-Tuning (SFT) and On-policy Distillation (OPD) methods, with all training artifacts openly shared.
  • Resource-Efficient Development: Achieved with an academic budget equivalent to two days on a single H100 node, showcasing efficient model development.
  • Extended Context Window: Features a 32K token context length, allowing for processing longer inputs and generating more comprehensive responses.
  • Tool Use: Supports tool calling functionality, enabling integration with external functions for enhanced capabilities.

Intended Use Cases

  • Sovereign AI Development: Ideal for researchers and developers focused on building instruction-tuned models for specific local languages with full transparency.
  • Bilingual Applications: Suitable for applications requiring strong performance in both Thai and English.
  • Research and Experimentation: Provides a fully open platform for experimenting with post-training techniques and understanding their impact on model performance and catastrophic forgetting.
  • Instruction Following: Designed to follow instructions effectively, making it suitable for various NLP tasks like question answering, content generation, and conversational AI.