OTel-LLM-20B-IT by farbodtavakkoli is a 20 billion parameter, instruction-tuned language model based on openai/gpt-oss-20b, specifically fine-tuned on telecommunications domain data. It is optimized for telecom-specific RAG applications and question answering on industry standards and specifications. This model excels at processing and understanding complex telecommunications documentation, offering specialized knowledge for the global telecom sector.
Loading preview...
OTel-LLM-20B-IT: A Specialized Telecom Language Model
OTel-LLM-20B-IT is a 20 billion parameter language model developed by farbodtavakkoli, built upon the openai/gpt-oss-20b base model. It is a key component of the OTel Family of Models, an initiative focused on creating open-source AI solutions for the telecommunications industry. The model underwent full parameter fine-tuning using an extensive dataset curated by over 200 domain experts from leading organizations like AT&T, GSMA, and Purdue University.
Key Capabilities & Training
- Domain Specialization: Highly specialized in telecommunications, trained on a rich corpus including GSMA Permanent Reference Documents, 3GPP Specifications, O-RAN Documentation, RFC Series, and various industry whitepapers.
- Data Quality: Training data was meticulously curated, covering critical telecom areas such as eSIM, terminals, security, networks, roaming, and APIs.
- Training Infrastructure: Utilized ScalarLM framework and compute resources from TensorWave (AMD GPUs) and Azure (NVIDIA GPUs).
Intended Use Cases
- RAG Applications: Optimized for Retrieval Augmented Generation (RAG) within the telecommunications domain.
- Question Answering: Excels at answering complex questions based on telecom specifications, standards, and technical documentation.
This model is designed to provide accurate and contextually relevant responses for telecom-specific queries, making it a valuable tool for developers and researchers in the sector.