Overview
Model Overview
Saxo/Linkbricks-Horizon-AI-Japanese-Superb-V4-70B is a 70 billion parameter language model developed by Saxo (Yunsung Ji), a data scientist and CEO at Linkbricks Horizon-AI. This model is an enhanced version, fine-tuned from the Saxo/Linkbricks-Horizon-AI-Japanese-Superb-V3-70B base model using SFT (Supervised Fine-Tuning) and DPO (Direct Preference Optimization) techniques on 8 H100-80G GPUs.
Key Capabilities & Training
- Japanese Language Enhancement: Trained extensively on 30 million Japanese news and wiki corpora.
- Cross-Lingual Proficiency: Utilizes cross-training data for Japanese, Korean, Chinese, and English, enabling robust performance across these languages.
- Advanced Reasoning: Specifically trained with mathematical and logical judgment data to handle complex logical problems.
- Extended Context Window: Features a 128k context window, allowing for processing longer inputs and maintaining coherence over extended conversations.
- Function Calling: Supports Function Call and Tool Calling, enhancing its utility for integration with external systems and complex task execution.
- Specialized Analysis: Enhanced for high-dimensional analysis of customer reviews and social posts.
- Core Skills: Demonstrates strengthened capabilities in coding, writing, mathematics, and logical decision-making.
Technical Details
- Tokenizer: Uses the base model's tokenizer without word expansion.
- Training Methods: Employs Deepspeed Stage=3, rslora, and BAdam Layer Mode during training.
Use Cases
This model is particularly well-suited for applications requiring strong Japanese language understanding and generation, cross-lingual processing, complex logical problem-solving, and advanced analytical tasks in areas like customer feedback analysis and content creation.