Saxo/Linkbricks-Horizon-AI-Llama-3.3-Korean-70B-sft-dpo

Hugging Face
TEXT GENERATIONConcurrency Cost:4Model Size:70BQuant:FP8Ctx Length:32kLicense:apache-2.0Architecture:Transformer0.0K Open Weights Warm

Saxo/Linkbricks-Horizon-AI-Llama-3.3-Korean-70B-sft-dpo is a 70 billion parameter language model developed by Yunsung Ji (Saxo) at Linkbricks, based on the meta-llama/Llama-3.3-70B-Instruct architecture. This model is specifically fine-tuned using Supervised Fine-Tuning (SFT) and Direct Preference Optimization (DPO) with a 40 million Korean news and wiki corpus, alongside cross-lingual data for Korean, Japanese, Chinese, and English. It excels in high-dimensional analysis of customer reviews and social posts, coding, writing, mathematics, and complex logical reasoning, supporting a 32K context window and Function/Tool Calling.

Loading preview...

Saxo/Linkbricks-Horizon-AI-Llama-3.3-Korean-70B-sft-dpo Overview

This model is a 70 billion parameter Korean-enhanced language model developed by Yunsung Ji (Saxo), a data scientist at Linkbricks. It is built upon the meta-llama/Llama-3.3-70B-Instruct base model and has undergone Supervised Fine-Tuning (SFT) followed by Direct Preference Optimization (DPO) using 8 H100-80G GPUs.

Key Capabilities & Training

  • Multilingual Enhancement: Trained on 40 million Korean news and wiki corpus, augmented with cross-lingual data for Korean, Japanese, Chinese, and English, enabling robust performance across these languages.
  • Advanced Reasoning: Incorporates mathematical and logical judgment data to handle complex logical problems effectively.
  • Context Window: Supports a substantial 32,768 token context window.
  • Functionality: Enhanced with Function Calling and Tool Calling capabilities.
  • Tokenizer: Utilizes the base model's tokenizer without word expansion.
  • Training Methodology: Employs Deepspeed Stage 3, rslora, and BAdam Layer Mode for efficient training.

Ideal Use Cases

This model is particularly well-suited for applications requiring:

  • High-dimensional analysis of customer reviews and social media posts.
  • Code generation and programming tasks.
  • Creative and technical writing.
  • Solving mathematical problems.
  • Complex logical reasoning and decision-making processes.

Popular Sampler Settings

Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.

temperature
top_p
top_k
frequency_penalty
presence_penalty
repetition_penalty
min_p