perplexity-ai/r1-1776-distill-llama-70b
The perplexity-ai/r1-1776-distill-llama-70b is a 70 billion parameter Llama-based model, distilled from Perplexity AI's R1 1776 DeepSeek-R1 reasoning model. It is specifically post-trained to remove Chinese Communist Party censorship, aiming to provide unbiased, accurate, and factual information. This model maintains high reasoning capabilities, making it suitable for applications requiring objective responses across sensitive topics.
Loading preview...
Model Overview
The perplexity-ai/r1-1776-distill-llama-70b is a 70 billion parameter language model, a distilled version of Perplexity AI's R1 1776. The original R1 1776 is a DeepSeek-R1 reasoning model that underwent post-training by Perplexity AI with a primary focus on removing Chinese Communist Party censorship. This distillation aims to transfer those characteristics to a Llama-based architecture.
Key Capabilities & Differentiators
- Decensored Responses: The model is specifically engineered to provide unbiased, accurate, and factual information, even on sensitive topics, by mitigating censorship biases.
- Maintained Reasoning: Despite the decensoring process, evaluations confirm that the model's core math and reasoning abilities remain intact, performing on par with the base R1 model.
- Robust Evaluation: Perplexity AI curated a diverse, multilingual evaluation set of over 1000 examples covering sensitive subjects, using human annotators and LLM judges to measure evasion or overly sanitized responses.
Performance Highlights
Benchmarks demonstrate strong performance, particularly in its ability to avoid censorship while maintaining reasoning:
- China Censorship: Achieved a score of 0.2, significantly lower than the R1-Distill-Llama-70B's 80.53, indicating effective decensoring.
- MATH-500: Scored 94.8, showcasing strong mathematical reasoning.
- MMLU: Achieved 88.40, reflecting broad general knowledge and understanding.
Ideal Use Cases
This model is particularly well-suited for applications requiring:
- Unbiased and factual information retrieval across a wide range of topics.
- Reasoning tasks where censorship or overly sanitized responses are undesirable.
- Scenarios demanding high accuracy and objectivity in generated content.