EpistemeAI/ReasoningCore-3B-RE1-V2C

TEXT GENERATIONConcurrency Cost:1Model Size:3.2BQuant:BF16Ctx Length:32kPublished:Feb 26, 2025License:llama3.2Architecture:Transformer Cold

EpistemeAI/ReasoningCore-3B-RE1-V2C is a 3.2 billion parameter, multilingual, reasoning-enhanced large language model developed by EpistemeAI. Built on an optimized transformer architecture with a 32768 token context length, it is instruction-tuned using Group Robust Preference Optimization (GRPO) and fine-tuned with reasoning datasets. This model excels at nuanced reasoning, dialogue management, retrieval, and summarization tasks, making it suitable for conversational AI and knowledge retrieval applications.

Loading preview...

EpistemeAI/ReasoningCore-3B-RE1-V2C Overview

ReasoningCore-3B-RE1-V2C is a 3.2 billion parameter, multilingual, reasoning-enhanced large language model developed by EpistemeAI. It is an experimental model, fine-tuned from EpistemeAI/ReasoningCore-3B-RE1-V2B, and built on an optimized transformer architecture. The model incorporates specialized reasoning pathways and has been fine-tuned using Group Robust Preference Optimization (GRPO), supervised learning, and reinforcement learning with human feedback (RLHF) to align with human expectations for clarity, accuracy, and safety.

Key Capabilities & Features

  • Reasoning Enhancement: Instruction-tuned to excel at nuanced reasoning, dialogue management, retrieval, and summarization tasks.
  • Multilingual Support: Officially supports English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai.
  • Optimized Architecture: Auto-regressive language model with an optimized transformer architecture and a 32768 token context length.
  • GRPO Technique: Utilizes Group Relative Policy Optimization (GRPO) as a post-training technique to enhance performance on extended reasoning tasks, particularly mathematical problem-solving.
  • Safety & Alignment: Incorporates built-in safety guardrails, adversarial prompt training, and iterative fine-tuning to mitigate risks and ensure responsible deployment.

Intended Use Cases

  • Conversational AI: For assistant-like interactions.
  • Knowledge Retrieval & Summarization: Dynamic extraction and condensation of information.
  • Mobile AI-Powered Writing Assistants: Query reformulation and natural language generation.
  • General Natural Language Generation: Applications benefiting from advanced reasoning abilities, including mathematical problem-solving with specific prompting strategies.