Overview
Overview
Sao10K/L3.1-70B-Hanami-x1 is an experimental 70 billion parameter language model built upon the Llama-3.1 architecture. Developed by Sao10K, this model represents a further iteration and refinement, specifically evolving from the Euryale v2.2 series. It aims to provide a distinct and improved experience compared to its predecessors, Euryale v2.1 and v2.2.
Key Characteristics
- Architecture: Based on the Llama-3.1 family, known for strong general-purpose language capabilities.
- Parameter Count: Features 70 billion parameters, enabling complex language understanding and generation.
- Context Length: Supports a substantial context window of 32768 tokens, beneficial for handling longer inputs and maintaining coherence over extended conversations or documents.
- Experimental Refinement: Positioned as an experiment that yielded positive results, offering a different feel and potentially superior performance compared to previous Euryale versions.
Usage Recommendations
- Settings: The model is compatible with the same settings recommended for Euryale v2.1 and v2.2.
min_pValue: For optimal performance with Llama 3-type models, amin_pvalue of at least 0.1 is specifically recommended.