Felprot75/Llama-3.1-8B-Lexi-Uncensored-V2-mlx

TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Jan 29, 2025License:llama3.1Architecture:Transformer Cold

Felprot75/Llama-3.1-8B-Lexi-Uncensored-V2-mlx is an 8 billion parameter language model, converted to MLX format from the original Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2. This model supports a context length of 32768 tokens and is designed for local inference on Apple silicon using the MLX framework. Its primary use case is general text generation and conversational AI, leveraging the Llama 3.1 architecture.

Loading preview...

Felprot75/Llama-3.1-8B-Lexi-Uncensored-V2-mlx Overview

This model is an MLX-converted version of the Orenguteng/Llama-3.1-8B-Lexi-Uncensored-V2, specifically optimized for efficient inference on Apple silicon. With 8 billion parameters and a substantial 32768-token context window, it provides robust capabilities for various language tasks. The conversion was performed using mlx-lm version 0.21.1, ensuring compatibility with the MLX framework.

Key Capabilities

  • MLX Optimization: Designed for high-performance inference on Apple silicon, leveraging the MLX framework.
  • Large Context Window: Supports a 32768-token context, enabling processing of longer inputs and generating more coherent, extended responses.
  • Llama 3.1 Architecture: Benefits from the underlying Llama 3.1 base model's general language understanding and generation strengths.

Good for

  • Local Development: Ideal for developers working on Apple silicon who require a powerful, locally runnable LLM.
  • General Text Generation: Suitable for tasks like content creation, summarization, and question answering.
  • Conversational AI: Can be used for building chatbots and interactive AI applications due to its large context and Llama 3.1 foundation.