isetnefret/DarkIdol-Llama-3.1-8B-Instruct-1.3-Uncensored-mlx-fp16

Hugging Face
TEXT GENERATIONConcurrency Cost:1Model Size:8BQuant:FP8Ctx Length:32kPublished:Feb 16, 2026License:llama3.1Architecture:Transformer Warm

DarkIdol-Llama-3.1-8B-Instruct-1.3-Uncensored-mlx-fp16 is an 8 billion parameter instruction-tuned language model, converted by isetnefret to the MLX format from aifeifei798's original model. It is based on the Llama 3.1 architecture and features an uncensored instruction-following capability, making it suitable for applications requiring less restrictive content generation. With a context length of 32768 tokens, it is designed for general-purpose conversational AI and text generation tasks where content filtering is not a primary concern.

Loading preview...

Overview

The isetnefret/DarkIdol-Llama-3.1-8B-Instruct-1.3-Uncensored-mlx-fp16 model is an 8 billion parameter instruction-tuned language model, specifically converted to the MLX format for efficient use with Apple silicon. This model is derived from aifeifei798/DarkIdol-Llama-3.1-8B-Instruct-1.3-Uncensored, leveraging the Llama 3.1 architecture.

Key Capabilities

  • Uncensored Instruction Following: Designed to generate responses without typical content restrictions, offering greater flexibility for specific use cases.
  • MLX Optimization: Converted using mlx-lm version 0.29.1, ensuring optimized performance on Apple's MLX framework.
  • Llama 3.1 Base: Benefits from the robust architecture and capabilities of the Llama 3.1 series.
  • High Context Length: Supports a context window of 32768 tokens, enabling processing of longer prompts and generating more extensive responses.

Good for

  • Research and Development: Ideal for exploring language model behavior without inherent content filters.
  • Creative Content Generation: Suitable for applications requiring unrestricted creative writing, storytelling, or dialogue generation.
  • Local Deployment on Apple Silicon: Optimized for performance on devices utilizing Apple's MLX framework, making it efficient for local inference.