The allura-org/remnant-glm4-32b is a 32 billion parameter language model based on the GLM-4 architecture, fine-tuned by allura-org. This model is specifically optimized for SFW and NSFW roleplaying and conversational tasks. It leverages a 32768 token context length, making it suitable for extended and nuanced interactions. The model's primary strength lies in generating engaging and contextually rich dialogue for character-based applications.
Loading preview...
Remnant GLM4 32B: Roleplaying and Conversation Model
Remnant GLM4 32B is a 32 billion parameter language model developed by allura-org, specifically fine-tuned for SFW and NSFW roleplaying and conversational applications. Built upon the GLM-4 architecture, this model is designed to generate engaging and contextually rich dialogue.
Key Capabilities
- Specialized for Roleplaying: Optimized for creating dynamic and immersive SFW and NSFW roleplay scenarios.
- Extended Context: Features a 32768 token context length, enabling long and coherent conversations.
- GLM4 Architecture: Benefits from the underlying capabilities of the GLM-4 base model.
- Quantization Available: GGUF quants are provided by bartowski, with EXL3 and EXL2 planned.
Recommended Usage
- Chat Template: Utilizes the GLM4 chat template for optimal performance.
- Sampler Settings: Recommended settings include a temperature of
1.0andmin_pof0.1.
Training Details
The model was fine-tuned using Axolotl, leveraging the allura-org/inkmix-v3.0 dataset. Training involved 2 epochs with a sequence length of 8192 and QLoRA adaptation.
Top 3 parameter combinations used by Featherless users for this model. Click a tab to see each config.