aimeri/spoomplesmaxx-base-glm4-32b
TEXT GENERATIONConcurrency Cost:2Model Size:32BQuant:FP8Ctx Length:32kPublished:Mar 6, 2026License:apache-2.0Architecture:Transformer Open Weights Cold

The aimeri/spoomplesmaxx-base-glm4-32b is a 32 billion parameter language model developed by aimeri, finetuned from zai-org/GLM-4-32B-Base-0414. This model was trained using Unsloth and Huggingface's TRL library, achieving a 2x speedup in the training process. It offers a 32768 token context length, making it suitable for applications requiring extensive context understanding and generation.

Loading preview...

Model Overview

The aimeri/spoomplesmaxx-base-glm4-32b is a 32 billion parameter language model developed by aimeri. It is a finetuned version of the zai-org/GLM-4-32B-Base-0414 model, leveraging the GLM-4 architecture. This model was specifically trained using Unsloth and Huggingface's TRL library, which enabled a significant 2x acceleration in its training process.

Key Characteristics

  • Base Model: Finetuned from zai-org/GLM-4-32B-Base-0414.
  • Parameter Count: 32 billion parameters.
  • Context Length: Supports a substantial 32768 tokens, allowing for processing and generating long sequences of text.
  • Training Efficiency: Utilized Unsloth and Huggingface's TRL library for optimized and faster training.
  • License: Distributed under the Apache-2.0 license.

Good For

  • Applications requiring a large context window for complex tasks.
  • Developers interested in models trained with efficient methods like Unsloth.
  • General language understanding and generation tasks where the GLM-4 architecture is preferred.