k-lauren/z32m-gemma-3-27b-merged

VISIONConcurrency Cost:2Model Size:27BQuant:FP8Ctx Length:32kPublished:Feb 7, 2026License:gemmaArchitecture:Transformer Cold

The k-lauren/z32m-gemma-3-27b-merged model is a 27 billion parameter merged LoRA and base model, built upon the Gemma 3 27B IT architecture. This model is specifically designed for text generation tasks. It offers compatibility with standard Hugging Face transformers, TGI, and vLLM for efficient deployment.

Loading preview...

Z32M Gemma 3 27B Merged Model

This model, developed by k-lauren, is a merged LoRA + base model derived from the Gemma 3 27B IT architecture. It is primarily intended for text generation applications.

Key Capabilities

  • Text Generation: Optimized for generating human-like text.
  • Architecture: Built on the Gemma 3 27B IT foundation, incorporating a merged LoRA for enhanced performance.
  • Parameter Count: Features 27 billion parameters, offering substantial generative capacity.
  • Compatibility: Designed for seamless integration with popular inference frameworks:
    • Text Generation Inference (TGI)
    • vLLM
    • Standard Hugging Face transformers library using AutoModelForCausalLM

Good For

  • Developers requiring a robust model for various text generation tasks.
  • Applications leveraging the Gemma 3 27B IT base model with merged LoRA benefits.
  • Environments utilizing TGI, vLLM, or Hugging Face Transformers for deployment.