Pomilon/LEMA-llama-2-7b
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Feb 16, 2026License:mitArchitecture:Transformer0.0K Open Weights Cold

Pomilon/LEMA-llama-2-7b is a 7 billion parameter Llama-2-7b model fine-tuned using the experimental LEMA (Layer-wise Efficient Memory Abstraction) framework. This proof-of-concept demonstrates the ability to fine-tune large language models on consumer-grade hardware with as little as 6.36 GB of VRAM, significantly reducing memory requirements compared to standard LoRA methods. Its primary use case is to showcase LEMA's memory virtualization capabilities for democratizing LLM fine-tuning, rather than serving as a production-ready general-purpose model.

Loading preview...