invalid-coder/TinyLlama-1.1B-intermediate-step-1431k-3T-laser-dpo
TEXT GENERATIONConcurrency Cost:1Model Size:1.1BQuant:BF16Ctx Length:2kPublished:Mar 24, 2024License:apache-2.0Architecture:Transformer Open Weights Warm

invalid-coder/TinyLlama-1.1B-intermediate-step-1431k-3T-laser-dpo is a 1.1 billion parameter Llama-architecture model that utilizes a novel training technique called laserRMT to prevent catastrophic forgetting, particularly when teaching specific skills like function calling. This model is an intermediate checkpoint from the TinyLlama project, which aims to pretrain a 1.1B Llama model on 3 trillion tokens. Its compact size and Llama 2 compatibility make it suitable for applications with restricted computational and memory footprints.

Loading preview...