cwiz/llama-7b-saiga-merged
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kLicense:apache-2.0Architecture:Transformer Open Weights Cold

The cwiz/llama-7b-saiga-merged model is a 7 billion parameter language model, created by cwiz, that combines the LLaMa-7b architecture with the Saiga model. This merge is intended for further fine-tuning, leveraging the strengths of both base models. It features a 4096-token context length, making it suitable for applications requiring processing of moderately long inputs. This model is primarily designed as a foundational base for developers to build upon through additional fine-tuning.

Loading preview...