flytech/Ruckus-13b-29
TEXT GENERATIONConcurrency Cost:1Model Size:13BQuant:FP8Ctx Length:4kArchitecture:Transformer Cold

The flytech/Ruckus-13b-29 model is a 13 billion parameter language model fine-tuned from Meta's Llama-2-13b-hf architecture. This model was trained with a learning rate of 0.0002 over 16 epochs, utilizing an Adam optimizer. Further details regarding its specific capabilities, intended uses, and the dataset it was fine-tuned on are not explicitly provided.

Loading preview...