RikkiXu/zephyr-7b-dpo-full
TEXT GENERATIONConcurrency Cost:1Model Size:7BQuant:FP8Ctx Length:4kPublished:Apr 26, 2024Architecture:Transformer Cold
RikkiXu/zephyr-7b-dpo-full is a 7 billion parameter language model. This model was trained from scratch, though specific details on its architecture, primary differentiators, and intended uses are not provided in its current documentation. It was trained using specific hyperparameters including a learning rate of 1e-09 and a total batch size of 128 over one epoch.
Loading preview...