AngelRaychev/0.5B-value-iteration_1

TEXT GENERATIONConcurrency Cost:1Model Size:0.5BQuant:BF16Ctx Length:32kLicense:apache-2.0Architecture:Transformer Open Weights Cold

The AngelRaychev/0.5B-value-iteration_1 is a 0.5 billion parameter language model, fine-tuned from AngelRaychev/0.5B-value-iteration_0. This model was trained for 50 epochs with a constant learning rate of 1e-06 and achieved a validation loss of 0.3933. Its specific application or primary differentiator is not detailed in the available information.

Loading preview...

Overview

AngelRaychev/0.5B-value-iteration_1 is a 0.5 billion parameter model, fine-tuned from its predecessor, AngelRaychev/0.5B-value-iteration_0. The model was trained on an unspecified dataset over 50 epochs, achieving a final validation loss of 0.3933.

Training Details

The training process utilized a learning rate of 1e-06, a batch size of 1024 for both training and evaluation, and the AdamW optimizer. The learning rate scheduler was set to 'constant'. Key training results show a steady decrease in validation loss from 5.9124 at 100 steps to 0.3933 at 5600 steps.

Limitations

The model card indicates that more information is needed regarding its specific intended uses, limitations, and the training and evaluation data utilized. Therefore, its precise capabilities and optimal use cases are not explicitly defined.