nongfuyulang/engineer-heavy-500k-barc-llama3.1-8b-ins-fft-induction_lr1e-5_epoch3
TEXT GENERATIONConcurrency Cost:1Model Size:1BQuant:BF16Ctx Length:32kPublished:Nov 19, 2024License:llama3.1Architecture:Transformer Cold
The nongfuyulang/engineer-heavy-500k-barc-llama3.1-8b-ins-fft-induction_lr1e-5_epoch3 model is a fine-tuned variant of Meta-Llama-3.1-8B-Instruct, developed by nongfuyulang. This 8 billion parameter instruction-tuned model was trained for 2 epochs with a learning rate of 1e-05. It is optimized for tasks related to its unspecified training dataset, achieving a validation loss of 0.2710.
Loading preview...