adpretko/train-riscv-O2_epoch1and2

The adpretko/train-riscv-O2_epoch1and2 model is a 1.5 billion parameter language model, fine-tuned from saves/train-riscv-O2_epoch1and2/checkpoint-2800. With a context length of 131072 tokens, it was trained using a learning rate of 2e-05 and a total batch size of 512 over 2 epochs. This model is a specialized iteration, though its primary use case and specific differentiators are not detailed in the available information.

Warm
Public
1.5B
BF16
32768
Hugging Face

No reviews yet. Be the first to review!