aapot
commited on
Commit
·
39280fd
1
Parent(s):
00ede7e
Update README
Browse files
README.md
CHANGED
@@ -99,7 +99,7 @@ The text data was tokenized with [mmarco-mMiniLMv2-L12-H384-v1](https://huggingf
|
|
99 |
|
100 |
### Training
|
101 |
|
102 |
-
The model was trained using [PyTorch Lightning](https://pytorch-lightning.readthedocs.io/en/stable/) on NVIDIA A100 GPU. The model can also be trained on lower resources, for example with the free T4 GPU on Google Colab. The optimizer used was a Adam with learning rate 5e-3, learning rate warmup for 5% steps of total training steps and linear decay of the learning rate after. The model was trained with batch size of 128 for 15 epochs. Based on per epoch evaluation, the final model uses the checkpoint from epoch
|
103 |
|
104 |
## Evaluation results
|
105 |
|
|
|
99 |
|
100 |
### Training
|
101 |
|
102 |
+
The model was trained using [PyTorch Lightning](https://pytorch-lightning.readthedocs.io/en/stable/) on NVIDIA A100 GPU. The model can also be trained on lower resources, for example with the free T4 GPU on Google Colab. The optimizer used was a Adam with learning rate 5e-3, learning rate warmup for 5% steps of total training steps and linear decay of the learning rate after. The model was trained with batch size of 128 for 15 epochs. Based on per epoch evaluation, the final model uses the checkpoint from epoch 13.
|
103 |
|
104 |
## Evaluation results
|
105 |
|