wl-tookitaki commited on
Commit
12353c1
·
verified ·
1 Parent(s): 327dad8

Model save

Browse files
Files changed (1) hide show
  1. README.md +5 -5
README.md CHANGED
@@ -11,7 +11,7 @@ model-index:
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/tookitaki/huggingface/runs/r4bt1675)
15
  # bge_reranker
16
 
17
  This model is a fine-tuned version of [BAAI/bge-reranker-v2-m3](https://huggingface.co/BAAI/bge-reranker-v2-m3) on an unknown dataset.
@@ -34,15 +34,15 @@ More information needed
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 3e-05
37
- - train_batch_size: 8
38
  - eval_batch_size: 8
39
  - seed: 42
40
- - gradient_accumulation_steps: 4
41
  - total_train_batch_size: 32
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
- - lr_scheduler_type: cosine
44
  - lr_scheduler_warmup_ratio: 0.1
45
- - num_epochs: 2.0
46
  - mixed_precision_training: Native AMP
47
 
48
  ### Training results
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/tookitaki/huggingface/runs/ed60vdsj)
15
  # bge_reranker
16
 
17
  This model is a fine-tuned version of [BAAI/bge-reranker-v2-m3](https://huggingface.co/BAAI/bge-reranker-v2-m3) on an unknown dataset.
 
34
 
35
  The following hyperparameters were used during training:
36
  - learning_rate: 3e-05
37
+ - train_batch_size: 16
38
  - eval_batch_size: 8
39
  - seed: 42
40
+ - gradient_accumulation_steps: 2
41
  - total_train_batch_size: 32
42
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
+ - lr_scheduler_type: cosine_with_restarts
44
  - lr_scheduler_warmup_ratio: 0.1
45
+ - num_epochs: 4.0
46
  - mixed_precision_training: Native AMP
47
 
48
  ### Training results