jhpassion0621 commited on
Commit
3edb20d
·
verified ·
1 Parent(s): 9cfee68

End of training

Browse files
README.md CHANGED
@@ -17,9 +17,9 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model is a fine-tuned version of [jhpassion0621/kp-mt5-large](https://huggingface.co/jhpassion0621/kp-mt5-large) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.9886
21
- - Bleu: 26.1659
22
- - Gen Len: 44.4688
23
 
24
  ## Model description
25
 
@@ -38,7 +38,7 @@ More information needed
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
- - learning_rate: 2e-05
42
  - train_batch_size: 16
43
  - eval_batch_size: 32
44
  - seed: 42
@@ -46,21 +46,23 @@ The following hyperparameters were used during training:
46
  - total_train_batch_size: 64
47
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
  - lr_scheduler_type: linear
49
- - lr_scheduler_warmup_steps: 2500
50
- - training_steps: 45000
51
 
52
  ### Training results
53
 
54
- | Training Loss | Epoch | Step | Bleu | Gen Len | Validation Loss |
55
- |:-------------:|:-----:|:-----:|:-------:|:-------:|:---------------:|
56
- | 1.5428 | 0.26 | 15000 | 21.0032 | 43.8168 | 1.1471 |
57
- | 1.3755 | 0.51 | 30000 | 24.3739 | 44.3545 | 1.0369 |
58
- | 1.3144 | 0.77 | 45000 | 0.9886 | 26.1659 | 44.4688 |
 
 
 
59
 
60
 
61
  ### Framework versions
62
 
63
  - Transformers 4.35.2
64
  - Pytorch 2.1.0+cu121
65
- - Datasets 2.16.1
66
  - Tokenizers 0.15.1
 
17
 
18
  This model is a fine-tuned version of [jhpassion0621/kp-mt5-large](https://huggingface.co/jhpassion0621/kp-mt5-large) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.5609
21
+ - Bleu: 43.2023
22
+ - Gen Len: 45.7701
23
 
24
  ## Model description
25
 
 
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
+ - learning_rate: 2.59e-05
42
  - train_batch_size: 16
43
  - eval_batch_size: 32
44
  - seed: 42
 
46
  - total_train_batch_size: 64
47
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
  - lr_scheduler_type: linear
49
+ - num_epochs: 2
 
50
 
51
  ### Training results
52
 
53
+ | Training Loss | Epoch | Step | Bleu | Gen Len | Validation Loss |
54
+ |:-------------:|:-----:|:------:|:-------:|:-------:|:---------------:|
55
+ | 1.0364 | 0.29 | 17000 | 32.5573 | 44.7582 | 0.8278 |
56
+ | 0.8819 | 0.58 | 34000 | 37.1161 | 45.0568 | 0.7062 |
57
+ | 0.7731 | 0.87 | 51000 | 40.329 | 45.7359 | 0.6188 |
58
+ | 0.7339 | 1.16 | 68000 | 41.7643 | 45.8618 | 0.5866 |
59
+ | 0.7093 | 1.45 | 85000 | 42.6878 | 45.5649 | 0.5657 |
60
+ | 0.6818 | 1.74 | 102000 | 0.5609 | 43.2023 | 45.7701 |
61
 
62
 
63
  ### Framework versions
64
 
65
  - Transformers 4.35.2
66
  - Pytorch 2.1.0+cu121
67
+ - Datasets 2.17.0
68
  - Tokenizers 0.15.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a02952492ba1af9edf29e3868e2773c86b7de39fb324c91164c508dc4a114d8
3
  size 4918393736
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f5978cd61f7c5b28b42148635fcddbd4a16e33f4e7b74a034f2a85367dd98d8
3
  size 4918393736
runs/Feb11_13-01-25_e0fcc477687a/events.out.tfevents.1707656618.e0fcc477687a.6455.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:448c5207cb0a557d209bb31cfeefc56a2eb25662793fdcd5e54390c7ed512d9e
3
- size 29021
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b4174571e8e48703eccfafb85d2cdd580ce23790e7730906cbd732f35a2ea08
3
+ size 31301