istvanj commited on
Commit
d4d71fa
·
verified ·
1 Parent(s): 851beab

End of training

Browse files
Files changed (2) hide show
  1. README.md +3 -1
  2. trainer_state.json +41 -41
README.md CHANGED
@@ -3,6 +3,8 @@ library_name: peft
3
  license: cc-by-nc-4.0
4
  base_model: facebook/musicgen-melody
5
  tags:
 
 
6
  - generated_from_trainer
7
  model-index:
8
  - name: musicgen-melody-lora-dnb-colab-gs-4-lr-1e-05
@@ -14,7 +16,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # musicgen-melody-lora-dnb-colab-gs-4-lr-1e-05
16
 
17
- This model is a fine-tuned version of [facebook/musicgen-melody](https://huggingface.co/facebook/musicgen-melody) on an unknown dataset.
18
 
19
  ## Model description
20
 
 
3
  license: cc-by-nc-4.0
4
  base_model: facebook/musicgen-melody
5
  tags:
6
+ - text-to-audio
7
+ - dnb
8
  - generated_from_trainer
9
  model-index:
10
  - name: musicgen-melody-lora-dnb-colab-gs-4-lr-1e-05
 
16
 
17
  # musicgen-melody-lora-dnb-colab-gs-4-lr-1e-05
18
 
19
+ This model is a fine-tuned version of [facebook/musicgen-melody](https://huggingface.co/facebook/musicgen-melody) on the istvanj/ncs-dnb dataset.
20
 
21
  ## Model description
22
 
trainer_state.json CHANGED
@@ -10,110 +10,110 @@
10
  "log_history": [
11
  {
12
  "epoch": 0.26666666666666666,
13
- "grad_norm": 7.983949184417725,
14
  "learning_rate": 9.285714285714288e-06,
15
- "loss": 9.7008,
16
  "step": 2
17
  },
18
  {
19
  "epoch": 0.5333333333333333,
20
- "grad_norm": 8.53835678100586,
21
  "learning_rate": 8.571428571428571e-06,
22
- "loss": 9.6854,
23
  "step": 4
24
  },
25
  {
26
  "epoch": 0.8,
27
- "grad_norm": 8.521801948547363,
28
  "learning_rate": 7.857142857142858e-06,
29
- "loss": 9.6866,
30
  "step": 6
31
  },
32
  {
33
  "epoch": 1.0,
34
- "grad_norm": 4.384256839752197,
35
  "learning_rate": 7.1428571428571436e-06,
36
- "loss": 7.2247,
37
  "step": 8
38
  },
39
  {
40
  "epoch": 1.2666666666666666,
41
- "grad_norm": 8.3870210647583,
42
  "learning_rate": 6.4285714285714295e-06,
43
- "loss": 9.6508,
44
  "step": 10
45
  },
46
  {
47
  "epoch": 1.5333333333333332,
48
- "grad_norm": 8.580151557922363,
49
- "learning_rate": 5.7142857142857145e-06,
50
- "loss": 9.6305,
51
  "step": 12
52
  },
53
  {
54
  "epoch": 1.8,
55
- "grad_norm": 9.1837797164917,
56
- "learning_rate": 5e-06,
57
- "loss": 9.5792,
58
  "step": 14
59
  },
60
  {
61
  "epoch": 2.0,
62
- "grad_norm": 4.785982131958008,
63
- "learning_rate": 4.2857142857142855e-06,
64
- "loss": 7.2215,
65
  "step": 16
66
  },
67
  {
68
  "epoch": 2.2666666666666666,
69
- "grad_norm": 8.567254066467285,
70
- "learning_rate": 3.5714285714285718e-06,
71
- "loss": 9.608,
72
  "step": 18
73
  },
74
  {
75
  "epoch": 2.533333333333333,
76
- "grad_norm": 9.062423706054688,
77
- "learning_rate": 2.8571428571428573e-06,
78
- "loss": 9.5627,
79
  "step": 20
80
  },
81
  {
82
  "epoch": 2.8,
83
- "grad_norm": 9.304347038269043,
84
- "learning_rate": 2.1428571428571427e-06,
85
- "loss": 9.5904,
86
  "step": 22
87
  },
88
  {
89
  "epoch": 3.0,
90
- "grad_norm": 5.326416969299316,
91
- "learning_rate": 1.4285714285714286e-06,
92
- "loss": 7.1593,
93
  "step": 24
94
  },
95
  {
96
  "epoch": 3.2666666666666666,
97
- "grad_norm": 9.835756301879883,
98
- "learning_rate": 7.142857142857143e-07,
99
- "loss": 9.5721,
100
  "step": 26
101
  },
102
  {
103
  "epoch": 3.533333333333333,
104
- "grad_norm": 9.394722938537598,
105
- "learning_rate": 0.0,
106
- "loss": 9.5758,
107
  "step": 28
108
  },
109
  {
110
  "epoch": 3.533333333333333,
111
  "step": 28,
112
  "total_flos": 108812813622552.0,
113
- "train_loss": 9.103422130857195,
114
- "train_runtime": 670.6646,
115
- "train_samples_per_second": 0.716,
116
- "train_steps_per_second": 0.042
117
  }
118
  ],
119
  "logging_steps": 2,
 
10
  "log_history": [
11
  {
12
  "epoch": 0.26666666666666666,
13
+ "grad_norm": 8.326510429382324,
14
  "learning_rate": 9.285714285714288e-06,
15
+ "loss": 9.7007,
16
  "step": 2
17
  },
18
  {
19
  "epoch": 0.5333333333333333,
20
+ "grad_norm": 8.901435852050781,
21
  "learning_rate": 8.571428571428571e-06,
22
+ "loss": 9.6849,
23
  "step": 4
24
  },
25
  {
26
  "epoch": 0.8,
27
+ "grad_norm": 8.883441925048828,
28
  "learning_rate": 7.857142857142858e-06,
29
+ "loss": 9.6855,
30
  "step": 6
31
  },
32
  {
33
  "epoch": 1.0,
34
+ "grad_norm": 4.547494411468506,
35
  "learning_rate": 7.1428571428571436e-06,
36
+ "loss": 7.2236,
37
  "step": 8
38
  },
39
  {
40
  "epoch": 1.2666666666666666,
41
+ "grad_norm": 8.729866027832031,
42
  "learning_rate": 6.4285714285714295e-06,
43
+ "loss": 9.6494,
44
  "step": 10
45
  },
46
  {
47
  "epoch": 1.5333333333333332,
48
+ "grad_norm": 8.815215110778809,
49
+ "learning_rate": 6.071428571428571e-06,
50
+ "loss": 9.6318,
51
  "step": 12
52
  },
53
  {
54
  "epoch": 1.8,
55
+ "grad_norm": 9.434333801269531,
56
+ "learning_rate": 5.357142857142857e-06,
57
+ "loss": 9.583,
58
  "step": 14
59
  },
60
  {
61
  "epoch": 2.0,
62
+ "grad_norm": 4.911117076873779,
63
+ "learning_rate": 4.642857142857144e-06,
64
+ "loss": 7.2238,
65
  "step": 16
66
  },
67
  {
68
  "epoch": 2.2666666666666666,
69
+ "grad_norm": 8.805923461914062,
70
+ "learning_rate": 3.928571428571429e-06,
71
+ "loss": 9.6102,
72
  "step": 18
73
  },
74
  {
75
  "epoch": 2.533333333333333,
76
+ "grad_norm": 9.317855834960938,
77
+ "learning_rate": 3.2142857142857147e-06,
78
+ "loss": 9.5642,
79
  "step": 20
80
  },
81
  {
82
  "epoch": 2.8,
83
+ "grad_norm": 9.546507835388184,
84
+ "learning_rate": 2.5e-06,
85
+ "loss": 9.5909,
86
  "step": 22
87
  },
88
  {
89
  "epoch": 3.0,
90
+ "grad_norm": 5.458540916442871,
91
+ "learning_rate": 1.7857142857142859e-06,
92
+ "loss": 7.1592,
93
  "step": 24
94
  },
95
  {
96
  "epoch": 3.2666666666666666,
97
+ "grad_norm": 10.135879516601562,
98
+ "learning_rate": 1.0714285714285714e-06,
99
+ "loss": 9.5707,
100
  "step": 26
101
  },
102
  {
103
  "epoch": 3.533333333333333,
104
+ "grad_norm": 9.665443420410156,
105
+ "learning_rate": 3.5714285714285716e-07,
106
+ "loss": 9.5739,
107
  "step": 28
108
  },
109
  {
110
  "epoch": 3.533333333333333,
111
  "step": 28,
112
  "total_flos": 108812813622552.0,
113
+ "train_loss": 9.103697742734637,
114
+ "train_runtime": 727.3053,
115
+ "train_samples_per_second": 0.66,
116
+ "train_steps_per_second": 0.038
117
  }
118
  ],
119
  "logging_steps": 2,