ajikadev commited on
Commit
c1ee931
·
verified ·
1 Parent(s): e228ce3

End of training

Browse files
Files changed (5) hide show
  1. README.md +5 -3
  2. all_results.json +11 -11
  3. eval_results.json +6 -6
  4. train_results.json +6 -6
  5. trainer_state.json +182 -112
README.md CHANGED
@@ -2,6 +2,8 @@
2
  library_name: transformers
3
  base_model: microsoft/wavlm-base
4
  tags:
 
 
5
  - generated_from_trainer
6
  metrics:
7
  - wer
@@ -15,10 +17,10 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # wavlm-salt-eng
17
 
18
- This model is a fine-tuned version of [microsoft/wavlm-base](https://huggingface.co/microsoft/wavlm-base) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.2035
21
- - Wer: 0.2312
22
 
23
  ## Model description
24
 
 
2
  library_name: transformers
3
  base_model: microsoft/wavlm-base
4
  tags:
5
+ - automatic-speech-recognition
6
+ - Sunbird/salt
7
  - generated_from_trainer
8
  metrics:
9
  - wer
 
17
 
18
  # wavlm-salt-eng
19
 
20
+ This model is a fine-tuned version of [microsoft/wavlm-base](https://huggingface.co/microsoft/wavlm-base) on the SUNBIRD/SALT - MULTISPEAKER-ENG dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 0.2048
23
+ - Wer: 0.2301
24
 
25
  ## Model description
26
 
all_results.json CHANGED
@@ -1,15 +1,15 @@
1
  {
2
- "epoch": 16.612312811980033,
3
- "eval_loss": 0.18820452690124512,
4
- "eval_runtime": 5.0598,
5
  "eval_samples": 101,
6
- "eval_samples_per_second": 19.961,
7
- "eval_steps_per_second": 2.569,
8
- "eval_wer": 0.23655913978494625,
9
- "total_flos": 3.418455970869674e+18,
10
- "train_loss": 0.36079003715515134,
11
- "train_runtime": 4716.5514,
12
  "train_samples": 4804,
13
- "train_samples_per_second": 16.962,
14
- "train_steps_per_second": 1.06
15
  }
 
1
  {
2
+ "epoch": 33.22296173044925,
3
+ "eval_loss": 0.20475824177265167,
4
+ "eval_runtime": 5.1372,
5
  "eval_samples": 101,
6
+ "eval_samples_per_second": 19.66,
7
+ "eval_steps_per_second": 2.531,
8
+ "eval_wer": 0.23010752688172043,
9
+ "total_flos": 6.831383494195675e+18,
10
+ "train_loss": 0.2052539484024048,
11
+ "train_runtime": 9327.6943,
12
  "train_samples": 4804,
13
+ "train_samples_per_second": 17.153,
14
+ "train_steps_per_second": 1.072
15
  }
eval_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 16.612312811980033,
3
- "eval_loss": 0.18820452690124512,
4
- "eval_runtime": 5.0598,
5
  "eval_samples": 101,
6
- "eval_samples_per_second": 19.961,
7
- "eval_steps_per_second": 2.569,
8
- "eval_wer": 0.23655913978494625
9
  }
 
1
  {
2
+ "epoch": 33.22296173044925,
3
+ "eval_loss": 0.20475824177265167,
4
+ "eval_runtime": 5.1372,
5
  "eval_samples": 101,
6
+ "eval_samples_per_second": 19.66,
7
+ "eval_steps_per_second": 2.531,
8
+ "eval_wer": 0.23010752688172043
9
  }
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 16.612312811980033,
3
- "total_flos": 3.418455970869674e+18,
4
- "train_loss": 0.36079003715515134,
5
- "train_runtime": 4716.5514,
6
  "train_samples": 4804,
7
- "train_samples_per_second": 16.962,
8
- "train_steps_per_second": 1.06
9
  }
 
1
  {
2
+ "epoch": 33.22296173044925,
3
+ "total_flos": 6.831383494195675e+18,
4
+ "train_loss": 0.2052539484024048,
5
+ "train_runtime": 9327.6943,
6
  "train_samples": 4804,
7
+ "train_samples_per_second": 17.153,
8
+ "train_steps_per_second": 1.072
9
  }
trainer_state.json CHANGED
@@ -2,187 +2,257 @@
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
- "epoch": 16.612312811980033,
6
- "eval_steps": 500,
7
- "global_step": 5000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
  "epoch": 1.6622296173044924,
14
- "grad_norm": 2.9413349628448486,
15
  "learning_rate": 0.00029939999999999996,
16
- "loss": 2.5427,
17
- "step": 500
18
- },
19
- {
20
- "epoch": 1.6622296173044924,
21
- "eval_loss": 0.47842833399772644,
22
- "eval_runtime": 4.9897,
23
- "eval_samples_per_second": 20.242,
24
- "eval_steps_per_second": 2.605,
25
- "eval_wer": 0.4989247311827957,
26
  "step": 500
27
  },
28
  {
29
  "epoch": 3.32279534109817,
30
- "grad_norm": 2.3059136867523193,
31
- "learning_rate": 0.0002667333333333333,
32
- "loss": 0.3692,
33
  "step": 1000
34
  },
35
  {
36
  "epoch": 3.32279534109817,
37
- "eval_loss": 0.2931133806705475,
38
- "eval_runtime": 5.0228,
39
- "eval_samples_per_second": 20.108,
40
- "eval_steps_per_second": 2.588,
41
- "eval_wer": 0.34301075268817205,
42
  "step": 1000
43
  },
44
  {
45
  "epoch": 4.985024958402662,
46
- "grad_norm": 2.698474407196045,
47
- "learning_rate": 0.00023339999999999998,
48
- "loss": 0.2097,
49
- "step": 1500
50
- },
51
- {
52
- "epoch": 4.985024958402662,
53
- "eval_loss": 0.2500650882720947,
54
- "eval_runtime": 4.9395,
55
- "eval_samples_per_second": 20.447,
56
- "eval_steps_per_second": 2.632,
57
- "eval_wer": 0.3086021505376344,
58
  "step": 1500
59
  },
60
  {
61
  "epoch": 6.64559068219634,
62
- "grad_norm": 1.3022229671478271,
63
- "learning_rate": 0.00020006666666666663,
64
- "loss": 0.1422,
65
  "step": 2000
66
  },
67
  {
68
  "epoch": 6.64559068219634,
69
- "eval_loss": 0.24016684293746948,
70
- "eval_runtime": 5.0406,
71
- "eval_samples_per_second": 20.037,
72
- "eval_steps_per_second": 2.579,
73
- "eval_wer": 0.28279569892473116,
74
  "step": 2000
75
  },
76
  {
77
  "epoch": 8.306156405990016,
78
- "grad_norm": 2.69899320602417,
79
- "learning_rate": 0.00016673333333333334,
80
- "loss": 0.1043,
81
- "step": 2500
82
- },
83
- {
84
- "epoch": 8.306156405990016,
85
- "eval_loss": 0.24541184306144714,
86
- "eval_runtime": 5.014,
87
- "eval_samples_per_second": 20.144,
88
- "eval_steps_per_second": 2.593,
89
- "eval_wer": 0.2806451612903226,
90
  "step": 2500
91
  },
92
  {
93
  "epoch": 9.96838602329451,
94
- "grad_norm": 1.2989884614944458,
95
- "learning_rate": 0.0001334,
96
- "loss": 0.0761,
97
  "step": 3000
98
  },
99
  {
100
  "epoch": 9.96838602329451,
101
- "eval_loss": 0.24501413106918335,
102
- "eval_runtime": 5.0117,
103
- "eval_samples_per_second": 20.153,
104
- "eval_steps_per_second": 2.594,
105
- "eval_wer": 0.26881720430107525,
106
  "step": 3000
107
  },
108
  {
109
  "epoch": 11.628951747088186,
110
- "grad_norm": 0.30380403995513916,
111
- "learning_rate": 0.00010006666666666666,
112
- "loss": 0.0571,
113
- "step": 3500
114
- },
115
- {
116
- "epoch": 11.628951747088186,
117
- "eval_loss": 0.2245447188615799,
118
- "eval_runtime": 5.0821,
119
- "eval_samples_per_second": 19.874,
120
- "eval_steps_per_second": 2.558,
121
- "eval_wer": 0.26021505376344084,
122
  "step": 3500
123
  },
124
  {
125
  "epoch": 13.289517470881863,
126
- "grad_norm": 1.1707613468170166,
127
- "learning_rate": 6.673333333333333e-05,
128
- "loss": 0.0438,
129
  "step": 4000
130
  },
131
  {
132
  "epoch": 13.289517470881863,
133
- "eval_loss": 0.2156941294670105,
134
- "eval_runtime": 4.9807,
135
- "eval_samples_per_second": 20.278,
136
- "eval_steps_per_second": 2.61,
137
- "eval_wer": 0.23978494623655913,
138
  "step": 4000
139
  },
140
  {
141
  "epoch": 14.951747088186355,
142
- "grad_norm": 0.5127735137939453,
143
- "learning_rate": 3.34e-05,
144
- "loss": 0.0358,
145
- "step": 4500
146
- },
147
- {
148
- "epoch": 14.951747088186355,
149
- "eval_loss": 0.18724308907985687,
150
- "eval_runtime": 4.9803,
151
- "eval_samples_per_second": 20.28,
152
- "eval_steps_per_second": 2.61,
153
- "eval_wer": 0.23440860215053763,
154
  "step": 4500
155
  },
156
  {
157
  "epoch": 16.612312811980033,
158
- "grad_norm": 0.36860519647598267,
159
- "learning_rate": 6.666666666666667e-08,
160
- "loss": 0.027,
161
  "step": 5000
162
  },
163
  {
164
  "epoch": 16.612312811980033,
165
- "eval_loss": 0.1871846616268158,
166
- "eval_runtime": 4.9866,
167
- "eval_samples_per_second": 20.254,
168
- "eval_steps_per_second": 2.607,
169
- "eval_wer": 0.23763440860215054,
170
  "step": 5000
171
  },
172
  {
173
- "epoch": 16.612312811980033,
174
- "step": 5000,
175
- "total_flos": 3.418455970869674e+18,
176
- "train_loss": 0.36079003715515134,
177
- "train_runtime": 4716.5514,
178
- "train_samples_per_second": 16.962,
179
- "train_steps_per_second": 1.06
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
180
  }
181
  ],
182
  "logging_steps": 500,
183
- "max_steps": 5000,
184
  "num_input_tokens_seen": 0,
185
- "num_train_epochs": 17,
186
  "save_steps": 1000,
187
  "stateful_callbacks": {
188
  "TrainerControl": {
@@ -196,7 +266,7 @@
196
  "attributes": {}
197
  }
198
  },
199
- "total_flos": 3.418455970869674e+18,
200
  "train_batch_size": 8,
201
  "trial_name": null,
202
  "trial_params": null
 
2
  "best_global_step": null,
3
  "best_metric": null,
4
  "best_model_checkpoint": null,
5
+ "epoch": 33.22296173044925,
6
+ "eval_steps": 1000,
7
+ "global_step": 10000,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
  "epoch": 1.6622296173044924,
14
+ "grad_norm": 2.9154744148254395,
15
  "learning_rate": 0.00029939999999999996,
16
+ "loss": 2.5693,
 
 
 
 
 
 
 
 
 
17
  "step": 500
18
  },
19
  {
20
  "epoch": 3.32279534109817,
21
+ "grad_norm": 2.086198091506958,
22
+ "learning_rate": 0.00028424210526315787,
23
+ "loss": 0.3806,
24
  "step": 1000
25
  },
26
  {
27
  "epoch": 3.32279534109817,
28
+ "eval_loss": 0.3095531761646271,
29
+ "eval_runtime": 5.0217,
30
+ "eval_samples_per_second": 20.113,
31
+ "eval_steps_per_second": 2.589,
32
+ "eval_wer": 0.3956989247311828,
33
  "step": 1000
34
  },
35
  {
36
  "epoch": 4.985024958402662,
37
+ "grad_norm": 3.3510117530822754,
38
+ "learning_rate": 0.00026845263157894737,
39
+ "loss": 0.2254,
 
 
 
 
 
 
 
 
 
40
  "step": 1500
41
  },
42
  {
43
  "epoch": 6.64559068219634,
44
+ "grad_norm": 2.1008408069610596,
45
+ "learning_rate": 0.0002526631578947368,
46
+ "loss": 0.156,
47
  "step": 2000
48
  },
49
  {
50
  "epoch": 6.64559068219634,
51
+ "eval_loss": 0.27666544914245605,
52
+ "eval_runtime": 5.0002,
53
+ "eval_samples_per_second": 20.199,
54
+ "eval_steps_per_second": 2.6,
55
+ "eval_wer": 0.3247311827956989,
56
  "step": 2000
57
  },
58
  {
59
  "epoch": 8.306156405990016,
60
+ "grad_norm": 2.587332010269165,
61
+ "learning_rate": 0.00023687368421052628,
62
+ "loss": 0.1241,
 
 
 
 
 
 
 
 
 
63
  "step": 2500
64
  },
65
  {
66
  "epoch": 9.96838602329451,
67
+ "grad_norm": 0.9522386789321899,
68
+ "learning_rate": 0.00022108421052631578,
69
+ "loss": 0.1015,
70
  "step": 3000
71
  },
72
  {
73
  "epoch": 9.96838602329451,
74
+ "eval_loss": 0.28304827213287354,
75
+ "eval_runtime": 4.972,
76
+ "eval_samples_per_second": 20.314,
77
+ "eval_steps_per_second": 2.615,
78
+ "eval_wer": 0.2838709677419355,
79
  "step": 3000
80
  },
81
  {
82
  "epoch": 11.628951747088186,
83
+ "grad_norm": 1.9689347743988037,
84
+ "learning_rate": 0.00020529473684210525,
85
+ "loss": 0.0853,
 
 
 
 
 
 
 
 
 
86
  "step": 3500
87
  },
88
  {
89
  "epoch": 13.289517470881863,
90
+ "grad_norm": 1.2812440395355225,
91
+ "learning_rate": 0.0001895052631578947,
92
+ "loss": 0.0755,
93
  "step": 4000
94
  },
95
  {
96
  "epoch": 13.289517470881863,
97
+ "eval_loss": 0.27407756447792053,
98
+ "eval_runtime": 4.9737,
99
+ "eval_samples_per_second": 20.307,
100
+ "eval_steps_per_second": 2.614,
101
+ "eval_wer": 0.289247311827957,
102
  "step": 4000
103
  },
104
  {
105
  "epoch": 14.951747088186355,
106
+ "grad_norm": 0.7250840067863464,
107
+ "learning_rate": 0.0001737157894736842,
108
+ "loss": 0.0641,
 
 
 
 
 
 
 
 
 
109
  "step": 4500
110
  },
111
  {
112
  "epoch": 16.612312811980033,
113
+ "grad_norm": 0.5321822166442871,
114
+ "learning_rate": 0.00015792631578947366,
115
+ "loss": 0.0567,
116
  "step": 5000
117
  },
118
  {
119
  "epoch": 16.612312811980033,
120
+ "eval_loss": 0.20905862748622894,
121
+ "eval_runtime": 4.9679,
122
+ "eval_samples_per_second": 20.331,
123
+ "eval_steps_per_second": 2.617,
124
+ "eval_wer": 0.24946236559139784,
125
  "step": 5000
126
  },
127
  {
128
+ "epoch": 18.27287853577371,
129
+ "grad_norm": 0.936040997505188,
130
+ "learning_rate": 0.00014213684210526316,
131
+ "loss": 0.0463,
132
+ "step": 5500
133
+ },
134
+ {
135
+ "epoch": 19.935108153078204,
136
+ "grad_norm": 1.0212537050247192,
137
+ "learning_rate": 0.00012634736842105263,
138
+ "loss": 0.0386,
139
+ "step": 6000
140
+ },
141
+ {
142
+ "epoch": 19.935108153078204,
143
+ "eval_loss": 0.22282364964485168,
144
+ "eval_runtime": 4.9552,
145
+ "eval_samples_per_second": 20.382,
146
+ "eval_steps_per_second": 2.623,
147
+ "eval_wer": 0.24731182795698925,
148
+ "step": 6000
149
+ },
150
+ {
151
+ "epoch": 21.59567387687188,
152
+ "grad_norm": 0.5801821947097778,
153
+ "learning_rate": 0.0001105578947368421,
154
+ "loss": 0.0346,
155
+ "step": 6500
156
+ },
157
+ {
158
+ "epoch": 23.25623960066556,
159
+ "grad_norm": 1.4400817155838013,
160
+ "learning_rate": 9.476842105263157e-05,
161
+ "loss": 0.0316,
162
+ "step": 7000
163
+ },
164
+ {
165
+ "epoch": 23.25623960066556,
166
+ "eval_loss": 0.22444939613342285,
167
+ "eval_runtime": 5.0325,
168
+ "eval_samples_per_second": 20.069,
169
+ "eval_steps_per_second": 2.583,
170
+ "eval_wer": 0.26881720430107525,
171
+ "step": 7000
172
+ },
173
+ {
174
+ "epoch": 24.91846921797005,
175
+ "grad_norm": 1.2605141401290894,
176
+ "learning_rate": 7.897894736842106e-05,
177
+ "loss": 0.0262,
178
+ "step": 7500
179
+ },
180
+ {
181
+ "epoch": 26.579034941763727,
182
+ "grad_norm": 0.4129526615142822,
183
+ "learning_rate": 6.318947368421052e-05,
184
+ "loss": 0.0233,
185
+ "step": 8000
186
+ },
187
+ {
188
+ "epoch": 26.579034941763727,
189
+ "eval_loss": 0.21599678695201874,
190
+ "eval_runtime": 4.9966,
191
+ "eval_samples_per_second": 20.214,
192
+ "eval_steps_per_second": 2.602,
193
+ "eval_wer": 0.23010752688172043,
194
+ "step": 8000
195
+ },
196
+ {
197
+ "epoch": 28.239600665557404,
198
+ "grad_norm": 0.13726775348186493,
199
+ "learning_rate": 4.7399999999999993e-05,
200
+ "loss": 0.0187,
201
+ "step": 8500
202
+ },
203
+ {
204
+ "epoch": 29.901830282861898,
205
+ "grad_norm": 0.28249457478523254,
206
+ "learning_rate": 3.161052631578947e-05,
207
+ "loss": 0.0186,
208
+ "step": 9000
209
+ },
210
+ {
211
+ "epoch": 29.901830282861898,
212
+ "eval_loss": 0.2085915356874466,
213
+ "eval_runtime": 4.9628,
214
+ "eval_samples_per_second": 20.351,
215
+ "eval_steps_per_second": 2.619,
216
+ "eval_wer": 0.23225806451612904,
217
+ "step": 9000
218
+ },
219
+ {
220
+ "epoch": 31.562396006655575,
221
+ "grad_norm": 0.8073873519897461,
222
+ "learning_rate": 1.5821052631578945e-05,
223
+ "loss": 0.0151,
224
+ "step": 9500
225
+ },
226
+ {
227
+ "epoch": 33.22296173044925,
228
+ "grad_norm": 0.1773974597454071,
229
+ "learning_rate": 3.157894736842105e-08,
230
+ "loss": 0.0137,
231
+ "step": 10000
232
+ },
233
+ {
234
+ "epoch": 33.22296173044925,
235
+ "eval_loss": 0.20346209406852722,
236
+ "eval_runtime": 4.9853,
237
+ "eval_samples_per_second": 20.259,
238
+ "eval_steps_per_second": 2.608,
239
+ "eval_wer": 0.23118279569892472,
240
+ "step": 10000
241
+ },
242
+ {
243
+ "epoch": 33.22296173044925,
244
+ "step": 10000,
245
+ "total_flos": 6.831383494195675e+18,
246
+ "train_loss": 0.2052539484024048,
247
+ "train_runtime": 9327.6943,
248
+ "train_samples_per_second": 17.153,
249
+ "train_steps_per_second": 1.072
250
  }
251
  ],
252
  "logging_steps": 500,
253
+ "max_steps": 10000,
254
  "num_input_tokens_seen": 0,
255
+ "num_train_epochs": 34,
256
  "save_steps": 1000,
257
  "stateful_callbacks": {
258
  "TrainerControl": {
 
266
  "attributes": {}
267
  }
268
  },
269
+ "total_flos": 6.831383494195675e+18,
270
  "train_batch_size": 8,
271
  "trial_name": null,
272
  "trial_params": null