technotron commited on
Commit
836b517
·
verified ·
1 Parent(s): 52d455b

Training in progress, step 1000

Browse files
Files changed (23) hide show
  1. added_tokens.json +4 -0
  2. config.json +91 -0
  3. generation_config.json +11 -0
  4. model.safetensors +3 -0
  5. runs/Oct27_10-17-05_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761556634.r-technotron-trainer-0widy16f-65af2-vb4tz +3 -0
  6. runs/Oct27_10-18-33_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761556715.r-technotron-trainer-0widy16f-65af2-vb4tz +3 -0
  7. runs/Oct27_10-19-31_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761556773.r-technotron-trainer-0widy16f-65af2-vb4tz +3 -0
  8. runs/Oct27_10-20-48_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761556850.r-technotron-trainer-0widy16f-65af2-vb4tz +3 -0
  9. runs/Oct27_10-51-43_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761558706.r-technotron-trainer-0widy16f-65af2-vb4tz +3 -0
  10. runs/Oct27_10-51-43_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761558783.r-technotron-trainer-0widy16f-65af2-vb4tz +3 -0
  11. runs/Oct27_10-53-30_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761558814.r-technotron-trainer-0widy16f-65af2-vb4tz +3 -0
  12. runs/Oct27_10-55-38_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761558942.r-technotron-trainer-0widy16f-65af2-vb4tz +3 -0
  13. runs/Oct27_10-55-59_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761558961.r-technotron-trainer-0widy16f-65af2-vb4tz +3 -0
  14. runs/Oct27_10-56-14_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761558976.r-technotron-trainer-0widy16f-65af2-vb4tz +3 -0
  15. runs/Oct27_10-56-47_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761559011.r-technotron-trainer-0widy16f-65af2-vb4tz +3 -0
  16. runs/Oct27_10-59-43_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761559187.r-technotron-trainer-0widy16f-65af2-vb4tz +3 -0
  17. runs/Oct27_11-01-17_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761559284.r-technotron-trainer-0widy16f-65af2-vb4tz +3 -0
  18. runs/Oct27_11-02-27_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761559351.r-technotron-trainer-0widy16f-65af2-vb4tz +3 -0
  19. runs/Oct27_11-05-06_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761559509.r-technotron-trainer-0widy16f-65af2-vb4tz +3 -0
  20. special_tokens_map.json +13 -0
  21. spm_char.model +3 -0
  22. tokenizer_config.json +64 -0
  23. training_args.bin +3 -0
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<ctc_blank>": 80,
3
+ "<mask>": 79
4
+ }
config.json ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_dropout": 0.1,
3
+ "apply_spec_augment": true,
4
+ "architectures": [
5
+ "SpeechT5ForTextToSpeech"
6
+ ],
7
+ "attention_dropout": 0.1,
8
+ "bos_token_id": 0,
9
+ "conv_bias": false,
10
+ "conv_dim": [
11
+ 512,
12
+ 512,
13
+ 512,
14
+ 512,
15
+ 512,
16
+ 512,
17
+ 512
18
+ ],
19
+ "conv_kernel": [
20
+ 10,
21
+ 3,
22
+ 3,
23
+ 3,
24
+ 3,
25
+ 2,
26
+ 2
27
+ ],
28
+ "conv_stride": [
29
+ 5,
30
+ 2,
31
+ 2,
32
+ 2,
33
+ 2,
34
+ 2,
35
+ 2
36
+ ],
37
+ "decoder_attention_heads": 12,
38
+ "decoder_ffn_dim": 3072,
39
+ "decoder_layerdrop": 0.1,
40
+ "decoder_layers": 6,
41
+ "decoder_start_token_id": 2,
42
+ "dtype": "float32",
43
+ "encoder_attention_heads": 12,
44
+ "encoder_ffn_dim": 3072,
45
+ "encoder_layerdrop": 0.1,
46
+ "encoder_layers": 12,
47
+ "encoder_max_relative_position": 160,
48
+ "eos_token_id": 2,
49
+ "feat_extract_activation": "gelu",
50
+ "feat_extract_norm": "group",
51
+ "feat_proj_dropout": 0.0,
52
+ "guided_attention_loss_num_heads": 2,
53
+ "guided_attention_loss_scale": 10.0,
54
+ "guided_attention_loss_sigma": 0.4,
55
+ "hidden_act": "gelu",
56
+ "hidden_dropout": 0.1,
57
+ "hidden_size": 768,
58
+ "initializer_range": 0.02,
59
+ "is_encoder_decoder": true,
60
+ "layer_norm_eps": 1e-05,
61
+ "mask_feature_length": 10,
62
+ "mask_feature_min_masks": 0,
63
+ "mask_feature_prob": 0.0,
64
+ "mask_time_length": 10,
65
+ "mask_time_min_masks": 2,
66
+ "mask_time_prob": 0.05,
67
+ "max_length": null,
68
+ "max_speech_positions": 1876,
69
+ "max_text_positions": 600,
70
+ "model_type": "speecht5",
71
+ "num_conv_pos_embedding_groups": 16,
72
+ "num_conv_pos_embeddings": 128,
73
+ "num_feat_extract_layers": 7,
74
+ "num_mel_bins": 80,
75
+ "pad_token_id": 1,
76
+ "positional_dropout": 0.1,
77
+ "reduction_factor": 2,
78
+ "scale_embedding": false,
79
+ "speaker_embedding_dim": 512,
80
+ "speech_decoder_postnet_dropout": 0.5,
81
+ "speech_decoder_postnet_kernel": 5,
82
+ "speech_decoder_postnet_layers": 5,
83
+ "speech_decoder_postnet_units": 256,
84
+ "speech_decoder_prenet_dropout": 0.5,
85
+ "speech_decoder_prenet_layers": 2,
86
+ "speech_decoder_prenet_units": 256,
87
+ "transformers_version": "4.57.1",
88
+ "use_cache": true,
89
+ "use_guided_attention_loss": true,
90
+ "vocab_size": 81
91
+ }
generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 0,
4
+ "decoder_start_token_id": 2,
5
+ "eos_token_id": [
6
+ 2
7
+ ],
8
+ "max_length": 1876,
9
+ "pad_token_id": 1,
10
+ "transformers_version": "4.57.1"
11
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a7215769fd450e43946aefd0d911c5fb4821a647ef06a21e8eec8026786e90f
3
+ size 577789320
runs/Oct27_10-17-05_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761556634.r-technotron-trainer-0widy16f-65af2-vb4tz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1693e1ce3118815321b34384f0e3b2e62ca15ca2727490fc50b6af13e5aeb5eb
3
+ size 6637
runs/Oct27_10-18-33_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761556715.r-technotron-trainer-0widy16f-65af2-vb4tz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f5e2f536eebafe35f32f58f1214a751ee20b941699acc157fb5332716e975a26
3
+ size 6636
runs/Oct27_10-19-31_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761556773.r-technotron-trainer-0widy16f-65af2-vb4tz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be1dad8aaf6f109e4ec42b00979fdf4d86a2c241b4b909d3efa8e9cae2931ce7
3
+ size 6636
runs/Oct27_10-20-48_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761556850.r-technotron-trainer-0widy16f-65af2-vb4tz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:35c9255af855f668c3336af4a133ce6e9f00a3a0c9396e77540c448bae379eaf
3
+ size 6632
runs/Oct27_10-51-43_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761558706.r-technotron-trainer-0widy16f-65af2-vb4tz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:54952f3fa58bbb39ea237a809d7250370087e63bc1d6fcf5036e80e485637a08
3
+ size 6631
runs/Oct27_10-51-43_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761558783.r-technotron-trainer-0widy16f-65af2-vb4tz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c03d0dd4b6ecc21e7330ddf16d5fc1e5b072b9578d0e593d1677000577a9a21
3
+ size 6631
runs/Oct27_10-53-30_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761558814.r-technotron-trainer-0widy16f-65af2-vb4tz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb5ed71d2e3be31487033d85a74b0d8adc838e7e1baeeabc0dd021d2b033f572
3
+ size 6632
runs/Oct27_10-55-38_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761558942.r-technotron-trainer-0widy16f-65af2-vb4tz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94108cb4ce08b1037cbd7d7e0156a85e74da9ca43fbd5eac7a79be311a98034e
3
+ size 6632
runs/Oct27_10-55-59_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761558961.r-technotron-trainer-0widy16f-65af2-vb4tz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0f824b0ee609185e70ca938575973903d11486ee15df1744ed5c4334b357025
3
+ size 6633
runs/Oct27_10-56-14_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761558976.r-technotron-trainer-0widy16f-65af2-vb4tz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50cc97df9154e948be8ed15d2c2d5740142f6ae5483df6e86ae77903d48cdb74
3
+ size 6631
runs/Oct27_10-56-47_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761559011.r-technotron-trainer-0widy16f-65af2-vb4tz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c66c7801a6d0f82c12445635eb53c624d9bce3300d5adaa176147c76b7d6e80
3
+ size 6632
runs/Oct27_10-59-43_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761559187.r-technotron-trainer-0widy16f-65af2-vb4tz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d009bbe59746044abdbdcf4d390028c670e626930fd2927659b9eb4dac7d8056
3
+ size 6632
runs/Oct27_11-01-17_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761559284.r-technotron-trainer-0widy16f-65af2-vb4tz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe67387b7ee3d4656582c8c99fd56470fa2ccdc3d4a2efca464558004176725c
3
+ size 6632
runs/Oct27_11-02-27_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761559351.r-technotron-trainer-0widy16f-65af2-vb4tz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48c4f0ba9252f37f75fc6e4da6e1e814e13d2b65355851ffde2fd24b1a1155f1
3
+ size 6632
runs/Oct27_11-05-06_r-technotron-trainer-0widy16f-65af2-vb4tz/events.out.tfevents.1761559509.r-technotron-trainer-0widy16f-65af2-vb4tz ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0abdb783d335db438ad236d7e2974ba3f0a07082058f7be66916b22ba7624b3c
3
+ size 8767
special_tokens_map.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<s>",
3
+ "eos_token": "</s>",
4
+ "mask_token": {
5
+ "content": "<mask>",
6
+ "lstrip": true,
7
+ "normalized": true,
8
+ "rstrip": false,
9
+ "single_word": false
10
+ },
11
+ "pad_token": "<pad>",
12
+ "unk_token": "<unk>"
13
+ }
spm_char.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fcc48f3e225f627b1641db410ceb0c8649bd2b0c982e150b03f8be3728ab560
3
+ size 238473
tokenizer_config.json ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<s>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<pad>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "<unk>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "79": {
36
+ "content": "<mask>",
37
+ "lstrip": true,
38
+ "normalized": true,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "80": {
44
+ "content": "<ctc_blank>",
45
+ "lstrip": false,
46
+ "normalized": true,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": false
50
+ }
51
+ },
52
+ "bos_token": "<s>",
53
+ "clean_up_tokenization_spaces": false,
54
+ "eos_token": "</s>",
55
+ "extra_special_tokens": {},
56
+ "mask_token": "<mask>",
57
+ "model_max_length": 600,
58
+ "normalize": false,
59
+ "pad_token": "<pad>",
60
+ "processor_class": "SpeechT5Processor",
61
+ "sp_model_kwargs": {},
62
+ "tokenizer_class": "SpeechT5Tokenizer",
63
+ "unk_token": "<unk>"
64
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc56af2e51780b7659f079d02dd25a8becdd142e303d75b75a0915b42676782a
3
+ size 6097