RobCaamano commited on
Commit
2e4825a
·
verified ·
1 Parent(s): a707d23

Upload folder using huggingface_hub

Browse files
Files changed (4) hide show
  1. config.json +201 -0
  2. ipykernel_launcher.py +18 -0
  3. model.pth +3 -0
  4. trainer_0_log.txt +486 -0
config.json ADDED
@@ -0,0 +1,201 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "output_path": "./run/training/",
3
+ "logger_uri": null,
4
+ "run_name": "Sherlock-Holmes-2-epochs",
5
+ "project_name": "XTTS-v2 Finetune",
6
+ "run_description": "\n GPT XTTS training\n ",
7
+ "print_step": 50,
8
+ "plot_step": 100,
9
+ "model_param_stats": false,
10
+ "wandb_entity": null,
11
+ "dashboard_logger": "wandb",
12
+ "save_on_interrupt": true,
13
+ "log_model_step": 1000,
14
+ "save_step": 1000,
15
+ "save_n_checkpoints": 1,
16
+ "save_checkpoints": true,
17
+ "save_all_best": false,
18
+ "save_best_after": 0,
19
+ "target_loss": null,
20
+ "print_eval": true,
21
+ "test_delay_epochs": 0,
22
+ "run_eval": true,
23
+ "run_eval_steps": null,
24
+ "distributed_backend": "nccl",
25
+ "distributed_url": "tcp://localhost:54321",
26
+ "mixed_precision": false,
27
+ "precision": "fp16",
28
+ "epochs": 2,
29
+ "batch_size": 3,
30
+ "eval_batch_size": 3,
31
+ "grad_clip": null,
32
+ "scheduler_after_epoch": true,
33
+ "lr": 5e-06,
34
+ "optimizer": "AdamW",
35
+ "optimizer_params": {
36
+ "betas": [
37
+ 0.9,
38
+ 0.96
39
+ ],
40
+ "eps": 1e-08,
41
+ "weight_decay": 0.01
42
+ },
43
+ "lr_scheduler": "MultiStepLR",
44
+ "lr_scheduler_params": {
45
+ "milestones": [
46
+ 900000,
47
+ 2700000,
48
+ 5400000
49
+ ],
50
+ "gamma": 0.5,
51
+ "last_epoch": -1
52
+ },
53
+ "use_grad_scaler": false,
54
+ "allow_tf32": false,
55
+ "cudnn_enable": true,
56
+ "cudnn_deterministic": false,
57
+ "cudnn_benchmark": false,
58
+ "training_seed": 1,
59
+ "model": "xtts",
60
+ "num_loader_workers": 0,
61
+ "num_eval_loader_workers": 0,
62
+ "use_noise_augment": false,
63
+ "audio": {
64
+ "sample_rate": 16000,
65
+ "output_sample_rate": 24000,
66
+ "dvae_sample_rate": 16000
67
+ },
68
+ "use_phonemes": false,
69
+ "phonemizer": null,
70
+ "phoneme_language": null,
71
+ "compute_input_seq_cache": false,
72
+ "text_cleaner": null,
73
+ "enable_eos_bos_chars": false,
74
+ "test_sentences_file": "",
75
+ "phoneme_cache_path": null,
76
+ "characters": null,
77
+ "add_blank": false,
78
+ "batch_group_size": 48,
79
+ "loss_masking": null,
80
+ "min_audio_len": 1,
81
+ "max_audio_len": Infinity,
82
+ "min_text_len": 1,
83
+ "max_text_len": Infinity,
84
+ "compute_f0": false,
85
+ "compute_energy": false,
86
+ "compute_linear_spec": false,
87
+ "precompute_num_workers": 0,
88
+ "start_by_longest": false,
89
+ "shuffle": false,
90
+ "drop_last": false,
91
+ "datasets": [
92
+ {
93
+ "formatter": "",
94
+ "dataset_name": "",
95
+ "path": "",
96
+ "meta_file_train": "",
97
+ "ignored_speakers": null,
98
+ "language": "",
99
+ "phonemizer": "",
100
+ "meta_file_val": "",
101
+ "meta_file_attn_mask": ""
102
+ }
103
+ ],
104
+ "test_sentences": [
105
+ {
106
+ "text": "It took me quite a long time to develop a voice, and now that I have it I'm not going to be silent.",
107
+ "speaker_wav": "datasets/Sherlock Holmes Stories Read by Benedict Cumberbatch/wavs/chunk_0220.wav",
108
+ "language": "en"
109
+ },
110
+ {
111
+ "text": "This cake is great. It's so delicious and moist.",
112
+ "speaker_wav": "datasets/Sherlock Holmes Stories Read by Benedict Cumberbatch/wavs/chunk_0220.wav",
113
+ "language": "en"
114
+ }
115
+ ],
116
+ "eval_split_max_size": 256,
117
+ "eval_split_size": 0.01,
118
+ "use_speaker_weighted_sampler": false,
119
+ "speaker_weighted_sampler_alpha": 1.0,
120
+ "use_language_weighted_sampler": false,
121
+ "language_weighted_sampler_alpha": 1.0,
122
+ "use_length_weighted_sampler": false,
123
+ "length_weighted_sampler_alpha": 1.0,
124
+ "model_args": {
125
+ "gpt_batch_size": 1,
126
+ "enable_redaction": false,
127
+ "kv_cache": true,
128
+ "gpt_checkpoint": "",
129
+ "clvp_checkpoint": null,
130
+ "decoder_checkpoint": null,
131
+ "num_chars": 255,
132
+ "tokenizer_file": "./XTTS-files/vocab.json",
133
+ "gpt_max_audio_tokens": 605,
134
+ "gpt_max_text_tokens": 402,
135
+ "gpt_max_prompt_tokens": 70,
136
+ "gpt_layers": 30,
137
+ "gpt_n_model_channels": 1024,
138
+ "gpt_n_heads": 16,
139
+ "gpt_number_text_tokens": 6681,
140
+ "gpt_start_text_token": 261,
141
+ "gpt_stop_text_token": 0,
142
+ "gpt_num_audio_tokens": 1026,
143
+ "gpt_start_audio_token": 1024,
144
+ "gpt_stop_audio_token": 1025,
145
+ "gpt_code_stride_len": 1024,
146
+ "gpt_use_masking_gt_prompt_approach": true,
147
+ "gpt_use_perceiver_resampler": true,
148
+ "input_sample_rate": 22050,
149
+ "output_sample_rate": 24000,
150
+ "output_hop_length": 256,
151
+ "decoder_input_dim": 1024,
152
+ "d_vector_dim": 512,
153
+ "cond_d_vector_in_each_upsampling_layer": true,
154
+ "duration_const": 102400,
155
+ "min_conditioning_length": 66150,
156
+ "max_conditioning_length": 143677,
157
+ "gpt_loss_text_ce_weight": 0.01,
158
+ "gpt_loss_mel_ce_weight": 1.0,
159
+ "debug_loading_failures": true,
160
+ "max_wav_length": 255995,
161
+ "max_text_length": 66150,
162
+ "mel_norm_file": "./XTTS-files/mel_stats.pth",
163
+ "dvae_checkpoint": "./XTTS-files/dvae.pth",
164
+ "xtts_checkpoint": "./XTTS-files/model.pth",
165
+ "vocoder": ""
166
+ },
167
+ "model_dir": null,
168
+ "languages": [
169
+ "en",
170
+ "es",
171
+ "fr",
172
+ "de",
173
+ "it",
174
+ "pt",
175
+ "pl",
176
+ "tr",
177
+ "ru",
178
+ "nl",
179
+ "cs",
180
+ "ar",
181
+ "zh-cn",
182
+ "hu",
183
+ "ko",
184
+ "ja",
185
+ "hi"
186
+ ],
187
+ "temperature": 0.85,
188
+ "length_penalty": 1.0,
189
+ "repetition_penalty": 2.0,
190
+ "top_k": 50,
191
+ "top_p": 0.85,
192
+ "num_gpt_outputs": 1,
193
+ "gpt_cond_len": 12,
194
+ "gpt_cond_chunk_len": 4,
195
+ "max_ref_len": 10,
196
+ "sound_norm_refs": false,
197
+ "optimizer_wd_only_on_weights": true,
198
+ "weighted_loss_attrs": null,
199
+ "weighted_loss_multipliers": null,
200
+ "github_branch": "unknown"
201
+ }
ipykernel_launcher.py ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ """Entry point for launching an IPython kernel.
2
+
3
+ This is separate from the ipykernel package so we can avoid doing imports until
4
+ after removing the cwd from sys.path.
5
+ """
6
+
7
+ import sys
8
+ from pathlib import Path
9
+
10
+ if __name__ == "__main__":
11
+ # Remove the CWD from sys.path while we load stuff.
12
+ # This is added back by InteractiveShellApp.init_path()
13
+ if sys.path[0] == "" or Path(sys.path[0]) == Path.cwd():
14
+ del sys.path[0]
15
+
16
+ from ipykernel import kernelapp as app
17
+
18
+ app.launch_new_instance()
model.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15583983f6227c2c94dd690afb57bcd499b678865ba7ac2abc82ac18dc2b63ab
3
+ size 5607926869
trainer_0_log.txt ADDED
@@ -0,0 +1,486 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ > Training Environment:
2
+ | > Backend: Torch
3
+ | > Mixed precision: False
4
+ | > Precision: float32
5
+ | > Current device: 0
6
+ | > Num. of GPUs: 1
7
+ | > Num. of CPUs: 20
8
+ | > Num. of Torch Threads: 1
9
+ | > Torch seed: 1
10
+ | > Torch CUDNN: True
11
+ | > Torch CUDNN deterministic: False
12
+ | > Torch CUDNN benchmark: False
13
+ | > Torch TF32 MatMul: False
14
+
15
+ > Model has 518442047 parameters
16
+
17
+  > EPOCH: 0/2
18
+ --> run\training\Sherlock-Holmes-2-epochs-April-25-2025_03+08PM-0000000
19
+
20
+  > EVALUATION 
21
+
22
+  --> STEP: 0
23
+ | > loss_text_ce: 0.02065003104507923 (0.02065003104507923)
24
+ | > loss_mel_ce: 4.8800153732299805 (4.8800153732299805)
25
+ | > loss: 4.900665283203125 (4.900665283203125)
26
+
27
+  --> STEP: 1
28
+ | > loss_text_ce: 0.02369523048400879 (0.02369523048400879)
29
+ | > loss_mel_ce: 4.812253952026367 (4.812253952026367)
30
+ | > loss: 4.835948944091797 (4.835948944091797)
31
+
32
+  --> STEP: 2
33
+ | > loss_text_ce: 0.021101634949445724 (0.022398432716727257)
34
+ | > loss_mel_ce: 4.779322624206543 (4.795788288116455)
35
+ | > loss: 4.800424098968506 (4.818186521530151)
36
+
37
+  --> STEP: 3
38
+ | > loss_text_ce: 0.021626941859722137 (0.02214126909772555)
39
+ | > loss_mel_ce: 4.4567179679870605 (4.682764848073323)
40
+ | > loss: 4.478344917297363 (4.704905986785889)
41
+
42
+  --> STEP: 4
43
+ | > loss_text_ce: 0.021944627165794373 (0.022092108614742756)
44
+ | > loss_mel_ce: 4.730851650238037 (4.694786548614502)
45
+ | > loss: 4.752796173095703 (4.716878533363342)
46
+
47
+  --> STEP: 5
48
+ | > loss_text_ce: 0.021313846111297607 (0.021936456114053725)
49
+ | > loss_mel_ce: 4.908026218414307 (4.7374344825744625)
50
+ | > loss: 4.92933988571167 (4.759370803833008)
51
+
52
+  --> STEP: 6
53
+ | > loss_text_ce: 0.02120722085237503 (0.021814916903773945)
54
+ | > loss_mel_ce: 4.877439022064209 (4.760768572489421)
55
+ | > loss: 4.898646354675293 (4.782583395640056)
56
+
57
+
58
+ --> EVAL PERFORMANCE
59
+ | > avg_loader_time: 0.06062706311543783 (+0.0)
60
+ | > avg_loss_text_ce: 0.021814916903773945 (+0.0)
61
+ | > avg_loss_mel_ce: 4.760768572489421 (+0.0)
62
+ | > avg_loss: 4.782583395640056 (+0.0)
63
+
64
+
65
+  > EPOCH: 1/2
66
+ --> run\training\Sherlock-Holmes-2-epochs-April-25-2025_03+08PM-0000000
67
+
68
+  > TRAINING (2025-04-25 15:08:41) 
69
+
70
+  --> TIME: 2025-04-25 15:08:41 -- STEP: 0/370 -- GLOBAL_STEP: 0
71
+ | > loss_text_ce: 0.022243821993470192 (0.022243821993470192)
72
+ | > loss_mel_ce: 5.101950168609619 (5.101950168609619)
73
+ | > loss: 0.06100231409072876 (0.06100231409072876)
74
+ | > current_lr: 5e-06
75
+ | > step_time: 0.1968 (0.1967782974243164)
76
+ | > loader_time: 0.054 (0.0540008544921875)
77
+
78
+
79
+  --> TIME: 2025-04-25 15:08:52 -- STEP: 50/370 -- GLOBAL_STEP: 50
80
+ | > loss_text_ce: 0.021904677152633667 (0.023353138603270054)
81
+ | > loss_mel_ce: 4.993193626403809 (4.857127528190612)
82
+ | > loss: 0.05970355123281479 (0.05810096099972725)
83
+ | > current_lr: 5e-06
84
+ | > step_time: 0.1014 (0.09666483402252198)
85
+ | > loader_time: 0.053 (0.053735547065734864)
86
+
87
+
88
+  --> TIME: 2025-04-25 15:09:02 -- STEP: 100/370 -- GLOBAL_STEP: 100
89
+ | > loss_text_ce: 0.02186979539692402 (0.023663721550256013)
90
+ | > loss_mel_ce: 5.097776889801025 (4.844553368091585)
91
+ | > loss: 0.06094817444682121 (0.057954966314136984)
92
+ | > current_lr: 5e-06
93
+ | > step_time: 0.1057 (0.09815988302230835)
94
+ | > loader_time: 0.0514 (0.05377385139465332)
95
+
96
+
97
+  --> TIME: 2025-04-25 15:09:19 -- STEP: 150/370 -- GLOBAL_STEP: 150
98
+ | > loss_text_ce: 0.026617947965860367 (0.023607976113756497)
99
+ | > loss_mel_ce: 4.41134786605835 (4.817071544329327)
100
+ | > loss: 0.05283292755484581 (0.0576271382222573)
101
+ | > current_lr: 5e-06
102
+ | > step_time: 0.1661 (0.14195078055063884)
103
+ | > loader_time: 0.046 (0.053330375353495274)
104
+
105
+
106
+  --> TIME: 2025-04-25 15:09:37 -- STEP: 200/370 -- GLOBAL_STEP: 200
107
+ | > loss_text_ce: 0.021193694323301315 (0.02364832037128508)
108
+ | > loss_mel_ce: 5.093357086181641 (4.778294647932054)
109
+ | > loss: 0.06088750809431076 (0.057165988851338626)
110
+ | > current_lr: 5e-06
111
+ | > step_time: 0.3385 (0.16869270443916323)
112
+ | > loader_time: 0.053 (0.052930202484130856)
113
+
114
+
115
+  --> TIME: 2025-04-25 15:09:57 -- STEP: 250/370 -- GLOBAL_STEP: 250
116
+ | > loss_text_ce: 0.025131795555353165 (0.02353109458088875)
117
+ | > loss_mel_ce: 4.591306686401367 (4.7446932544708265)
118
+ | > loss: 0.05495760217308998 (0.05676457673311233)
119
+ | > current_lr: 5e-06
120
+ | > step_time: 0.4551 (0.19466437816619875)
121
+ | > loader_time: 0.0546 (0.05248429012298584)
122
+
123
+
124
+  --> TIME: 2025-04-25 15:10:23 -- STEP: 300/370 -- GLOBAL_STEP: 300
125
+ | > loss_text_ce: 0.02108745090663433 (0.023446181671073044)
126
+ | > loss_mel_ce: 4.584284782409668 (4.713035838603975)
127
+ | > loss: 0.05482586473226547 (0.05638669184098641)
128
+ | > current_lr: 5e-06
129
+ | > step_time: 0.3054 (0.22047462224960337)
130
+ | > loader_time: 0.0541 (0.05223121722539266)
131
+
132
+
133
+  --> TIME: 2025-04-25 15:10:49 -- STEP: 350/370 -- GLOBAL_STEP: 350
134
+ | > loss_text_ce: 0.025105265900492668 (0.023427489916128773)
135
+ | > loss_mel_ce: 4.507345676422119 (4.689699961798533)
136
+ | > loss: 0.053957752883434296 (0.056108661315270834)
137
+ | > current_lr: 5e-06
138
+ | > step_time: 0.2924 (0.24373621872493206)
139
+ | > loader_time: 0.0491 (0.05209174701145717)
140
+
141
+
142
+  > EVALUATION 
143
+
144
+  --> STEP: 0
145
+ | > loss_text_ce: 0.020510854199528694 (0.020510854199528694)
146
+ | > loss_mel_ce: 4.415481090545654 (4.415481090545654)
147
+ | > loss: 4.4359917640686035 (4.4359917640686035)
148
+
149
+  --> STEP: 1
150
+ | > loss_text_ce: 0.023326078429818153 (0.023326078429818153)
151
+ | > loss_mel_ce: 4.377121448516846 (4.377121448516846)
152
+ | > loss: 4.400447368621826 (4.400447368621826)
153
+
154
+  --> STEP: 2
155
+ | > loss_text_ce: 0.020727859809994698 (0.022026969119906425)
156
+ | > loss_mel_ce: 4.428055763244629 (4.402588605880737)
157
+ | > loss: 4.4487833976745605 (4.424615383148193)
158
+
159
+  --> STEP: 3
160
+ | > loss_text_ce: 0.021357735618948936 (0.02180389128625393)
161
+ | > loss_mel_ce: 4.123639106750488 (4.309605439503987)
162
+ | > loss: 4.144996643066406 (4.331409136454265)
163
+
164
+  --> STEP: 4
165
+ | > loss_text_ce: 0.02157936803996563 (0.021747760474681854)
166
+ | > loss_mel_ce: 4.35579252243042 (4.321152210235596)
167
+ | > loss: 4.377371788024902 (4.342899799346924)
168
+
169
+  --> STEP: 5
170
+ | > loss_text_ce: 0.021084001287817955 (0.021615008637309075)
171
+ | > loss_mel_ce: 4.472175121307373 (4.351356792449951)
172
+ | > loss: 4.493258953094482 (4.372971630096435)
173
+
174
+  --> STEP: 6
175
+ | > loss_text_ce: 0.02100624516606331 (0.021513548058768112)
176
+ | > loss_mel_ce: 4.452880382537842 (4.368277390797933)
177
+ | > loss: 4.473886489868164 (4.389790773391724)
178
+
179
+
180
+ --> EVAL PERFORMANCE
181
+ | > avg_loader_time: 0.054283698399861656 (-0.006343364715576172)
182
+ | > avg_loss_text_ce: 0.021513548058768112 (-0.00030136884500583244)
183
+ | > avg_loss_mel_ce: 4.368277390797933 (-0.3924911816914882)
184
+ | > avg_loss: 4.389790773391724 (-0.392792622248332)
185
+
186
+ > BEST MODEL : run\training\Sherlock-Holmes-2-epochs-April-25-2025_03+08PM-0000000\best_model_370.pth
187
+ > Training Environment:
188
+ | > Backend: Torch
189
+ | > Mixed precision: False
190
+ | > Precision: float32
191
+ | > Current device: 0
192
+ | > Num. of GPUs: 1
193
+ | > Num. of CPUs: 20
194
+ | > Num. of Torch Threads: 1
195
+ | > Torch seed: 1
196
+ | > Torch CUDNN: True
197
+ | > Torch CUDNN deterministic: False
198
+ | > Torch CUDNN benchmark: False
199
+ | > Torch TF32 MatMul: False
200
+
201
+ > Model has 518442047 parameters
202
+
203
+  > EPOCH: 0/3
204
+ --> run\training\Sherlock-Holmes-3-epochs-April-25-2025_03+11PM-0000000
205
+
206
+  > EVALUATION 
207
+
208
+  --> STEP: 0
209
+ | > loss_text_ce: 0.02065003104507923 (0.02065003104507923)
210
+ | > loss_mel_ce: 4.8800153732299805 (4.8800153732299805)
211
+ | > loss: 4.900665283203125 (4.900665283203125)
212
+
213
+  --> STEP: 1
214
+ | > loss_text_ce: 0.02369523048400879 (0.02369523048400879)
215
+ | > loss_mel_ce: 4.812253952026367 (4.812253952026367)
216
+ | > loss: 4.835948944091797 (4.835948944091797)
217
+
218
+  --> STEP: 2
219
+ | > loss_text_ce: 0.021101634949445724 (0.022398432716727257)
220
+ | > loss_mel_ce: 4.779322624206543 (4.795788288116455)
221
+ | > loss: 4.800424098968506 (4.818186521530151)
222
+
223
+  --> STEP: 3
224
+ | > loss_text_ce: 0.021626941859722137 (0.02214126909772555)
225
+ | > loss_mel_ce: 4.4567179679870605 (4.682764848073323)
226
+ | > loss: 4.478344917297363 (4.704905986785889)
227
+
228
+  --> STEP: 4
229
+ | > loss_text_ce: 0.021944627165794373 (0.022092108614742756)
230
+ | > loss_mel_ce: 4.730851650238037 (4.694786548614502)
231
+ | > loss: 4.752796173095703 (4.716878533363342)
232
+
233
+  --> STEP: 5
234
+ | > loss_text_ce: 0.021313846111297607 (0.021936456114053725)
235
+ | > loss_mel_ce: 4.908026218414307 (4.7374344825744625)
236
+ | > loss: 4.92933988571167 (4.759370803833008)
237
+
238
+  --> STEP: 6
239
+ | > loss_text_ce: 0.02120722085237503 (0.021814916903773945)
240
+ | > loss_mel_ce: 4.877439022064209 (4.760768572489421)
241
+ | > loss: 4.898646354675293 (4.782583395640056)
242
+
243
+
244
+ --> EVAL PERFORMANCE
245
+ | > avg_loader_time: 0.061905860900878906 (+0.0)
246
+ | > avg_loss_text_ce: 0.021814916903773945 (+0.0)
247
+ | > avg_loss_mel_ce: 4.760768572489421 (+0.0)
248
+ | > avg_loss: 4.782583395640056 (+0.0)
249
+
250
+
251
+  > EPOCH: 1/3
252
+ --> run\training\Sherlock-Holmes-3-epochs-April-25-2025_03+11PM-0000000
253
+
254
+  > TRAINING (2025-04-25 15:12:02) 
255
+
256
+  --> TIME: 2025-04-25 15:12:03 -- STEP: 0/370 -- GLOBAL_STEP: 0
257
+ | > loss_text_ce: 0.022243821993470192 (0.022243821993470192)
258
+ | > loss_mel_ce: 5.101950645446777 (5.101950645446777)
259
+ | > loss: 0.06100231781601906 (0.06100231781601906)
260
+ | > current_lr: 5e-06
261
+ | > step_time: 0.0796 (0.07958102226257324)
262
+ | > loader_time: 0.0468 (0.04683685302734375)
263
+
264
+
265
+  --> TIME: 2025-04-25 15:12:13 -- STEP: 50/370 -- GLOBAL_STEP: 50
266
+ | > loss_text_ce: 0.021904677152633667 (0.023353138603270054)
267
+ | > loss_mel_ce: 4.993193626403809 (4.857127528190612)
268
+ | > loss: 0.05970355123281479 (0.05810096099972725)
269
+ | > current_lr: 5e-06
270
+ | > step_time: 0.0971 (0.09639417648315429)
271
+ | > loader_time: 0.0485 (0.050910625457763675)
272
+
273
+
274
+  --> TIME: 2025-04-25 15:12:27 -- STEP: 100/370 -- GLOBAL_STEP: 100
275
+ | > loss_text_ce: 0.02186979539692402 (0.023663721568882464)
276
+ | > loss_mel_ce: 5.097776889801025 (4.844553372859956)
277
+ | > loss: 0.06094817444682121 (0.05795496646314859)
278
+ | > current_lr: 5e-06
279
+ | > step_time: 0.2023 (0.12174204587936405)
280
+ | > loader_time: 0.049 (0.05165000915527344)
281
+
282
+
283
+  --> TIME: 2025-04-25 15:13:04 -- STEP: 150/370 -- GLOBAL_STEP: 150
284
+ | > loss_text_ce: 0.026617947965860367 (0.023607976151009402)
285
+ | > loss_mel_ce: 4.41134786605835 (4.817071537971498)
286
+ | > loss: 0.05283292755484581 (0.057627138197422025)
287
+ | > current_lr: 5e-06
288
+ | > step_time: 0.2527 (0.25384192148844403)
289
+ | > loader_time: 0.0494 (0.051709008216857914)
290
+
291
+
292
+  --> TIME: 2025-04-25 15:13:41 -- STEP: 200/370 -- GLOBAL_STEP: 200
293
+ | > loss_text_ce: 0.021193763241171837 (0.023648322029039263)
294
+ | > loss_mel_ce: 5.093355655670166 (4.778295999765397)
295
+ | > loss: 0.060887496918439865 (0.057166005037724975)
296
+ | > current_lr: 5e-06
297
+ | > step_time: 0.4276 (0.3186194431781769)
298
+ | > loader_time: 0.056 (0.051824249029159546)
299
+
300
+
301
+  --> TIME: 2025-04-25 15:14:17 -- STEP: 250/370 -- GLOBAL_STEP: 250
302
+ | > loss_text_ce: 0.025131838396191597 (0.023531100630760193)
303
+ | > loss_mel_ce: 4.591292381286621 (4.744693719863892)
304
+ | > loss: 0.05495743080973625 (0.05676458239555359)
305
+ | > current_lr: 5e-06
306
+ | > step_time: 0.7552 (0.35306662845611575)
307
+ | > loader_time: 0.0536 (0.05188997745513916)
308
+
309
+
310
+  --> TIME: 2025-04-25 15:15:09 -- STEP: 300/370 -- GLOBAL_STEP: 300
311
+ | > loss_text_ce: 0.021087417379021645 (0.023446188066154727)
312
+ | > loss_mel_ce: 4.584291934967041 (4.713036131064096)
313
+ | > loss: 0.05482594668865204 (0.05638669546693564)
314
+ | > current_lr: 5e-06
315
+ | > step_time: 0.5866 (0.41808438301086415)
316
+ | > loader_time: 0.0495 (0.051725912888844816)
317
+
318
+
319
+  --> TIME: 2025-04-25 15:15:56 -- STEP: 350/370 -- GLOBAL_STEP: 350
320
+ | > loss_text_ce: 0.0251054335385561 (0.02342749488140857)
321
+ | > loss_mel_ce: 4.507317066192627 (4.689699707031248)
322
+ | > loss: 0.053957413882017136 (0.05610865838825703)
323
+ | > current_lr: 5e-06
324
+ | > step_time: 0.4707 (0.45240074838910777)
325
+ | > loader_time: 0.0487 (0.05162254674094064)
326
+
327
+
328
+  > EVALUATION 
329
+
330
+  --> STEP: 0
331
+ | > loss_text_ce: 0.020510999485850334 (0.020510999485850334)
332
+ | > loss_mel_ce: 4.415628910064697 (4.415628910064697)
333
+ | > loss: 4.436140060424805 (4.436140060424805)
334
+
335
+  --> STEP: 1
336
+ | > loss_text_ce: 0.023326119408011436 (0.023326119408011436)
337
+ | > loss_mel_ce: 4.377127647399902 (4.377127647399902)
338
+ | > loss: 4.400453567504883 (4.400453567504883)
339
+
340
+  --> STEP: 2
341
+ | > loss_text_ce: 0.02072770707309246 (0.02202691324055195)
342
+ | > loss_mel_ce: 4.428041458129883 (4.402584552764893)
343
+ | > loss: 4.4487690925598145 (4.424611330032349)
344
+
345
+  --> STEP: 3
346
+ | > loss_text_ce: 0.02135772444307804 (0.021803850308060646)
347
+ | > loss_mel_ce: 4.123610973358154 (4.3095933596293134)
348
+ | > loss: 4.144968509674072 (4.33139705657959)
349
+
350
+  --> STEP: 4
351
+ | > loss_text_ce: 0.021579395979642868 (0.0217477367259562)
352
+ | > loss_mel_ce: 4.355801105499268 (4.321145296096802)
353
+ | > loss: 4.37738037109375 (4.34289288520813)
354
+
355
+  --> STEP: 5
356
+ | > loss_text_ce: 0.02108406275510788 (0.02161500193178654)
357
+ | > loss_mel_ce: 4.472189903259277 (4.3513542175292965)
358
+ | > loss: 4.493273735046387 (4.372969055175782)
359
+
360
+  --> STEP: 6
361
+ | > loss_text_ce: 0.02100609987974167 (0.021513518256445725)
362
+ | > loss_mel_ce: 4.452908515930176 (4.368279933929443)
363
+ | > loss: 4.473914623260498 (4.389793316523234)
364
+
365
+
366
+ --> EVAL PERFORMANCE
367
+ | > avg_loader_time: 0.0531382163365682 (-0.008767644564310707)
368
+ | > avg_loss_text_ce: 0.021513518256445725 (-0.00030139864732822014)
369
+ | > avg_loss_mel_ce: 4.368279933929443 (-0.3924886385599775)
370
+ | > avg_loss: 4.389793316523234 (-0.3927900791168213)
371
+
372
+ > BEST MODEL : run\training\Sherlock-Holmes-3-epochs-April-25-2025_03+11PM-0000000\best_model_370.pth
373
+
374
+  > EPOCH: 2/3
375
+ --> run\training\Sherlock-Holmes-3-epochs-April-25-2025_03+11PM-0000000
376
+
377
+  > TRAINING (2025-04-25 15:17:32) 
378
+
379
+  --> TIME: 2025-04-25 15:18:36 -- STEP: 30/370 -- GLOBAL_STEP: 400
380
+ | > loss_text_ce: 0.026931729167699814 (0.023552863424022994)
381
+ | > loss_mel_ce: 4.232264518737793 (4.534945948918659)
382
+ | > loss: 0.0507047176361084 (0.05426784406105677)
383
+ | > current_lr: 5e-06
384
+ | > step_time: 1.5333 (1.5590810775756836)
385
+ | > loader_time: 0.0677 (0.06684258778889975)
386
+
387
+
388
+  --> TIME: 2025-04-25 15:20:26 -- STEP: 80/370 -- GLOBAL_STEP: 450
389
+ | > loss_text_ce: 0.020132526755332947 (0.023305393802002073)
390
+ | > loss_mel_ce: 4.565151214599609 (4.511084467172622)
391
+ | > loss: 0.05458671227097511 (0.05398083268664777)
392
+ | > current_lr: 5e-06
393
+ | > step_time: 1.7984 (1.6084659487009048)
394
+ | > loader_time: 0.0573 (0.062030056118965136)
395
+
396
+
397
+  --> TIME: 2025-04-25 15:22:09 -- STEP: 130/370 -- GLOBAL_STEP: 500
398
+ | > loss_text_ce: 0.022751471027731895 (0.02329650246180021)
399
+ | > loss_mel_ce: 4.295829772949219 (4.466045319117033)
400
+ | > loss: 0.05141168087720871 (0.05344454657572966)
401
+ | > current_lr: 5e-06
402
+ | > step_time: 1.6361 (1.58745659498068)
403
+ | > loader_time: 0.0562 (0.06053082759563739)
404
+
405
+
406
+  --> TIME: 2025-04-25 15:23:50 -- STEP: 180/370 -- GLOBAL_STEP: 550
407
+ | > loss_text_ce: 0.024590985849499702 (0.02325609718552894)
408
+ | > loss_mel_ce: 4.346665382385254 (4.431982457637788)
409
+ | > loss: 0.0520387664437294 (0.05303855517672168)
410
+ | > current_lr: 5e-06
411
+ | > step_time: 1.4534 (1.5663768596119352)
412
+ | > loader_time: 0.053 (0.05971963670518663)
413
+
414
+
415
+  --> TIME: 2025-04-25 15:25:27 -- STEP: 230/370 -- GLOBAL_STEP: 600
416
+ | > loss_text_ce: 0.022246310487389565 (0.023372901101475177)
417
+ | > loss_mel_ce: 4.374294281005859 (4.418170050952747)
418
+ | > loss: 0.05233976989984512 (0.05287551223907782)
419
+ | > current_lr: 5e-06
420
+ | > step_time: 1.1929 (1.5410345740940259)
421
+ | > loader_time: 0.0506 (0.05902964861496635)
422
+
423
+
424
+  --> TIME: 2025-04-25 15:27:11 -- STEP: 280/370 -- GLOBAL_STEP: 650
425
+ | > loss_text_ce: 0.022131072357296944 (0.023339813382231768)
426
+ | > loss_mel_ce: 4.388411521911621 (4.400581210851671)
427
+ | > loss: 0.0525064580142498 (0.05266572743920344)
428
+ | > current_lr: 5e-06
429
+ | > step_time: 1.9263 (1.53961922611509)
430
+ | > loader_time: 0.0595 (0.05847194194793701)
431
+
432
+
433
+  --> TIME: 2025-04-25 15:29:05 -- STEP: 330/370 -- GLOBAL_STEP: 700
434
+ | > loss_text_ce: 0.023882780224084854 (0.023317814375640778)
435
+ | > loss_mel_ce: 4.366799831390381 (4.392894887924197)
436
+ | > loss: 0.052270032465457916 (0.052573961662975226)
437
+ | > current_lr: 5e-06
438
+ | > step_time: 1.6675 (1.555066782055479)
439
+ | > loader_time: 0.0466 (0.05809157906156598)
440
+
441
+
442
+  > EVALUATION 
443
+
444
+  --> STEP: 0
445
+ | > loss_text_ce: 0.020527256652712822 (0.020527256652712822)
446
+ | > loss_mel_ce: 4.198864936828613 (4.198864936828613)
447
+ | > loss: 4.2193922996521 (4.2193922996521)
448
+
449
+  --> STEP: 1
450
+ | > loss_text_ce: 0.023331278935074806 (0.023331278935074806)
451
+ | > loss_mel_ce: 4.249141693115234 (4.249141693115234)
452
+ | > loss: 4.272472858428955 (4.272472858428955)
453
+
454
+  --> STEP: 2
455
+ | > loss_text_ce: 0.02069421485066414 (0.022012746892869473)
456
+ | > loss_mel_ce: 4.16215181350708 (4.205646753311157)
457
+ | > loss: 4.1828460693359375 (4.227659463882446)
458
+
459
+  --> STEP: 3
460
+ | > loss_text_ce: 0.021366318687796593 (0.021797270824511845)
461
+ | > loss_mel_ce: 3.9948670864105225 (4.135386864344279)
462
+ | > loss: 4.016233444213867 (4.15718412399292)
463
+
464
+  --> STEP: 4
465
+ | > loss_text_ce: 0.021582989022135735 (0.021743700373917818)
466
+ | > loss_mel_ce: 4.194777965545654 (4.150234639644623)
467
+ | > loss: 4.216361045837402 (4.1719783544540405)
468
+
469
+  --> STEP: 5
470
+ | > loss_text_ce: 0.021038610488176346 (0.021602682396769525)
471
+ | > loss_mel_ce: 4.166754245758057 (4.153538560867309)
472
+ | > loss: 4.187792778015137 (4.17514123916626)
473
+
474
+  --> STEP: 6
475
+ | > loss_text_ce: 0.021066777408123016 (0.021513364898661774)
476
+ | > loss_mel_ce: 4.266147136688232 (4.172306656837463)
477
+ | > loss: 4.2872138023376465 (4.193819999694824)
478
+
479
+
480
+ --> EVAL PERFORMANCE
481
+ | > avg_loader_time: 0.05410194396972656 (+0.0009637276331583636)
482
+ | > avg_loss_text_ce: 0.021513364898661774 (-1.5335778395103583e-07)
483
+ | > avg_loss_mel_ce: 4.172306656837463 (-0.19597327709197998)
484
+ | > avg_loss: 4.193819999694824 (-0.19597331682841013)
485
+
486
+ > BEST MODEL : run\training\Sherlock-Holmes-3-epochs-April-25-2025_03+11PM-0000000\best_model_740.pth