seastar105 commited on
Commit
2771491
·
verified ·
1 Parent(s): dd0d681

Training in progress, step 5000

Browse files
.hydra/config.yaml ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ paths:
2
+ root_dir: .
3
+ data_dir: ${paths.root_dir}/data/
4
+ log_dir: ${paths.root_dir}/logs/
5
+ output_dir: ${hydra:runtime.output_dir}
6
+ work_dir: ${hydra:runtime.cwd}
7
+ data_config:
8
+ name: krishnakalyan3/emo_webds_2
9
+ model_config:
10
+ name: openai/whisper-tiny
11
+ attn_implementation: sdpa
12
+ trainer_config:
13
+ tf32: true
14
+ bf16: true
15
+ fp16: false
16
+ dataloader_num_workers: 16
17
+ eval_strategy: 'no'
18
+ eval_steps: null
19
+ save_strategy: steps
20
+ save_steps: 5000
21
+ save_total_limit: 5
22
+ gradient_checkpointing: false
23
+ gradient_checkpointing_kwargs:
24
+ use_reentrant: false
25
+ ddp_find_unused_parameters: false
26
+ logging_steps: 50
27
+ seed: 998244353
28
+ optim: adamw_torch
29
+ adam_beta1: 0.9
30
+ adam_beta2: 0.98
31
+ max_grad_norm: 1.0
32
+ per_device_train_batch_size: 128
33
+ output_dir: ${paths.output_dir}
34
+ report_to: tensorboard
35
+ logging_dir: ${trainer_config.output_dir}/tb
36
+ lr_scheduler_type: cosine
37
+ learning_rate: 0.0001
38
+ warmup_steps: 2000
39
+ max_steps: 25000
40
+ label_smoothing_factor: 0.0
41
+ push_to_hub: true
42
+ hub_model_id: seastar105/whisper-base-emo-speech-caption
43
+ split_batches: true
44
+ dataloader_prefetch_factor: 4
45
+ task_name: tiny-test
.hydra/hydra.yaml ADDED
@@ -0,0 +1,164 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ hydra:
2
+ run:
3
+ dir: ${paths.log_dir}/${task_name}/runs/${now:%Y-%m-%d}_${now:%H-%M-%S}
4
+ sweep:
5
+ dir: ${paths.log_dir}/${task_name}/multiruns/${now:%Y-%m-%d}_${now:%H-%M-%S}
6
+ subdir: ${hydra.job.num}
7
+ launcher:
8
+ _target_: hydra._internal.core_plugins.basic_launcher.BasicLauncher
9
+ sweeper:
10
+ _target_: hydra._internal.core_plugins.basic_sweeper.BasicSweeper
11
+ max_batch_size: null
12
+ params: null
13
+ help:
14
+ app_name: ${hydra.job.name}
15
+ header: '${hydra.help.app_name} is powered by Hydra.
16
+
17
+ '
18
+ footer: 'Powered by Hydra (https://hydra.cc)
19
+
20
+ Use --hydra-help to view Hydra specific help
21
+
22
+ '
23
+ template: '${hydra.help.header}
24
+
25
+ == Configuration groups ==
26
+
27
+ Compose your configuration from those groups (group=option)
28
+
29
+
30
+ $APP_CONFIG_GROUPS
31
+
32
+
33
+ == Config ==
34
+
35
+ Override anything in the config (foo.bar=value)
36
+
37
+
38
+ $CONFIG
39
+
40
+
41
+ ${hydra.help.footer}
42
+
43
+ '
44
+ hydra_help:
45
+ template: 'Hydra (${hydra.runtime.version})
46
+
47
+ See https://hydra.cc for more info.
48
+
49
+
50
+ == Flags ==
51
+
52
+ $FLAGS_HELP
53
+
54
+
55
+ == Configuration groups ==
56
+
57
+ Compose your configuration from those groups (For example, append hydra/job_logging=disabled
58
+ to command line)
59
+
60
+
61
+ $HYDRA_CONFIG_GROUPS
62
+
63
+
64
+ Use ''--cfg hydra'' to Show the Hydra config.
65
+
66
+ '
67
+ hydra_help: ???
68
+ hydra_logging:
69
+ version: 1
70
+ formatters:
71
+ simple:
72
+ format: '[%(asctime)s][HYDRA] %(message)s'
73
+ handlers:
74
+ console:
75
+ class: logging.StreamHandler
76
+ formatter: simple
77
+ stream: ext://sys.stdout
78
+ root:
79
+ level: INFO
80
+ handlers:
81
+ - console
82
+ loggers:
83
+ logging_example:
84
+ level: DEBUG
85
+ disable_existing_loggers: false
86
+ job_logging:
87
+ version: 1
88
+ formatters:
89
+ simple:
90
+ format: '[%(asctime)s][%(name)s][%(levelname)s] - %(message)s'
91
+ handlers:
92
+ console:
93
+ class: logging.StreamHandler
94
+ formatter: simple
95
+ stream: ext://sys.stdout
96
+ file:
97
+ class: logging.FileHandler
98
+ formatter: simple
99
+ filename: ${hydra.runtime.output_dir}/${task_name}.log
100
+ root:
101
+ level: INFO
102
+ handlers:
103
+ - console
104
+ - file
105
+ disable_existing_loggers: false
106
+ env: {}
107
+ mode: RUN
108
+ searchpath: []
109
+ callbacks: {}
110
+ output_subdir: .hydra
111
+ overrides:
112
+ hydra:
113
+ - hydra.mode=RUN
114
+ task:
115
+ - trainer_config.per_device_train_batch_size=128
116
+ - ++trainer_config.bf16=true
117
+ - trainer_config.max_steps=25000
118
+ - ++trainer_config.push_to_hub=true
119
+ - ++trainer_config.hub_model_id=seastar105/whisper-base-emo-speech-caption
120
+ - ++trainer_config.split_batches=true
121
+ - ++trainer_config.save_total_limit=5
122
+ - ++trainer_config.save_steps=5000
123
+ - ++trainer_config.dataloader_prefetch_factor=4
124
+ - ++trainer_config.dataloader_num_workers=16
125
+ job:
126
+ name: train
127
+ chdir: null
128
+ override_dirname: ++trainer_config.bf16=true,++trainer_config.dataloader_num_workers=16,++trainer_config.dataloader_prefetch_factor=4,++trainer_config.hub_model_id=seastar105/whisper-base-emo-speech-caption,++trainer_config.push_to_hub=true,++trainer_config.save_steps=5000,++trainer_config.save_total_limit=5,++trainer_config.split_batches=true,trainer_config.max_steps=25000,trainer_config.per_device_train_batch_size=128
129
+ id: ???
130
+ num: ???
131
+ config_name: main
132
+ env_set: {}
133
+ env_copy: []
134
+ config:
135
+ override_dirname:
136
+ kv_sep: '='
137
+ item_sep: ','
138
+ exclude_keys: []
139
+ runtime:
140
+ version: 1.3.2
141
+ version_base: '1.3'
142
+ cwd: /root/audio-caption
143
+ config_sources:
144
+ - path: hydra.conf
145
+ schema: pkg
146
+ provider: hydra
147
+ - path: /root/audio-caption/configs
148
+ schema: file
149
+ provider: main
150
+ - path: ''
151
+ schema: structured
152
+ provider: schema
153
+ output_dir: /root/audio-caption/logs/tiny-test/runs/2024-12-26_15-36-11
154
+ choices:
155
+ hydra/env: default
156
+ hydra/callbacks: null
157
+ hydra/job_logging: default
158
+ hydra/hydra_logging: default
159
+ hydra/hydra_help: default
160
+ hydra/help: default
161
+ hydra/sweeper: basic
162
+ hydra/launcher: basic
163
+ hydra/output: default
164
+ verbose: false
.hydra/overrides.yaml ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ - trainer_config.per_device_train_batch_size=128
2
+ - ++trainer_config.bf16=true
3
+ - trainer_config.max_steps=25000
4
+ - ++trainer_config.push_to_hub=true
5
+ - ++trainer_config.hub_model_id=seastar105/whisper-base-emo-speech-caption
6
+ - ++trainer_config.split_batches=true
7
+ - ++trainer_config.save_total_limit=5
8
+ - ++trainer_config.save_steps=5000
9
+ - ++trainer_config.dataloader_prefetch_factor=4
10
+ - ++trainer_config.dataloader_num_workers=16
config.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "openai/whisper-tiny",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "gelu",
5
+ "apply_spec_augment": false,
6
+ "architectures": [
7
+ "WhisperForConditionalGeneration"
8
+ ],
9
+ "attention_dropout": 0.0,
10
+ "begin_suppress_tokens": null,
11
+ "bos_token_id": 50257,
12
+ "classifier_proj_size": 256,
13
+ "d_model": 384,
14
+ "decoder_attention_heads": 6,
15
+ "decoder_ffn_dim": 1536,
16
+ "decoder_layerdrop": 0.0,
17
+ "decoder_layers": 4,
18
+ "decoder_start_token_id": 50258,
19
+ "dropout": 0.0,
20
+ "encoder_attention_heads": 6,
21
+ "encoder_ffn_dim": 1536,
22
+ "encoder_layerdrop": 0.0,
23
+ "encoder_layers": 4,
24
+ "eos_token_id": 50257,
25
+ "forced_decoder_ids": [
26
+ [
27
+ 1,
28
+ 50259
29
+ ],
30
+ [
31
+ 2,
32
+ 50359
33
+ ],
34
+ [
35
+ 3,
36
+ 50363
37
+ ]
38
+ ],
39
+ "init_std": 0.02,
40
+ "is_encoder_decoder": true,
41
+ "mask_feature_length": 10,
42
+ "mask_feature_min_masks": 0,
43
+ "mask_feature_prob": 0.0,
44
+ "mask_time_length": 10,
45
+ "mask_time_min_masks": 2,
46
+ "mask_time_prob": 0.05,
47
+ "max_length": null,
48
+ "max_source_positions": 1500,
49
+ "max_target_positions": 448,
50
+ "median_filter_width": 7,
51
+ "model_type": "whisper",
52
+ "num_hidden_layers": 4,
53
+ "num_mel_bins": 80,
54
+ "pad_token_id": 50257,
55
+ "scale_embedding": false,
56
+ "torch_dtype": "float32",
57
+ "transformers_version": "4.47.1",
58
+ "use_cache": true,
59
+ "use_weighted_layer_sum": false,
60
+ "vocab_size": 51865
61
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b86d035092ca3b34763dc3cffd8c20b0534b87e11b83d23f12ad615cd428f28
3
+ size 151061672
tb/events.out.tfevents.1735227385.c76029f3d320 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e73cd0bcc8d77f479ec5c4ffd15a72e9bd2bdac10c244be7c0d4c899d6199f7
3
+ size 27674
tiny-test.log ADDED
File without changes
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:784cdd034747ae293a66620e63a46c3dcd9545583f708cd6aaf2c774962d1099
3
+ size 5432