kishorekashyap commited on
Commit
ae56147
·
1 Parent(s): 152aac7

Upload hyperparams.yaml

Browse files
Files changed (1) hide show
  1. hyperparams.yaml +72 -0
hyperparams.yaml ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ # ################################
3
+ # Model: Tacotroon2 for TTS
4
+ # Authors: Jonas Freiknecht
5
+ # ################################
6
+
7
+ mask_padding: True
8
+ n_mel_channels: 80
9
+ n_symbols: 148
10
+ symbols_embedding_dim: 256
11
+ encoder_kernel_size: 5
12
+ encoder_n_convolutions: 3
13
+ encoder_embedding_dim: 256
14
+ attention_rnn_dim: 1024
15
+ attention_dim: 128
16
+ attention_location_n_filters: 32
17
+ attention_location_kernel_size: 31
18
+ n_frames_per_step: 1
19
+ decoder_rnn_dim: 512
20
+ prenet_dim: 128
21
+ max_decoder_steps: 1000
22
+ gate_threshold: 0.5
23
+ p_attention_dropout: 0.1
24
+ p_decoder_dropout: 0.1
25
+ postnet_embedding_dim: 256
26
+ postnet_kernel_size: 5
27
+ postnet_n_convolutions: 5
28
+ decoder_no_early_stopping: False
29
+ sample_rate: 22050
30
+
31
+ # Model
32
+ model: !new:speechbrain.lobes.models.Tacotron2.Tacotron2
33
+ mask_padding: !ref <mask_padding>
34
+ n_mel_channels: !ref <n_mel_channels>
35
+ # symbols
36
+ n_symbols: !ref <n_symbols>
37
+ symbols_embedding_dim: !ref <symbols_embedding_dim>
38
+ # encoder
39
+ encoder_kernel_size: !ref <encoder_kernel_size>
40
+ encoder_n_convolutions: !ref <encoder_n_convolutions>
41
+ encoder_embedding_dim: !ref <encoder_embedding_dim>
42
+ # attention
43
+ attention_rnn_dim: !ref <attention_rnn_dim>
44
+ attention_dim: !ref <attention_dim>
45
+ # attention location
46
+ attention_location_n_filters: !ref <attention_location_n_filters>
47
+ attention_location_kernel_size: !ref <attention_location_kernel_size>
48
+ # decoder
49
+ n_frames_per_step: !ref <n_frames_per_step>
50
+ decoder_rnn_dim: !ref <decoder_rnn_dim>
51
+ prenet_dim: !ref <prenet_dim>
52
+ max_decoder_steps: !ref <max_decoder_steps>
53
+ gate_threshold: !ref <gate_threshold>
54
+ p_attention_dropout: !ref <p_attention_dropout>
55
+ p_decoder_dropout: !ref <p_decoder_dropout>
56
+ # postnet
57
+ postnet_embedding_dim: !ref <postnet_embedding_dim>
58
+ postnet_kernel_size: !ref <postnet_kernel_size>
59
+ postnet_n_convolutions: !ref <postnet_n_convolutions>
60
+ decoder_no_early_stopping: !ref <decoder_no_early_stopping>
61
+
62
+ # Function that converts the text into a sequence of valid characters.
63
+ text_to_sequence: !name:speechbrain.utils.text_to_sequence.text_to_sequence
64
+
65
+ modules:
66
+ model: !ref <model>
67
+
68
+ pretrainer: !new:speechbrain.utils.parameter_transfer.Pretrainer
69
+ loadables:
70
+ model: !ref <model>
71
+
72
+