quim-motger commited on
Commit
fcf0c43
1 Parent(s): 2133f67

Upload 7 files

Browse files
Files changed (7) hide show
  1. config.json +27 -0
  2. optimizer.pt +3 -0
  3. pytorch_model.bin +3 -0
  4. rng_state.pth +3 -0
  5. scheduler.pt +3 -0
  6. trainer_state.json +140 -0
  7. training_args.bin +3 -0
config.json ADDED
@@ -0,0 +1,27 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "roberta-large",
3
+ "architectures": [
4
+ "RobertaForMaskedLM"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "bos_token_id": 0,
8
+ "classifier_dropout": null,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 1024,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 4096,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 514,
17
+ "model_type": "roberta",
18
+ "num_attention_heads": 16,
19
+ "num_hidden_layers": 24,
20
+ "pad_token_id": 1,
21
+ "position_embedding_type": "absolute",
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.30.2",
24
+ "type_vocab_size": 1,
25
+ "use_cache": true,
26
+ "vocab_size": 50265
27
+ }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a514c11c21646f7b9398f826c6e575cb417aaeb1d339b2b0fff372336c9ad6c2
3
+ size 2843527877
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c5a1e7653b1337ea34c8ba6221c70affd2c2392d066a6e50ea6898a64d81112
3
+ size 1421788537
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:688ee49884c0bfeb340b3679220313b2c39723e036d4dd745e0015c79dd40cd3
3
+ size 14575
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:141aa6c5621998cba44fdb6c642b02c82163a2385db36dc8e9e42187193a7b25
3
+ size 627
trainer_state.json ADDED
@@ -0,0 +1,140 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.00012142229388700798,
3
+ "best_model_checkpoint": "data/further_pretraining/roberta-large/checkpoint-9356",
4
+ "epoch": 2.0,
5
+ "global_step": 9356,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.11,
12
+ "learning_rate": 1.9786233433091067e-05,
13
+ "loss": 0.0753,
14
+ "step": 500
15
+ },
16
+ {
17
+ "epoch": 0.21,
18
+ "learning_rate": 1.9572466866182133e-05,
19
+ "loss": 0.0589,
20
+ "step": 1000
21
+ },
22
+ {
23
+ "epoch": 0.32,
24
+ "learning_rate": 1.9358700299273195e-05,
25
+ "loss": 0.0558,
26
+ "step": 1500
27
+ },
28
+ {
29
+ "epoch": 0.43,
30
+ "learning_rate": 1.914493373236426e-05,
31
+ "loss": 0.0551,
32
+ "step": 2000
33
+ },
34
+ {
35
+ "epoch": 0.53,
36
+ "learning_rate": 1.8931167165455325e-05,
37
+ "loss": 0.0541,
38
+ "step": 2500
39
+ },
40
+ {
41
+ "epoch": 0.64,
42
+ "learning_rate": 1.8717400598546387e-05,
43
+ "loss": 0.0385,
44
+ "step": 3000
45
+ },
46
+ {
47
+ "epoch": 0.75,
48
+ "learning_rate": 1.8503634031637453e-05,
49
+ "loss": 0.0123,
50
+ "step": 3500
51
+ },
52
+ {
53
+ "epoch": 0.86,
54
+ "learning_rate": 1.828986746472852e-05,
55
+ "loss": 0.0037,
56
+ "step": 4000
57
+ },
58
+ {
59
+ "epoch": 0.96,
60
+ "learning_rate": 1.807610089781958e-05,
61
+ "loss": 0.0029,
62
+ "step": 4500
63
+ },
64
+ {
65
+ "epoch": 1.0,
66
+ "eval_loss": 0.0009945582132786512,
67
+ "eval_runtime": 755.7373,
68
+ "eval_samples_per_second": 99.032,
69
+ "eval_steps_per_second": 6.19,
70
+ "step": 4678
71
+ },
72
+ {
73
+ "epoch": 1.07,
74
+ "learning_rate": 1.7862334330910646e-05,
75
+ "loss": 0.0017,
76
+ "step": 5000
77
+ },
78
+ {
79
+ "epoch": 1.18,
80
+ "learning_rate": 1.764856776400171e-05,
81
+ "loss": 0.0011,
82
+ "step": 5500
83
+ },
84
+ {
85
+ "epoch": 1.28,
86
+ "learning_rate": 1.7434801197092777e-05,
87
+ "loss": 0.0136,
88
+ "step": 6000
89
+ },
90
+ {
91
+ "epoch": 1.39,
92
+ "learning_rate": 1.7221034630183842e-05,
93
+ "loss": 0.001,
94
+ "step": 6500
95
+ },
96
+ {
97
+ "epoch": 1.5,
98
+ "learning_rate": 1.7007268063274904e-05,
99
+ "loss": 0.0009,
100
+ "step": 7000
101
+ },
102
+ {
103
+ "epoch": 1.6,
104
+ "learning_rate": 1.679350149636597e-05,
105
+ "loss": 0.0007,
106
+ "step": 7500
107
+ },
108
+ {
109
+ "epoch": 1.71,
110
+ "learning_rate": 1.6579734929457035e-05,
111
+ "loss": 0.0007,
112
+ "step": 8000
113
+ },
114
+ {
115
+ "epoch": 1.82,
116
+ "learning_rate": 1.63659683625481e-05,
117
+ "loss": 0.0013,
118
+ "step": 8500
119
+ },
120
+ {
121
+ "epoch": 1.92,
122
+ "learning_rate": 1.6152201795639163e-05,
123
+ "loss": 0.0012,
124
+ "step": 9000
125
+ },
126
+ {
127
+ "epoch": 2.0,
128
+ "eval_loss": 0.00012142229388700798,
129
+ "eval_runtime": 758.5152,
130
+ "eval_samples_per_second": 98.669,
131
+ "eval_steps_per_second": 6.167,
132
+ "step": 9356
133
+ }
134
+ ],
135
+ "max_steps": 46780,
136
+ "num_train_epochs": 10,
137
+ "total_flos": 3.4879581646470144e+16,
138
+ "trial_name": null,
139
+ "trial_params": null
140
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b5fa4df5c55437df819c9dab235ce56f6f579cc79813737dd030fcb1cdede79
3
+ size 3963