NDugar commited on
Commit
cf07427
·
verified ·
1 Parent(s): cbdb3d2

Upload 47 files

Browse files

Other related trainings and config specs.
Reality: Dumping old files that may be related tbh.

Files changed (47) hide show
  1. v2xl new/added_tokens.json +1 -0
  2. v2xl new/config.json +48 -0
  3. v2xl new/optimizer.pt +3 -0
  4. v2xl new/pytorch_model.bin +3 -0
  5. v2xl new/rng_state.pth +3 -0
  6. v2xl new/scheduler.pt +3 -0
  7. v2xl new/special_tokens_map.json +1 -0
  8. v2xl new/spm.model +3 -0
  9. v2xl new/tokenizer_config.json +1 -0
  10. v2xl new/trainer_state.json +142 -0
  11. v2xl new/training_args.bin +3 -0
  12. v2xl/added_tokens.json +1 -0
  13. v2xl/config.json +48 -0
  14. v2xl/optimizer.pt +3 -0
  15. v2xl/pytorch_model.bin +3 -0
  16. v2xl/rng_state.pth +3 -0
  17. v2xl/scaler.pt +3 -0
  18. v2xl/scheduler.pt +3 -0
  19. v2xl/special_tokens_map.json +1 -0
  20. v2xl/spm.model +3 -0
  21. v2xl/tokenizer_config.json +1 -0
  22. v2xl/trainer_state.json +61 -0
  23. v2xl/training_args.bin +3 -0
  24. v3-1 epoch/added_tokens.json +1 -0
  25. v3-1 epoch/config.json +45 -0
  26. v3-1 epoch/optimizer.pt +3 -0
  27. v3-1 epoch/pytorch_model.bin +3 -0
  28. v3-1 epoch/rng_state.pth +3 -0
  29. v3-1 epoch/scaler.pt +3 -0
  30. v3-1 epoch/scheduler.pt +3 -0
  31. v3-1 epoch/special_tokens_map.json +1 -0
  32. v3-1 epoch/spm.model +3 -0
  33. v3-1 epoch/tokenizer_config.json +1 -0
  34. v3-1 epoch/trainer_state.json +31 -0
  35. v3-1 epoch/training_args.bin +3 -0
  36. v3perfect/README.md +77 -0
  37. v3perfect/added_tokens.json +1 -0
  38. v3perfect/all_results.json +14 -0
  39. v3perfect/config.json +46 -0
  40. v3perfect/eval_results.json +9 -0
  41. v3perfect/pytorch_model.bin +3 -0
  42. v3perfect/special_tokens_map.json +1 -0
  43. v3perfect/spm.model +3 -0
  44. v3perfect/tokenizer_config.json +1 -0
  45. v3perfect/train_results.json +8 -0
  46. v3perfect/trainer_state.json +61 -0
  47. v3perfect/training_args.bin +3 -0
v2xl new/added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"[MASK]": 128000}
v2xl new/config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/deberta-v2-xlarge",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_head_size": 64,
7
+ "attention_probs_dropout_prob": 0.1,
8
+ "conv_act": "gelu",
9
+ "conv_kernel_size": 3,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 1536,
13
+ "id2label": {
14
+ "0": "entailment",
15
+ "1": "neutral",
16
+ "2": "contradiction"
17
+ },
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 6144,
20
+ "label2id": {
21
+ "contradiction": 2,
22
+ "entailment": 0,
23
+ "neutral": 1
24
+ },
25
+ "layer_norm_eps": 1e-07,
26
+ "max_position_embeddings": 512,
27
+ "max_relative_positions": -1,
28
+ "model_type": "deberta-v2",
29
+ "norm_rel_ebd": "layer_norm",
30
+ "num_attention_heads": 24,
31
+ "num_hidden_layers": 24,
32
+ "pad_token_id": 0,
33
+ "pooler_dropout": 0,
34
+ "pooler_hidden_act": "gelu",
35
+ "pooler_hidden_size": 1536,
36
+ "pos_att_type": [
37
+ "p2c",
38
+ "c2p"
39
+ ],
40
+ "position_biased_input": false,
41
+ "position_buckets": 256,
42
+ "relative_attention": true,
43
+ "share_att_key": true,
44
+ "torch_dtype": "float32",
45
+ "transformers_version": "4.13.0.dev0",
46
+ "type_vocab_size": 0,
47
+ "vocab_size": 128100
48
+ }
v2xl new/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:530418a25fba745d473ef256d49a8f3cca60ef89ae3debb6381fe55f8390def2
3
+ size 7095909461
v2xl new/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41fdfc7b911784d7dbc7dda1799ccf596dd5f714c996e63be6963589eac6dda7
3
+ size 3548003107
v2xl new/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51eeb1baaf5f33fb8143692004df6c9d9552068e7f238321952b0945ea8ffa6a
3
+ size 14503
v2xl new/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc7dbc6ad7398c51b4230f8710686075fb7cff6597c69383b0209345fdcfba63
3
+ size 623
v2xl new/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
v2xl new/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5598d5e96f339a8d980c15f9afd405a2e5e1be7db41de3ed13b0f03fac1e8c17
3
+ size 2447305
v2xl new/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "split_by_punct": false, "sp_model_kwargs": {}, "vocab_type": "spm", "model_max_length": 512, "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "microsoft/deberta-v2-xlarge", "tokenizer_class": "DebertaV2Tokenizer"}
v2xl new/trainer_state.json ADDED
@@ -0,0 +1,142 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.9068051101949407,
3
+ "best_model_checkpoint": "./resultsv3l32/output/checkpoint-294528",
4
+ "epoch": 4.0,
5
+ "global_step": 392704,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.25,
12
+ "learning_rate": 4.681694100391135e-06,
13
+ "loss": 0.5874,
14
+ "step": 25000
15
+ },
16
+ {
17
+ "epoch": 0.51,
18
+ "learning_rate": 4.363388200782269e-06,
19
+ "loss": 0.5086,
20
+ "step": 50000
21
+ },
22
+ {
23
+ "epoch": 0.76,
24
+ "learning_rate": 4.045082301173403e-06,
25
+ "loss": 0.48,
26
+ "step": 75000
27
+ },
28
+ {
29
+ "epoch": 1.0,
30
+ "eval_accuracy": 0.9011044943248333,
31
+ "eval_loss": 0.47157734632492065,
32
+ "eval_runtime": 213.0954,
33
+ "eval_samples_per_second": 92.198,
34
+ "eval_steps_per_second": 23.051,
35
+ "step": 98176
36
+ },
37
+ {
38
+ "epoch": 1.02,
39
+ "learning_rate": 3.7267764015645376e-06,
40
+ "loss": 0.4559,
41
+ "step": 100000
42
+ },
43
+ {
44
+ "epoch": 1.27,
45
+ "learning_rate": 3.408470501955672e-06,
46
+ "loss": 0.3309,
47
+ "step": 125000
48
+ },
49
+ {
50
+ "epoch": 1.53,
51
+ "learning_rate": 3.0901646023468064e-06,
52
+ "loss": 0.3398,
53
+ "step": 150000
54
+ },
55
+ {
56
+ "epoch": 1.78,
57
+ "learning_rate": 2.77185870273794e-06,
58
+ "loss": 0.3382,
59
+ "step": 175000
60
+ },
61
+ {
62
+ "epoch": 2.0,
63
+ "eval_accuracy": 0.9044128874637349,
64
+ "eval_loss": 0.5638315081596375,
65
+ "eval_runtime": 212.5015,
66
+ "eval_samples_per_second": 92.456,
67
+ "eval_steps_per_second": 23.115,
68
+ "step": 196352
69
+ },
70
+ {
71
+ "epoch": 2.04,
72
+ "learning_rate": 2.4535528031290744e-06,
73
+ "loss": 0.3082,
74
+ "step": 200000
75
+ },
76
+ {
77
+ "epoch": 2.29,
78
+ "learning_rate": 2.135246903520209e-06,
79
+ "loss": 0.1952,
80
+ "step": 225000
81
+ },
82
+ {
83
+ "epoch": 2.55,
84
+ "learning_rate": 1.816941003911343e-06,
85
+ "loss": 0.2029,
86
+ "step": 250000
87
+ },
88
+ {
89
+ "epoch": 2.8,
90
+ "learning_rate": 1.4986351043024774e-06,
91
+ "loss": 0.2027,
92
+ "step": 275000
93
+ },
94
+ {
95
+ "epoch": 3.0,
96
+ "eval_accuracy": 0.9068051101949407,
97
+ "eval_loss": 0.6681745052337646,
98
+ "eval_runtime": 212.8042,
99
+ "eval_samples_per_second": 92.324,
100
+ "eval_steps_per_second": 23.082,
101
+ "step": 294528
102
+ },
103
+ {
104
+ "epoch": 3.06,
105
+ "learning_rate": 1.1803292046936116e-06,
106
+ "loss": 0.1771,
107
+ "step": 300000
108
+ },
109
+ {
110
+ "epoch": 3.31,
111
+ "learning_rate": 8.620233050847459e-07,
112
+ "loss": 0.1124,
113
+ "step": 325000
114
+ },
115
+ {
116
+ "epoch": 3.57,
117
+ "learning_rate": 5.437174054758801e-07,
118
+ "loss": 0.1095,
119
+ "step": 350000
120
+ },
121
+ {
122
+ "epoch": 3.82,
123
+ "learning_rate": 2.2541150586701436e-07,
124
+ "loss": 0.1038,
125
+ "step": 375000
126
+ },
127
+ {
128
+ "epoch": 4.0,
129
+ "eval_accuracy": 0.9064997200590421,
130
+ "eval_loss": 0.7638580799102783,
131
+ "eval_runtime": 211.2725,
132
+ "eval_samples_per_second": 92.994,
133
+ "eval_steps_per_second": 23.25,
134
+ "step": 392704
135
+ }
136
+ ],
137
+ "max_steps": 392704,
138
+ "num_train_epochs": 4,
139
+ "total_flos": 3.755067886948521e+17,
140
+ "trial_name": null,
141
+ "trial_params": null
142
+ }
v2xl new/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:937df70044f985ff8dee511a2bc22fca38bfcadff7c2ef74fd252af4bd7e9dd2
3
+ size 2799
v2xl/added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"[MASK]": 128000}
v2xl/config.json ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/deberta-v2-xlarge",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_head_size": 64,
7
+ "attention_probs_dropout_prob": 0.1,
8
+ "conv_act": "gelu",
9
+ "conv_kernel_size": 3,
10
+ "hidden_act": "gelu",
11
+ "hidden_dropout_prob": 0.1,
12
+ "hidden_size": 1536,
13
+ "id2label": {
14
+ "0": "entailment",
15
+ "1": "neutral",
16
+ "2": "contradiction"
17
+ },
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 6144,
20
+ "label2id": {
21
+ "contradiction": 2,
22
+ "entailment": 0,
23
+ "neutral": 1
24
+ },
25
+ "layer_norm_eps": 1e-07,
26
+ "max_position_embeddings": 512,
27
+ "max_relative_positions": -1,
28
+ "model_type": "deberta-v2",
29
+ "norm_rel_ebd": "layer_norm",
30
+ "num_attention_heads": 24,
31
+ "num_hidden_layers": 24,
32
+ "pad_token_id": 0,
33
+ "pooler_dropout": 0,
34
+ "pooler_hidden_act": "gelu",
35
+ "pooler_hidden_size": 1536,
36
+ "pos_att_type": [
37
+ "p2c",
38
+ "c2p"
39
+ ],
40
+ "position_biased_input": false,
41
+ "position_buckets": 256,
42
+ "relative_attention": true,
43
+ "share_att_key": true,
44
+ "torch_dtype": "float32",
45
+ "transformers_version": "4.13.0.dev0",
46
+ "type_vocab_size": 0,
47
+ "vocab_size": 128100
48
+ }
v2xl/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aea3c98817c299ee60aacdebd14f4a08035dbe524378bcf64407b3e07da43d90
3
+ size 7095909461
v2xl/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63a1f786762c51cfec4d1514e2a566964637358788a892dbd9592dc4123ac1cc
3
+ size 3548003107
v2xl/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83f2eb52ce3fdf6765605cdca2051f8bbf6ea652fc1f29b5232a2dd3f0a25408
3
+ size 14503
v2xl/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1930bfa221eacebdb1f3b51ae8fa0d37b8a67db926a6a1c3b36f1929a2861dc1
3
+ size 559
v2xl/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b0b995dc2f9ae2800cc5b64a6adfaccac0ce1dea781c260e20894f495b81f6f
3
+ size 623
v2xl/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
v2xl/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5598d5e96f339a8d980c15f9afd405a2e5e1be7db41de3ed13b0f03fac1e8c17
3
+ size 2447305
v2xl/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "split_by_punct": false, "sp_model_kwargs": {}, "vocab_type": "spm", "model_max_length": 512, "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "microsoft/deberta-v2-xlarge", "tokenizer_class": "DebertaV2Tokenizer"}
v2xl/trainer_state.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.9058380414312618,
3
+ "best_model_checkpoint": "./resultsv2xl/output/checkpoint-294528",
4
+ "epoch": 3.0,
5
+ "global_step": 294528,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 1.0,
12
+ "learning_rate": 4.2562126159447515e-06,
13
+ "loss": 0.527,
14
+ "step": 98176
15
+ },
16
+ {
17
+ "epoch": 1.0,
18
+ "eval_accuracy": 0.8983559831017458,
19
+ "eval_loss": 0.4922527074813843,
20
+ "eval_runtime": 231.3267,
21
+ "eval_samples_per_second": 84.932,
22
+ "eval_steps_per_second": 21.234,
23
+ "step": 98176
24
+ },
25
+ {
26
+ "epoch": 2.0,
27
+ "learning_rate": 2.1293307712312536e-06,
28
+ "loss": 0.3335,
29
+ "step": 196352
30
+ },
31
+ {
32
+ "epoch": 2.0,
33
+ "eval_accuracy": 0.9027332417162925,
34
+ "eval_loss": 0.5514769554138184,
35
+ "eval_runtime": 231.2763,
36
+ "eval_samples_per_second": 84.95,
37
+ "eval_steps_per_second": 21.239,
38
+ "step": 196352
39
+ },
40
+ {
41
+ "epoch": 3.0,
42
+ "learning_rate": 2.4705984338428644e-09,
43
+ "loss": 0.1729,
44
+ "step": 294528
45
+ },
46
+ {
47
+ "epoch": 3.0,
48
+ "eval_accuracy": 0.9058380414312618,
49
+ "eval_loss": 0.65480637550354,
50
+ "eval_runtime": 230.3593,
51
+ "eval_samples_per_second": 85.289,
52
+ "eval_steps_per_second": 21.323,
53
+ "step": 294528
54
+ }
55
+ ],
56
+ "max_steps": 294528,
57
+ "num_train_epochs": 3,
58
+ "total_flos": 2.8162430874533802e+17,
59
+ "trial_name": null,
60
+ "trial_params": null
61
+ }
v2xl/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49fc87ca654f3e63fe7381a5b9e1d9f73d219885e1a7a116dbcf2a749c415f4b
3
+ size 2799
v3-1 epoch/added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"[MASK]": 128000}
v3-1 epoch/config.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/deberta-v3-large",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "hidden_act": "gelu",
8
+ "hidden_dropout_prob": 0.1,
9
+ "hidden_size": 1024,
10
+ "id2label": {
11
+ "0": "entailment",
12
+ "1": "neutral",
13
+ "2": "contradiction"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 4096,
17
+ "label2id": {
18
+ "contradiction": 2,
19
+ "entailment": 0,
20
+ "neutral": 1
21
+ },
22
+ "layer_norm_eps": 1e-07,
23
+ "max_position_embeddings": 512,
24
+ "max_relative_positions": -1,
25
+ "model_type": "deberta-v2",
26
+ "norm_rel_ebd": "layer_norm",
27
+ "num_attention_heads": 16,
28
+ "num_hidden_layers": 24,
29
+ "pad_token_id": 0,
30
+ "pooler_dropout": 0,
31
+ "pooler_hidden_act": "gelu",
32
+ "pooler_hidden_size": 1024,
33
+ "pos_att_type": [
34
+ "p2c",
35
+ "c2p"
36
+ ],
37
+ "position_biased_input": false,
38
+ "position_buckets": 256,
39
+ "relative_attention": true,
40
+ "share_att_key": true,
41
+ "torch_dtype": "float32",
42
+ "transformers_version": "4.13.0.dev0",
43
+ "type_vocab_size": 0,
44
+ "vocab_size": 128100
45
+ }
v3-1 epoch/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fc477f55816300ce73341afa1829884ca66d85b8ad8bf2184a39105a9e0ff0c
3
+ size 3480748761
v3-1 epoch/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ced5efa4b8b24ddeeba8b0d5884ac47af7b00aaf5fe0ea97be2a53ea418301b9
3
+ size 1740426539
v3-1 epoch/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c72fe6cbb3a9729e8eaec990f08c6133b4ee2ddd44df6ca26274ba6bb751261
3
+ size 14503
v3-1 epoch/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7b03ed43bb905000173ec810b3d544870bb440ffcecdfbe8c5f2bdc77248135
3
+ size 559
v3-1 epoch/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49da2828958bbc3229278c6d041c95960908ca65be2589040e4c80abbee39bde
3
+ size 623
v3-1 epoch/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
v3-1 epoch/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
v3-1 epoch/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "split_by_punct": false, "sp_model_kwargs": {}, "vocab_type": "spm", "model_max_length": 512, "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "microsoft/deberta-v3-large", "tokenizer_class": "DebertaV2Tokenizer"}
v3-1 epoch/trainer_state.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.9100117066218761,
3
+ "best_model_checkpoint": "./resultsv3p2/output/checkpoint-98176",
4
+ "epoch": 1.0,
5
+ "global_step": 98176,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 1.0,
12
+ "learning_rate": 3.1928048978707266e-06,
13
+ "loss": 0.4695,
14
+ "step": 98176
15
+ },
16
+ {
17
+ "epoch": 1.0,
18
+ "eval_accuracy": 0.9100117066218761,
19
+ "eval_loss": 0.3915119469165802,
20
+ "eval_runtime": 352.6229,
21
+ "eval_samples_per_second": 55.717,
22
+ "eval_steps_per_second": 13.93,
23
+ "step": 98176
24
+ }
25
+ ],
26
+ "max_steps": 196352,
27
+ "num_train_epochs": 2,
28
+ "total_flos": 4.154826179696498e+16,
29
+ "trial_name": null,
30
+ "trial_params": null
31
+ }
v3-1 epoch/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9409407acd571c783e5eb22459059a64f9129882be790fb39c9ffdda3344fd0b
3
+ size 2799
v3perfect/README.md ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: mit
5
+ tags:
6
+ - generated_from_trainer
7
+ datasets:
8
+ - glue
9
+ metrics:
10
+ - accuracy
11
+ model-index:
12
+ - name: mnlilearn
13
+ results:
14
+ - task:
15
+ name: Text Classification
16
+ type: text-classification
17
+ dataset:
18
+ name: GLUE MNLI
19
+ type: glue
20
+ args: mnli
21
+ metrics:
22
+ - name: Accuracy
23
+ type: accuracy
24
+ value: 0.9175142392188771
25
+ ---
26
+
27
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
28
+ should probably proofread and complete it, then remove this comment. -->
29
+
30
+ # mnlilearn
31
+
32
+ This model is a fine-tuned version of [microsoft/deberta-v3-large](https://huggingface.co/microsoft/deberta-v3-large) on the GLUE MNLI dataset.
33
+ It achieves the following results on the evaluation set:
34
+ - Loss: 0.4103
35
+ - Accuracy: 0.9175
36
+
37
+ ## Model description
38
+
39
+ More information needed
40
+
41
+ ## Intended uses & limitations
42
+
43
+ More information needed
44
+
45
+ ## Training and evaluation data
46
+
47
+ More information needed
48
+
49
+ ## Training procedure
50
+
51
+ ### Training hyperparameters
52
+
53
+ The following hyperparameters were used during training:
54
+ - learning_rate: 6e-06
55
+ - train_batch_size: 8
56
+ - eval_batch_size: 8
57
+ - seed: 42
58
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
59
+ - lr_scheduler_type: linear
60
+ - lr_scheduler_warmup_steps: 50
61
+ - num_epochs: 2.0
62
+
63
+ ### Training results
64
+
65
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
66
+ |:-------------:|:-----:|:-----:|:---------------:|:--------:|
67
+ | 0.3631 | 1.0 | 49088 | 0.3129 | 0.9130 |
68
+ | 0.2267 | 2.0 | 98176 | 0.4157 | 0.9153 |
69
+
70
+
71
+ ### Framework versions
72
+
73
+ - Transformers 4.13.0.dev0
74
+ - Pytorch 1.10.0
75
+ - Datasets 1.15.2.dev0
76
+ - Tokenizers 0.10.3
77
+ NDugar/v3-Large-mnli
v3perfect/added_tokens.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"[MASK]": 128000}
v3perfect/all_results.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.0,
3
+ "eval_accuracy": 0.9175142392188771,
4
+ "eval_loss": 0.41033002734184265,
5
+ "eval_runtime": 78.2213,
6
+ "eval_samples": 9832,
7
+ "eval_samples_per_second": 125.695,
8
+ "eval_steps_per_second": 15.712,
9
+ "train_loss": 0.2868618026395198,
10
+ "train_runtime": 21298.9707,
11
+ "train_samples": 392702,
12
+ "train_samples_per_second": 36.875,
13
+ "train_steps_per_second": 4.609
14
+ }
v3perfect/config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "microsoft/deberta-v3-large",
3
+ "architectures": [
4
+ "DebertaV2ForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "finetuning_task": "mnli",
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 1024,
11
+ "id2label": {
12
+ "0": "entailment",
13
+ "1": "neutral",
14
+ "2": "contradiction"
15
+ },
16
+ "initializer_range": 0.02,
17
+ "intermediate_size": 4096,
18
+ "label2id": {
19
+ "contradiction": 2,
20
+ "entailment": 0,
21
+ "neutral": 1
22
+ },
23
+ "layer_norm_eps": 1e-07,
24
+ "max_position_embeddings": 512,
25
+ "max_relative_positions": -1,
26
+ "model_type": "deberta-v2",
27
+ "norm_rel_ebd": "layer_norm",
28
+ "num_attention_heads": 16,
29
+ "num_hidden_layers": 24,
30
+ "pad_token_id": 0,
31
+ "pooler_dropout": 0,
32
+ "pooler_hidden_act": "gelu",
33
+ "pooler_hidden_size": 1024,
34
+ "pos_att_type": [
35
+ "p2c",
36
+ "c2p"
37
+ ],
38
+ "position_biased_input": false,
39
+ "position_buckets": 256,
40
+ "relative_attention": true,
41
+ "share_att_key": true,
42
+ "torch_dtype": "float32",
43
+ "transformers_version": "4.13.0.dev0",
44
+ "type_vocab_size": 0,
45
+ "vocab_size": 128100
46
+ }
v3perfect/eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.0,
3
+ "eval_accuracy": 0.9175142392188771,
4
+ "eval_loss": 0.41033002734184265,
5
+ "eval_runtime": 78.2213,
6
+ "eval_samples": 9832,
7
+ "eval_samples_per_second": 125.695,
8
+ "eval_steps_per_second": 15.712
9
+ }
v3perfect/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2723410e9906cee3b3636c827dcdd99e3badcb98768693a5dd20f6b4b7d8c91
3
+ size 1740426539
v3perfect/special_tokens_map.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]"}
v3perfect/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
v3perfect/tokenizer_config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"do_lower_case": false, "bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "split_by_punct": false, "sp_model_kwargs": {}, "vocab_type": "spm", "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "microsoft/deberta-v3-large", "tokenizer_class": "DebertaV2Tokenizer"}
v3perfect/train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 2.0,
3
+ "train_loss": 0.2868618026395198,
4
+ "train_runtime": 21298.9707,
5
+ "train_samples": 392702,
6
+ "train_samples_per_second": 36.875,
7
+ "train_steps_per_second": 4.609
8
+ }
v3perfect/trainer_state.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 2.0,
5
+ "global_step": 98176,
6
+ "is_hyper_param_search": false,
7
+ "is_local_process_zero": true,
8
+ "is_world_process_zero": true,
9
+ "log_history": [
10
+ {
11
+ "epoch": 0.61,
12
+ "learning_rate": 4.168681083504881e-06,
13
+ "loss": 0.3631,
14
+ "step": 30000
15
+ },
16
+ {
17
+ "epoch": 1.0,
18
+ "eval_accuracy": 0.9129903209373408,
19
+ "eval_loss": 0.31294044852256775,
20
+ "eval_runtime": 78.2833,
21
+ "eval_samples_per_second": 125.378,
22
+ "eval_steps_per_second": 15.674,
23
+ "step": 49088
24
+ },
25
+ {
26
+ "epoch": 1.22,
27
+ "learning_rate": 2.334304873326132e-06,
28
+ "loss": 0.2901,
29
+ "step": 60000
30
+ },
31
+ {
32
+ "epoch": 1.83,
33
+ "learning_rate": 4.999286631473819e-07,
34
+ "loss": 0.2267,
35
+ "step": 90000
36
+ },
37
+ {
38
+ "epoch": 2.0,
39
+ "eval_accuracy": 0.915333672949567,
40
+ "eval_loss": 0.41572055220603943,
41
+ "eval_runtime": 78.1399,
42
+ "eval_samples_per_second": 125.608,
43
+ "eval_steps_per_second": 15.703,
44
+ "step": 98176
45
+ },
46
+ {
47
+ "epoch": 2.0,
48
+ "step": 98176,
49
+ "total_flos": 3.65975017449044e+17,
50
+ "train_loss": 0.2868618026395198,
51
+ "train_runtime": 21298.9707,
52
+ "train_samples_per_second": 36.875,
53
+ "train_steps_per_second": 4.609
54
+ }
55
+ ],
56
+ "max_steps": 98176,
57
+ "num_train_epochs": 2,
58
+ "total_flos": 3.65975017449044e+17,
59
+ "trial_name": null,
60
+ "trial_params": null
61
+ }
v3perfect/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fe31ac00049c414b30f35dccac26a8c9e1eb745034ce6fbe76c79bad8d698e01
3
+ size 2799