NuNER_Zero_2025-02-17_12-11-11

#1
BSC-NLP4BIA/chagas-ner/README.md ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - generated_from_trainer
5
+ model-index:
6
+ - name: NuNER_Zero_2025-02-17_12-11-11
7
+ results: []
8
+ ---
9
+
10
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
+ should probably proofread and complete it, then remove this comment. -->
12
+
13
+ # NuNER_Zero_2025-02-17_12-11-11
14
+
15
+ This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
16
+ It achieves the following results on the evaluation set:
17
+ - Loss: 37.8851
18
+
19
+ ## Model description
20
+
21
+ More information needed
22
+
23
+ ## Intended uses & limitations
24
+
25
+ More information needed
26
+
27
+ ## Training and evaluation data
28
+
29
+ More information needed
30
+
31
+ ## Training procedure
32
+
33
+ ### Training hyperparameters
34
+
35
+ The following hyperparameters were used during training:
36
+ - learning_rate: 5e-06
37
+ - train_batch_size: 2
38
+ - eval_batch_size: 2
39
+ - seed: 42
40
+ - gradient_accumulation_steps: 8
41
+ - total_train_batch_size: 16
42
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
+ - lr_scheduler_type: linear
44
+ - lr_scheduler_warmup_ratio: 0.1
45
+ - num_epochs: 4
46
+
47
+ ### Training results
48
+
49
+ | Training Loss | Epoch | Step | Validation Loss |
50
+ |:-------------:|:------:|:----:|:---------------:|
51
+ | 30.303 | 0.6584 | 100 | 76.8218 |
52
+ | 4.6085 | 1.3226 | 200 | 57.4994 |
53
+ | 3.7912 | 1.9811 | 300 | 49.8785 |
54
+ | 3.0424 | 2.6453 | 400 | 41.6180 |
55
+ | 2.8036 | 3.3095 | 500 | 39.0891 |
56
+ | 2.5307 | 3.9679 | 600 | 37.8851 |
57
+
58
+
59
+ ### Framework versions
60
+
61
+ - Transformers 4.48.3
62
+ - Pytorch 2.5.1+cu124
63
+ - Datasets 3.3.0
64
+ - Tokenizers 0.21.0
BSC-NLP4BIA/chagas-ner/added_tokens.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "<<ENT>>": 128002,
3
+ "<<SEP>>": 128003,
4
+ "[FLERT]": 128001,
5
+ "[MASK]": 128000
6
+ }
BSC-NLP4BIA/chagas-ner/checkpoint-604/added_tokens.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "<<ENT>>": 128002,
3
+ "<<SEP>>": 128003,
4
+ "[FLERT]": 128001,
5
+ "[MASK]": 128000
6
+ }
BSC-NLP4BIA/chagas-ner/checkpoint-604/gliner_config.json ADDED
@@ -0,0 +1,131 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "class_token_index": 128002,
3
+ "dropout": 0.4,
4
+ "embed_ent_token": true,
5
+ "encoder_config": {
6
+ "_attn_implementation_autoset": false,
7
+ "_name_or_path": "microsoft/deberta-v3-large",
8
+ "add_cross_attention": false,
9
+ "architectures": null,
10
+ "attention_probs_dropout_prob": 0.1,
11
+ "bad_words_ids": null,
12
+ "begin_suppress_tokens": null,
13
+ "bos_token_id": null,
14
+ "chunk_size_feed_forward": 0,
15
+ "cross_attention_hidden_size": null,
16
+ "decoder_start_token_id": null,
17
+ "diversity_penalty": 0.0,
18
+ "do_sample": false,
19
+ "early_stopping": false,
20
+ "encoder_no_repeat_ngram_size": 0,
21
+ "eos_token_id": null,
22
+ "exponential_decay_length_penalty": null,
23
+ "finetuning_task": null,
24
+ "forced_bos_token_id": null,
25
+ "forced_eos_token_id": null,
26
+ "hidden_act": "gelu",
27
+ "hidden_dropout_prob": 0.1,
28
+ "hidden_size": 1024,
29
+ "id2label": {
30
+ "0": "LABEL_0",
31
+ "1": "LABEL_1"
32
+ },
33
+ "initializer_range": 0.02,
34
+ "intermediate_size": 4096,
35
+ "is_decoder": false,
36
+ "is_encoder_decoder": false,
37
+ "label2id": {
38
+ "LABEL_0": 0,
39
+ "LABEL_1": 1
40
+ },
41
+ "layer_norm_eps": 1e-07,
42
+ "legacy": true,
43
+ "length_penalty": 1.0,
44
+ "max_length": 20,
45
+ "max_position_embeddings": 512,
46
+ "max_relative_positions": -1,
47
+ "min_length": 0,
48
+ "model_type": "deberta-v2",
49
+ "no_repeat_ngram_size": 0,
50
+ "norm_rel_ebd": "layer_norm",
51
+ "num_attention_heads": 16,
52
+ "num_beam_groups": 1,
53
+ "num_beams": 1,
54
+ "num_hidden_layers": 24,
55
+ "num_return_sequences": 1,
56
+ "output_attentions": false,
57
+ "output_hidden_states": false,
58
+ "output_scores": false,
59
+ "pad_token_id": 0,
60
+ "pooler_dropout": 0,
61
+ "pooler_hidden_act": "gelu",
62
+ "pooler_hidden_size": 1024,
63
+ "pos_att_type": [
64
+ "p2c",
65
+ "c2p"
66
+ ],
67
+ "position_biased_input": false,
68
+ "position_buckets": 256,
69
+ "prefix": null,
70
+ "problem_type": null,
71
+ "pruned_heads": {},
72
+ "relative_attention": true,
73
+ "remove_invalid_values": false,
74
+ "repetition_penalty": 1.0,
75
+ "return_dict": true,
76
+ "return_dict_in_generate": false,
77
+ "sep_token_id": null,
78
+ "share_att_key": true,
79
+ "suppress_tokens": null,
80
+ "task_specific_params": null,
81
+ "temperature": 1.0,
82
+ "tf_legacy_loss": false,
83
+ "tie_encoder_decoder": false,
84
+ "tie_word_embeddings": true,
85
+ "tokenizer_class": null,
86
+ "top_k": 50,
87
+ "top_p": 1.0,
88
+ "torch_dtype": null,
89
+ "torchscript": false,
90
+ "type_vocab_size": 0,
91
+ "typical_p": 1.0,
92
+ "use_bfloat16": false,
93
+ "vocab_size": 128004
94
+ },
95
+ "ent_token": "<<ENT>>",
96
+ "eval_every": 2500,
97
+ "fine_tune": true,
98
+ "fuse_layers": false,
99
+ "gradient_accumulation_steps": 2,
100
+ "has_rnn": true,
101
+ "hidden_size": 768,
102
+ "labels_encoder": null,
103
+ "labels_encoder_config": null,
104
+ "log_dir": "logs",
105
+ "lr_encoder": "1e-5",
106
+ "lr_others": "5e-5",
107
+ "max_len": 384,
108
+ "max_neg_type_ratio": 1,
109
+ "max_types": 25,
110
+ "max_width": 1,
111
+ "model_name": "microsoft/deberta-v3-large",
112
+ "model_type": "gliner",
113
+ "name": "large",
114
+ "num_post_fusion_layers": 1,
115
+ "num_steps": 60000,
116
+ "post_fusion_schema": "",
117
+ "prev_path": "none",
118
+ "random_drop": true,
119
+ "root_dir": "ablation_backbone",
120
+ "sep_token": "<<SEP>>",
121
+ "shuffle_types": true,
122
+ "size_sup": -1,
123
+ "span_mode": "marker",
124
+ "subtoken_pooling": "first",
125
+ "train_batch_size": 4,
126
+ "train_data": "NuMinds_custom_data_mix.json",
127
+ "transformers_version": "4.48.3",
128
+ "vocab_size": 128004,
129
+ "warmup_ratio": 0.1,
130
+ "words_splitter_type": "whitespace"
131
+ }
BSC-NLP4BIA/chagas-ner/checkpoint-604/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ebf9c88e34b1d1a0dd8696140ce7b3276fb9014de15e4f326e28d4644cfa64f
3
+ size 3591324460
BSC-NLP4BIA/chagas-ner/checkpoint-604/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45c6fc56939820b78fda03d7dd5766055c6ca792e335743ef685564a446de6fb
3
+ size 1795640934
BSC-NLP4BIA/chagas-ner/checkpoint-604/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f6fe7b680821a170330572b71670a5d386db44084f7c16f24d42279fadfecdc
3
+ size 14244
BSC-NLP4BIA/chagas-ner/checkpoint-604/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e835f7f6bc97246ef77d213f5fc8d9d39c873b4d19ec47af4667388d9081ca4c
3
+ size 1064
BSC-NLP4BIA/chagas-ner/checkpoint-604/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
BSC-NLP4BIA/chagas-ner/checkpoint-604/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
BSC-NLP4BIA/chagas-ner/checkpoint-604/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6dba35d5dbe252aeafc5330bb2c1dbccd1632c829ea62a142b4082179db3fd94
3
+ size 8657331
BSC-NLP4BIA/chagas-ner/checkpoint-604/tokenizer_config.json ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "128001": {
44
+ "content": "[FLERT]",
45
+ "lstrip": false,
46
+ "normalized": true,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": false
50
+ },
51
+ "128002": {
52
+ "content": "<<ENT>>",
53
+ "lstrip": false,
54
+ "normalized": true,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": false
58
+ },
59
+ "128003": {
60
+ "content": "<<SEP>>",
61
+ "lstrip": false,
62
+ "normalized": true,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": false
66
+ }
67
+ },
68
+ "bos_token": "[CLS]",
69
+ "clean_up_tokenization_spaces": false,
70
+ "cls_token": "[CLS]",
71
+ "do_lower_case": false,
72
+ "eos_token": "[SEP]",
73
+ "extra_special_tokens": {},
74
+ "mask_token": "[MASK]",
75
+ "model_max_length": 1000000000000000019884624838656,
76
+ "pad_token": "[PAD]",
77
+ "sep_token": "[SEP]",
78
+ "sp_model_kwargs": {},
79
+ "split_by_punct": false,
80
+ "tokenizer_class": "DebertaV2Tokenizer",
81
+ "unk_token": "[UNK]",
82
+ "vocab_type": "spm"
83
+ }
BSC-NLP4BIA/chagas-ner/checkpoint-604/trainer_state.json ADDED
@@ -0,0 +1,123 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 3.994238683127572,
5
+ "eval_steps": 100,
6
+ "global_step": 604,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.6584362139917695,
13
+ "grad_norm": 317.5823974609375,
14
+ "learning_rate": 9.281767955801105e-06,
15
+ "loss": 30.303,
16
+ "step": 100
17
+ },
18
+ {
19
+ "epoch": 0.6584362139917695,
20
+ "eval_loss": 76.82180786132812,
21
+ "eval_runtime": 20.1289,
22
+ "eval_samples_per_second": 13.265,
23
+ "eval_steps_per_second": 6.657,
24
+ "step": 100
25
+ },
26
+ {
27
+ "epoch": 1.322633744855967,
28
+ "grad_norm": 779.8876342773438,
29
+ "learning_rate": 7.440147329650093e-06,
30
+ "loss": 4.6085,
31
+ "step": 200
32
+ },
33
+ {
34
+ "epoch": 1.322633744855967,
35
+ "eval_loss": 57.49937438964844,
36
+ "eval_runtime": 20.0303,
37
+ "eval_samples_per_second": 13.33,
38
+ "eval_steps_per_second": 6.69,
39
+ "step": 200
40
+ },
41
+ {
42
+ "epoch": 1.9810699588477365,
43
+ "grad_norm": 74.9585952758789,
44
+ "learning_rate": 5.59852670349908e-06,
45
+ "loss": 3.7912,
46
+ "step": 300
47
+ },
48
+ {
49
+ "epoch": 1.9810699588477365,
50
+ "eval_loss": 49.87845993041992,
51
+ "eval_runtime": 20.1122,
52
+ "eval_samples_per_second": 13.276,
53
+ "eval_steps_per_second": 6.663,
54
+ "step": 300
55
+ },
56
+ {
57
+ "epoch": 2.645267489711934,
58
+ "grad_norm": 734.7151489257812,
59
+ "learning_rate": 3.7569060773480665e-06,
60
+ "loss": 3.0424,
61
+ "step": 400
62
+ },
63
+ {
64
+ "epoch": 2.645267489711934,
65
+ "eval_loss": 41.61801528930664,
66
+ "eval_runtime": 20.2123,
67
+ "eval_samples_per_second": 13.21,
68
+ "eval_steps_per_second": 6.63,
69
+ "step": 400
70
+ },
71
+ {
72
+ "epoch": 3.309465020576132,
73
+ "grad_norm": 246.29994201660156,
74
+ "learning_rate": 1.9152854511970537e-06,
75
+ "loss": 2.8036,
76
+ "step": 500
77
+ },
78
+ {
79
+ "epoch": 3.309465020576132,
80
+ "eval_loss": 39.08913803100586,
81
+ "eval_runtime": 20.0366,
82
+ "eval_samples_per_second": 13.326,
83
+ "eval_steps_per_second": 6.688,
84
+ "step": 500
85
+ },
86
+ {
87
+ "epoch": 3.9679012345679014,
88
+ "grad_norm": 223.61874389648438,
89
+ "learning_rate": 7.366482504604052e-08,
90
+ "loss": 2.5307,
91
+ "step": 600
92
+ },
93
+ {
94
+ "epoch": 3.9679012345679014,
95
+ "eval_loss": 37.885135650634766,
96
+ "eval_runtime": 20.1417,
97
+ "eval_samples_per_second": 13.256,
98
+ "eval_steps_per_second": 6.653,
99
+ "step": 600
100
+ }
101
+ ],
102
+ "logging_steps": 100,
103
+ "max_steps": 604,
104
+ "num_input_tokens_seen": 0,
105
+ "num_train_epochs": 4,
106
+ "save_steps": 2500,
107
+ "stateful_callbacks": {
108
+ "TrainerControl": {
109
+ "args": {
110
+ "should_epoch_stop": false,
111
+ "should_evaluate": false,
112
+ "should_log": false,
113
+ "should_save": true,
114
+ "should_training_stop": true
115
+ },
116
+ "attributes": {}
117
+ }
118
+ },
119
+ "total_flos": 0.0,
120
+ "train_batch_size": 2,
121
+ "trial_name": null,
122
+ "trial_params": null
123
+ }
BSC-NLP4BIA/chagas-ner/gliner_config.json ADDED
@@ -0,0 +1,131 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "class_token_index": 128002,
3
+ "dropout": 0.4,
4
+ "embed_ent_token": true,
5
+ "encoder_config": {
6
+ "_attn_implementation_autoset": false,
7
+ "_name_or_path": "microsoft/deberta-v3-large",
8
+ "add_cross_attention": false,
9
+ "architectures": null,
10
+ "attention_probs_dropout_prob": 0.1,
11
+ "bad_words_ids": null,
12
+ "begin_suppress_tokens": null,
13
+ "bos_token_id": null,
14
+ "chunk_size_feed_forward": 0,
15
+ "cross_attention_hidden_size": null,
16
+ "decoder_start_token_id": null,
17
+ "diversity_penalty": 0.0,
18
+ "do_sample": false,
19
+ "early_stopping": false,
20
+ "encoder_no_repeat_ngram_size": 0,
21
+ "eos_token_id": null,
22
+ "exponential_decay_length_penalty": null,
23
+ "finetuning_task": null,
24
+ "forced_bos_token_id": null,
25
+ "forced_eos_token_id": null,
26
+ "hidden_act": "gelu",
27
+ "hidden_dropout_prob": 0.1,
28
+ "hidden_size": 1024,
29
+ "id2label": {
30
+ "0": "LABEL_0",
31
+ "1": "LABEL_1"
32
+ },
33
+ "initializer_range": 0.02,
34
+ "intermediate_size": 4096,
35
+ "is_decoder": false,
36
+ "is_encoder_decoder": false,
37
+ "label2id": {
38
+ "LABEL_0": 0,
39
+ "LABEL_1": 1
40
+ },
41
+ "layer_norm_eps": 1e-07,
42
+ "legacy": true,
43
+ "length_penalty": 1.0,
44
+ "max_length": 20,
45
+ "max_position_embeddings": 512,
46
+ "max_relative_positions": -1,
47
+ "min_length": 0,
48
+ "model_type": "deberta-v2",
49
+ "no_repeat_ngram_size": 0,
50
+ "norm_rel_ebd": "layer_norm",
51
+ "num_attention_heads": 16,
52
+ "num_beam_groups": 1,
53
+ "num_beams": 1,
54
+ "num_hidden_layers": 24,
55
+ "num_return_sequences": 1,
56
+ "output_attentions": false,
57
+ "output_hidden_states": false,
58
+ "output_scores": false,
59
+ "pad_token_id": 0,
60
+ "pooler_dropout": 0,
61
+ "pooler_hidden_act": "gelu",
62
+ "pooler_hidden_size": 1024,
63
+ "pos_att_type": [
64
+ "p2c",
65
+ "c2p"
66
+ ],
67
+ "position_biased_input": false,
68
+ "position_buckets": 256,
69
+ "prefix": null,
70
+ "problem_type": null,
71
+ "pruned_heads": {},
72
+ "relative_attention": true,
73
+ "remove_invalid_values": false,
74
+ "repetition_penalty": 1.0,
75
+ "return_dict": true,
76
+ "return_dict_in_generate": false,
77
+ "sep_token_id": null,
78
+ "share_att_key": true,
79
+ "suppress_tokens": null,
80
+ "task_specific_params": null,
81
+ "temperature": 1.0,
82
+ "tf_legacy_loss": false,
83
+ "tie_encoder_decoder": false,
84
+ "tie_word_embeddings": true,
85
+ "tokenizer_class": null,
86
+ "top_k": 50,
87
+ "top_p": 1.0,
88
+ "torch_dtype": null,
89
+ "torchscript": false,
90
+ "type_vocab_size": 0,
91
+ "typical_p": 1.0,
92
+ "use_bfloat16": false,
93
+ "vocab_size": 128004
94
+ },
95
+ "ent_token": "<<ENT>>",
96
+ "eval_every": 2500,
97
+ "fine_tune": true,
98
+ "fuse_layers": false,
99
+ "gradient_accumulation_steps": 2,
100
+ "has_rnn": true,
101
+ "hidden_size": 768,
102
+ "labels_encoder": null,
103
+ "labels_encoder_config": null,
104
+ "log_dir": "logs",
105
+ "lr_encoder": "1e-5",
106
+ "lr_others": "5e-5",
107
+ "max_len": 384,
108
+ "max_neg_type_ratio": 1,
109
+ "max_types": 25,
110
+ "max_width": 1,
111
+ "model_name": "microsoft/deberta-v3-large",
112
+ "model_type": "gliner",
113
+ "name": "large",
114
+ "num_post_fusion_layers": 1,
115
+ "num_steps": 60000,
116
+ "post_fusion_schema": "",
117
+ "prev_path": "none",
118
+ "random_drop": true,
119
+ "root_dir": "ablation_backbone",
120
+ "sep_token": "<<SEP>>",
121
+ "shuffle_types": true,
122
+ "size_sup": -1,
123
+ "span_mode": "marker",
124
+ "subtoken_pooling": "first",
125
+ "train_batch_size": 4,
126
+ "train_data": "NuMinds_custom_data_mix.json",
127
+ "transformers_version": "4.48.3",
128
+ "vocab_size": 128004,
129
+ "warmup_ratio": 0.1,
130
+ "words_splitter_type": "whitespace"
131
+ }
BSC-NLP4BIA/chagas-ner/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45c6fc56939820b78fda03d7dd5766055c6ca792e335743ef685564a446de6fb
3
+ size 1795640934
BSC-NLP4BIA/chagas-ner/special_tokens_map.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
+ "mask_token": "[MASK]",
6
+ "pad_token": "[PAD]",
7
+ "sep_token": "[SEP]",
8
+ "unk_token": {
9
+ "content": "[UNK]",
10
+ "lstrip": false,
11
+ "normalized": true,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ }
15
+ }
BSC-NLP4BIA/chagas-ner/spm.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c679fbf93643d19aab7ee10c0b99e460bdbc02fedf34b92b05af343b4af586fd
3
+ size 2464616
BSC-NLP4BIA/chagas-ner/tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6dba35d5dbe252aeafc5330bb2c1dbccd1632c829ea62a142b4082179db3fd94
3
+ size 8657331
BSC-NLP4BIA/chagas-ner/tokenizer_config.json ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[CLS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": true,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "128000": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "128001": {
44
+ "content": "[FLERT]",
45
+ "lstrip": false,
46
+ "normalized": true,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": false
50
+ },
51
+ "128002": {
52
+ "content": "<<ENT>>",
53
+ "lstrip": false,
54
+ "normalized": true,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": false
58
+ },
59
+ "128003": {
60
+ "content": "<<SEP>>",
61
+ "lstrip": false,
62
+ "normalized": true,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": false
66
+ }
67
+ },
68
+ "bos_token": "[CLS]",
69
+ "clean_up_tokenization_spaces": false,
70
+ "cls_token": "[CLS]",
71
+ "do_lower_case": false,
72
+ "eos_token": "[SEP]",
73
+ "extra_special_tokens": {},
74
+ "mask_token": "[MASK]",
75
+ "model_max_length": 1000000000000000019884624838656,
76
+ "pad_token": "[PAD]",
77
+ "sep_token": "[SEP]",
78
+ "sp_model_kwargs": {},
79
+ "split_by_punct": false,
80
+ "tokenizer_class": "DebertaV2Tokenizer",
81
+ "unk_token": "[UNK]",
82
+ "vocab_type": "spm"
83
+ }