narayan214 commited on
Commit
0480e3b
·
verified ·
1 Parent(s): 8ae890d

Upload fine-tuned PII NER model

Browse files
checkpoint-1053/config.json ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation": "gelu",
3
+ "architectures": [
4
+ "DistilBertForTokenClassification"
5
+ ],
6
+ "attention_dropout": 0.1,
7
+ "dim": 768,
8
+ "dropout": 0.1,
9
+ "dtype": "float32",
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "O",
13
+ "1": "B-AMOUNT",
14
+ "2": "B-COUNTRY",
15
+ "3": "B-CREDENTIALS",
16
+ "4": "B-DATE",
17
+ "5": "B-DATETIME",
18
+ "6": "B-EMAIL",
19
+ "7": "B-EXPIRYDATE",
20
+ "8": "B-FIRSTNAME",
21
+ "9": "B-IPADDRESS",
22
+ "10": "B-LASTNAME",
23
+ "11": "B-LOCATION",
24
+ "12": "B-MACADDRESS",
25
+ "13": "B-NATIONALITY",
26
+ "14": "B-NUMBER",
27
+ "15": "B-ORGANIZATION",
28
+ "16": "B-PERCENT",
29
+ "17": "B-PHONE",
30
+ "18": "B-TIME",
31
+ "19": "B-UID",
32
+ "20": "B-ZIPCODE",
33
+ "21": "I-AMOUNT",
34
+ "22": "I-COUNTRY",
35
+ "23": "I-CREDENTIALS",
36
+ "24": "I-DATE",
37
+ "25": "I-DATETIME",
38
+ "26": "I-EMAIL",
39
+ "27": "I-EXPIRYDATE",
40
+ "28": "I-FIRSTNAME",
41
+ "29": "I-IPADDRESS",
42
+ "30": "I-LASTNAME",
43
+ "31": "I-LOCATION",
44
+ "32": "I-MACADDRESS",
45
+ "33": "I-NATIONALITY",
46
+ "34": "I-NUMBER",
47
+ "35": "I-ORGANIZATION",
48
+ "36": "I-PERCENT",
49
+ "37": "I-PHONE",
50
+ "38": "I-TIME",
51
+ "39": "I-UID",
52
+ "40": "I-ZIPCODE"
53
+ },
54
+ "initializer_range": 0.02,
55
+ "label2id": {
56
+ "B-AMOUNT": 1,
57
+ "B-COUNTRY": 2,
58
+ "B-CREDENTIALS": 3,
59
+ "B-DATE": 4,
60
+ "B-DATETIME": 5,
61
+ "B-EMAIL": 6,
62
+ "B-EXPIRYDATE": 7,
63
+ "B-FIRSTNAME": 8,
64
+ "B-IPADDRESS": 9,
65
+ "B-LASTNAME": 10,
66
+ "B-LOCATION": 11,
67
+ "B-MACADDRESS": 12,
68
+ "B-NATIONALITY": 13,
69
+ "B-NUMBER": 14,
70
+ "B-ORGANIZATION": 15,
71
+ "B-PERCENT": 16,
72
+ "B-PHONE": 17,
73
+ "B-TIME": 18,
74
+ "B-UID": 19,
75
+ "B-ZIPCODE": 20,
76
+ "I-AMOUNT": 21,
77
+ "I-COUNTRY": 22,
78
+ "I-CREDENTIALS": 23,
79
+ "I-DATE": 24,
80
+ "I-DATETIME": 25,
81
+ "I-EMAIL": 26,
82
+ "I-EXPIRYDATE": 27,
83
+ "I-FIRSTNAME": 28,
84
+ "I-IPADDRESS": 29,
85
+ "I-LASTNAME": 30,
86
+ "I-LOCATION": 31,
87
+ "I-MACADDRESS": 32,
88
+ "I-NATIONALITY": 33,
89
+ "I-NUMBER": 34,
90
+ "I-ORGANIZATION": 35,
91
+ "I-PERCENT": 36,
92
+ "I-PHONE": 37,
93
+ "I-TIME": 38,
94
+ "I-UID": 39,
95
+ "I-ZIPCODE": 40,
96
+ "O": 0
97
+ },
98
+ "max_position_embeddings": 512,
99
+ "model_type": "distilbert",
100
+ "n_heads": 12,
101
+ "n_layers": 6,
102
+ "output_past": true,
103
+ "pad_token_id": 0,
104
+ "qa_dropout": 0.1,
105
+ "seq_classif_dropout": 0.2,
106
+ "sinusoidal_pos_embds": false,
107
+ "tie_weights_": true,
108
+ "transformers_version": "4.56.2",
109
+ "vocab_size": 28996
110
+ }
checkpoint-1053/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86332656eaca32aa36fd65b13c045170c787c54f2fbe4a38ff4e63e540136b60
3
+ size 260902108
checkpoint-1053/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a9f9cb82a30a2d3975277a5f17beeae875d6e7e12597e5ee98c0152da7cfd98
3
+ size 521866891
checkpoint-1053/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cbec5033c64a48f9a39d5f9d88a929ac670cc67d1a58b4c0c243a8e61614f770
3
+ size 14645
checkpoint-1053/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc90e2b1ac9c68ba58a7e5826a5d79a9af71e0667e1a52fc0f737deeaea0a5c5
3
+ size 1383
checkpoint-1053/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:099d50a3496c9efdbb2aaa403ccc3650cdf3bdbbf9abd678e060b3dcae93ef5c
3
+ size 1465
checkpoint-1053/special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
checkpoint-1053/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1053/tokenizer_config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": false,
47
+ "extra_special_tokens": {},
48
+ "mask_token": "[MASK]",
49
+ "max_length": 512,
50
+ "model_max_length": 512,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "stride": 0,
54
+ "strip_accents": null,
55
+ "tokenize_chinese_chars": true,
56
+ "tokenizer_class": "DistilBertTokenizer",
57
+ "truncation_side": "right",
58
+ "truncation_strategy": "longest_first",
59
+ "unk_token": "[UNK]"
60
+ }
checkpoint-1053/trainer_state.json ADDED
@@ -0,0 +1,226 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 1053,
3
+ "best_metric": 0.9034800222150778,
4
+ "best_model_checkpoint": "/content/drive/MyDrive/AI ML/PIIMsaker/model_training/model11102025/model/checkpoint-1053",
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 1053,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.14245014245014245,
14
+ "grad_norm": 4.053828716278076,
15
+ "learning_rate": 5.568181818181818e-06,
16
+ "loss": 2.8837,
17
+ "step": 50
18
+ },
19
+ {
20
+ "epoch": 0.2849002849002849,
21
+ "grad_norm": 2.1837286949157715,
22
+ "learning_rate": 1.125e-05,
23
+ "loss": 1.505,
24
+ "step": 100
25
+ },
26
+ {
27
+ "epoch": 0.42735042735042733,
28
+ "grad_norm": 2.4568073749542236,
29
+ "learning_rate": 1.6931818181818182e-05,
30
+ "loss": 0.6653,
31
+ "step": 150
32
+ },
33
+ {
34
+ "epoch": 0.5698005698005698,
35
+ "grad_norm": 1.7492278814315796,
36
+ "learning_rate": 1.9708676377454084e-05,
37
+ "loss": 0.3883,
38
+ "step": 200
39
+ },
40
+ {
41
+ "epoch": 0.7122507122507122,
42
+ "grad_norm": 2.4913177490234375,
43
+ "learning_rate": 1.9075364154528183e-05,
44
+ "loss": 0.235,
45
+ "step": 250
46
+ },
47
+ {
48
+ "epoch": 0.8547008547008547,
49
+ "grad_norm": 1.5097750425338745,
50
+ "learning_rate": 1.8442051931602282e-05,
51
+ "loss": 0.2108,
52
+ "step": 300
53
+ },
54
+ {
55
+ "epoch": 0.9971509971509972,
56
+ "grad_norm": 1.7614439725875854,
57
+ "learning_rate": 1.7808739708676378e-05,
58
+ "loss": 0.1995,
59
+ "step": 350
60
+ },
61
+ {
62
+ "epoch": 1.0,
63
+ "eval_accuracy": 0.0,
64
+ "eval_f1": 0.849518825626749,
65
+ "eval_loss": 0.15223319828510284,
66
+ "eval_precision": 0.8316818335794274,
67
+ "eval_recall": 0.8700500342964214,
68
+ "eval_runtime": 2.0848,
69
+ "eval_samples_per_second": 672.015,
70
+ "eval_steps_per_second": 21.105,
71
+ "step": 351
72
+ },
73
+ {
74
+ "epoch": 1.1396011396011396,
75
+ "grad_norm": 3.9141924381256104,
76
+ "learning_rate": 1.7175427485750477e-05,
77
+ "loss": 0.1588,
78
+ "step": 400
79
+ },
80
+ {
81
+ "epoch": 1.282051282051282,
82
+ "grad_norm": 0.9784330129623413,
83
+ "learning_rate": 1.6542115262824576e-05,
84
+ "loss": 0.1272,
85
+ "step": 450
86
+ },
87
+ {
88
+ "epoch": 1.4245014245014245,
89
+ "grad_norm": 2.1048169136047363,
90
+ "learning_rate": 1.590880303989867e-05,
91
+ "loss": 0.1345,
92
+ "step": 500
93
+ },
94
+ {
95
+ "epoch": 1.566951566951567,
96
+ "grad_norm": 2.9420387744903564,
97
+ "learning_rate": 1.527549081697277e-05,
98
+ "loss": 0.1323,
99
+ "step": 550
100
+ },
101
+ {
102
+ "epoch": 1.7094017094017095,
103
+ "grad_norm": 2.0811967849731445,
104
+ "learning_rate": 1.4642178594046866e-05,
105
+ "loss": 0.1458,
106
+ "step": 600
107
+ },
108
+ {
109
+ "epoch": 1.8518518518518519,
110
+ "grad_norm": 6.859551906585693,
111
+ "learning_rate": 1.4008866371120963e-05,
112
+ "loss": 0.1069,
113
+ "step": 650
114
+ },
115
+ {
116
+ "epoch": 1.9943019943019942,
117
+ "grad_norm": 1.1304482221603394,
118
+ "learning_rate": 1.337555414819506e-05,
119
+ "loss": 0.0973,
120
+ "step": 700
121
+ },
122
+ {
123
+ "epoch": 2.0,
124
+ "eval_accuracy": 0.0,
125
+ "eval_f1": 0.8943449857600699,
126
+ "eval_loss": 0.10938618332147598,
127
+ "eval_precision": 0.880242372075773,
128
+ "eval_recall": 0.9092690274163624,
129
+ "eval_runtime": 2.116,
130
+ "eval_samples_per_second": 662.11,
131
+ "eval_steps_per_second": 20.794,
132
+ "step": 702
133
+ },
134
+ {
135
+ "epoch": 2.1367521367521367,
136
+ "grad_norm": 1.131783366203308,
137
+ "learning_rate": 1.274224192526916e-05,
138
+ "loss": 0.0717,
139
+ "step": 750
140
+ },
141
+ {
142
+ "epoch": 2.2792022792022792,
143
+ "grad_norm": 2.2972805500030518,
144
+ "learning_rate": 1.2108929702343256e-05,
145
+ "loss": 0.0852,
146
+ "step": 800
147
+ },
148
+ {
149
+ "epoch": 2.421652421652422,
150
+ "grad_norm": 0.8555024862289429,
151
+ "learning_rate": 1.1475617479417354e-05,
152
+ "loss": 0.0742,
153
+ "step": 850
154
+ },
155
+ {
156
+ "epoch": 2.564102564102564,
157
+ "grad_norm": 1.5859020948410034,
158
+ "learning_rate": 1.0842305256491453e-05,
159
+ "loss": 0.0824,
160
+ "step": 900
161
+ },
162
+ {
163
+ "epoch": 2.7065527065527064,
164
+ "grad_norm": 0.9457488059997559,
165
+ "learning_rate": 1.0208993033565548e-05,
166
+ "loss": 0.0788,
167
+ "step": 950
168
+ },
169
+ {
170
+ "epoch": 2.849002849002849,
171
+ "grad_norm": 1.330013632774353,
172
+ "learning_rate": 9.575680810639646e-06,
173
+ "loss": 0.0751,
174
+ "step": 1000
175
+ },
176
+ {
177
+ "epoch": 2.9914529914529915,
178
+ "grad_norm": 0.8843093514442444,
179
+ "learning_rate": 8.942368587713743e-06,
180
+ "loss": 0.0992,
181
+ "step": 1050
182
+ },
183
+ {
184
+ "epoch": 3.0,
185
+ "eval_accuracy": 0.0,
186
+ "eval_f1": 0.9034800222150778,
187
+ "eval_loss": 0.09727968275547028,
188
+ "eval_precision": 0.8917970535296625,
189
+ "eval_recall": 0.9157629923674507,
190
+ "eval_runtime": 2.246,
191
+ "eval_samples_per_second": 623.788,
192
+ "eval_steps_per_second": 19.591,
193
+ "step": 1053
194
+ }
195
+ ],
196
+ "logging_steps": 50,
197
+ "max_steps": 1755,
198
+ "num_input_tokens_seen": 0,
199
+ "num_train_epochs": 5,
200
+ "save_steps": 500,
201
+ "stateful_callbacks": {
202
+ "EarlyStoppingCallback": {
203
+ "args": {
204
+ "early_stopping_patience": 3,
205
+ "early_stopping_threshold": 0.001
206
+ },
207
+ "attributes": {
208
+ "early_stopping_patience_counter": 0
209
+ }
210
+ },
211
+ "TrainerControl": {
212
+ "args": {
213
+ "should_epoch_stop": false,
214
+ "should_evaluate": false,
215
+ "should_log": false,
216
+ "should_save": true,
217
+ "should_training_stop": false
218
+ },
219
+ "attributes": {}
220
+ }
221
+ },
222
+ "total_flos": 549620270634240.0,
223
+ "train_batch_size": 16,
224
+ "trial_name": null,
225
+ "trial_params": null
226
+ }
checkpoint-1053/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dc0d1f41610192cd2e3fb0ddf624cd2ba54ef1a162452d0ad275908ce535518
3
+ size 5841
checkpoint-1053/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1404/config.json ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation": "gelu",
3
+ "architectures": [
4
+ "DistilBertForTokenClassification"
5
+ ],
6
+ "attention_dropout": 0.1,
7
+ "dim": 768,
8
+ "dropout": 0.1,
9
+ "dtype": "float32",
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "O",
13
+ "1": "B-AMOUNT",
14
+ "2": "B-COUNTRY",
15
+ "3": "B-CREDENTIALS",
16
+ "4": "B-DATE",
17
+ "5": "B-DATETIME",
18
+ "6": "B-EMAIL",
19
+ "7": "B-EXPIRYDATE",
20
+ "8": "B-FIRSTNAME",
21
+ "9": "B-IPADDRESS",
22
+ "10": "B-LASTNAME",
23
+ "11": "B-LOCATION",
24
+ "12": "B-MACADDRESS",
25
+ "13": "B-NATIONALITY",
26
+ "14": "B-NUMBER",
27
+ "15": "B-ORGANIZATION",
28
+ "16": "B-PERCENT",
29
+ "17": "B-PHONE",
30
+ "18": "B-TIME",
31
+ "19": "B-UID",
32
+ "20": "B-ZIPCODE",
33
+ "21": "I-AMOUNT",
34
+ "22": "I-COUNTRY",
35
+ "23": "I-CREDENTIALS",
36
+ "24": "I-DATE",
37
+ "25": "I-DATETIME",
38
+ "26": "I-EMAIL",
39
+ "27": "I-EXPIRYDATE",
40
+ "28": "I-FIRSTNAME",
41
+ "29": "I-IPADDRESS",
42
+ "30": "I-LASTNAME",
43
+ "31": "I-LOCATION",
44
+ "32": "I-MACADDRESS",
45
+ "33": "I-NATIONALITY",
46
+ "34": "I-NUMBER",
47
+ "35": "I-ORGANIZATION",
48
+ "36": "I-PERCENT",
49
+ "37": "I-PHONE",
50
+ "38": "I-TIME",
51
+ "39": "I-UID",
52
+ "40": "I-ZIPCODE"
53
+ },
54
+ "initializer_range": 0.02,
55
+ "label2id": {
56
+ "B-AMOUNT": 1,
57
+ "B-COUNTRY": 2,
58
+ "B-CREDENTIALS": 3,
59
+ "B-DATE": 4,
60
+ "B-DATETIME": 5,
61
+ "B-EMAIL": 6,
62
+ "B-EXPIRYDATE": 7,
63
+ "B-FIRSTNAME": 8,
64
+ "B-IPADDRESS": 9,
65
+ "B-LASTNAME": 10,
66
+ "B-LOCATION": 11,
67
+ "B-MACADDRESS": 12,
68
+ "B-NATIONALITY": 13,
69
+ "B-NUMBER": 14,
70
+ "B-ORGANIZATION": 15,
71
+ "B-PERCENT": 16,
72
+ "B-PHONE": 17,
73
+ "B-TIME": 18,
74
+ "B-UID": 19,
75
+ "B-ZIPCODE": 20,
76
+ "I-AMOUNT": 21,
77
+ "I-COUNTRY": 22,
78
+ "I-CREDENTIALS": 23,
79
+ "I-DATE": 24,
80
+ "I-DATETIME": 25,
81
+ "I-EMAIL": 26,
82
+ "I-EXPIRYDATE": 27,
83
+ "I-FIRSTNAME": 28,
84
+ "I-IPADDRESS": 29,
85
+ "I-LASTNAME": 30,
86
+ "I-LOCATION": 31,
87
+ "I-MACADDRESS": 32,
88
+ "I-NATIONALITY": 33,
89
+ "I-NUMBER": 34,
90
+ "I-ORGANIZATION": 35,
91
+ "I-PERCENT": 36,
92
+ "I-PHONE": 37,
93
+ "I-TIME": 38,
94
+ "I-UID": 39,
95
+ "I-ZIPCODE": 40,
96
+ "O": 0
97
+ },
98
+ "max_position_embeddings": 512,
99
+ "model_type": "distilbert",
100
+ "n_heads": 12,
101
+ "n_layers": 6,
102
+ "output_past": true,
103
+ "pad_token_id": 0,
104
+ "qa_dropout": 0.1,
105
+ "seq_classif_dropout": 0.2,
106
+ "sinusoidal_pos_embds": false,
107
+ "tie_weights_": true,
108
+ "transformers_version": "4.56.2",
109
+ "vocab_size": 28996
110
+ }
checkpoint-1404/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f3930070cf5001b24f4b3a754ddfb1981f4b0561fe7187dbe1adbd605766d69
3
+ size 260902108
checkpoint-1404/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78cb1f505c088b43a319b5b2095e3cbb5620f88cf26c0b4279f16d72adbdd39e
3
+ size 521866891
checkpoint-1404/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0560e9251864c5154d08bb46c59e977e904d8a522631d8d1592946ba2bfd1ef7
3
+ size 14645
checkpoint-1404/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3a782a4dbbfc98b2d4cc2a05c2aaf631ec6efdab2dc5da37bf69c5dc9d1ecc7
3
+ size 1383
checkpoint-1404/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4c59cf682ba754fdd25888a2cd59a753ba814ef164de84cbe57cd788df126f4
3
+ size 1465
checkpoint-1404/special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
checkpoint-1404/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1404/tokenizer_config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": false,
47
+ "extra_special_tokens": {},
48
+ "mask_token": "[MASK]",
49
+ "max_length": 512,
50
+ "model_max_length": 512,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "stride": 0,
54
+ "strip_accents": null,
55
+ "tokenize_chinese_chars": true,
56
+ "tokenizer_class": "DistilBertTokenizer",
57
+ "truncation_side": "right",
58
+ "truncation_strategy": "longest_first",
59
+ "unk_token": "[UNK]"
60
+ }
checkpoint-1404/trainer_state.json ADDED
@@ -0,0 +1,287 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 1404,
3
+ "best_metric": 0.9103603502831294,
4
+ "best_model_checkpoint": "/content/drive/MyDrive/AI ML/PIIMsaker/model_training/model11102025/model/checkpoint-1404",
5
+ "epoch": 4.0,
6
+ "eval_steps": 500,
7
+ "global_step": 1404,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.14245014245014245,
14
+ "grad_norm": 4.053828716278076,
15
+ "learning_rate": 5.568181818181818e-06,
16
+ "loss": 2.8837,
17
+ "step": 50
18
+ },
19
+ {
20
+ "epoch": 0.2849002849002849,
21
+ "grad_norm": 2.1837286949157715,
22
+ "learning_rate": 1.125e-05,
23
+ "loss": 1.505,
24
+ "step": 100
25
+ },
26
+ {
27
+ "epoch": 0.42735042735042733,
28
+ "grad_norm": 2.4568073749542236,
29
+ "learning_rate": 1.6931818181818182e-05,
30
+ "loss": 0.6653,
31
+ "step": 150
32
+ },
33
+ {
34
+ "epoch": 0.5698005698005698,
35
+ "grad_norm": 1.7492278814315796,
36
+ "learning_rate": 1.9708676377454084e-05,
37
+ "loss": 0.3883,
38
+ "step": 200
39
+ },
40
+ {
41
+ "epoch": 0.7122507122507122,
42
+ "grad_norm": 2.4913177490234375,
43
+ "learning_rate": 1.9075364154528183e-05,
44
+ "loss": 0.235,
45
+ "step": 250
46
+ },
47
+ {
48
+ "epoch": 0.8547008547008547,
49
+ "grad_norm": 1.5097750425338745,
50
+ "learning_rate": 1.8442051931602282e-05,
51
+ "loss": 0.2108,
52
+ "step": 300
53
+ },
54
+ {
55
+ "epoch": 0.9971509971509972,
56
+ "grad_norm": 1.7614439725875854,
57
+ "learning_rate": 1.7808739708676378e-05,
58
+ "loss": 0.1995,
59
+ "step": 350
60
+ },
61
+ {
62
+ "epoch": 1.0,
63
+ "eval_accuracy": 0.0,
64
+ "eval_f1": 0.849518825626749,
65
+ "eval_loss": 0.15223319828510284,
66
+ "eval_precision": 0.8316818335794274,
67
+ "eval_recall": 0.8700500342964214,
68
+ "eval_runtime": 2.0848,
69
+ "eval_samples_per_second": 672.015,
70
+ "eval_steps_per_second": 21.105,
71
+ "step": 351
72
+ },
73
+ {
74
+ "epoch": 1.1396011396011396,
75
+ "grad_norm": 3.9141924381256104,
76
+ "learning_rate": 1.7175427485750477e-05,
77
+ "loss": 0.1588,
78
+ "step": 400
79
+ },
80
+ {
81
+ "epoch": 1.282051282051282,
82
+ "grad_norm": 0.9784330129623413,
83
+ "learning_rate": 1.6542115262824576e-05,
84
+ "loss": 0.1272,
85
+ "step": 450
86
+ },
87
+ {
88
+ "epoch": 1.4245014245014245,
89
+ "grad_norm": 2.1048169136047363,
90
+ "learning_rate": 1.590880303989867e-05,
91
+ "loss": 0.1345,
92
+ "step": 500
93
+ },
94
+ {
95
+ "epoch": 1.566951566951567,
96
+ "grad_norm": 2.9420387744903564,
97
+ "learning_rate": 1.527549081697277e-05,
98
+ "loss": 0.1323,
99
+ "step": 550
100
+ },
101
+ {
102
+ "epoch": 1.7094017094017095,
103
+ "grad_norm": 2.0811967849731445,
104
+ "learning_rate": 1.4642178594046866e-05,
105
+ "loss": 0.1458,
106
+ "step": 600
107
+ },
108
+ {
109
+ "epoch": 1.8518518518518519,
110
+ "grad_norm": 6.859551906585693,
111
+ "learning_rate": 1.4008866371120963e-05,
112
+ "loss": 0.1069,
113
+ "step": 650
114
+ },
115
+ {
116
+ "epoch": 1.9943019943019942,
117
+ "grad_norm": 1.1304482221603394,
118
+ "learning_rate": 1.337555414819506e-05,
119
+ "loss": 0.0973,
120
+ "step": 700
121
+ },
122
+ {
123
+ "epoch": 2.0,
124
+ "eval_accuracy": 0.0,
125
+ "eval_f1": 0.8943449857600699,
126
+ "eval_loss": 0.10938618332147598,
127
+ "eval_precision": 0.880242372075773,
128
+ "eval_recall": 0.9092690274163624,
129
+ "eval_runtime": 2.116,
130
+ "eval_samples_per_second": 662.11,
131
+ "eval_steps_per_second": 20.794,
132
+ "step": 702
133
+ },
134
+ {
135
+ "epoch": 2.1367521367521367,
136
+ "grad_norm": 1.131783366203308,
137
+ "learning_rate": 1.274224192526916e-05,
138
+ "loss": 0.0717,
139
+ "step": 750
140
+ },
141
+ {
142
+ "epoch": 2.2792022792022792,
143
+ "grad_norm": 2.2972805500030518,
144
+ "learning_rate": 1.2108929702343256e-05,
145
+ "loss": 0.0852,
146
+ "step": 800
147
+ },
148
+ {
149
+ "epoch": 2.421652421652422,
150
+ "grad_norm": 0.8555024862289429,
151
+ "learning_rate": 1.1475617479417354e-05,
152
+ "loss": 0.0742,
153
+ "step": 850
154
+ },
155
+ {
156
+ "epoch": 2.564102564102564,
157
+ "grad_norm": 1.5859020948410034,
158
+ "learning_rate": 1.0842305256491453e-05,
159
+ "loss": 0.0824,
160
+ "step": 900
161
+ },
162
+ {
163
+ "epoch": 2.7065527065527064,
164
+ "grad_norm": 0.9457488059997559,
165
+ "learning_rate": 1.0208993033565548e-05,
166
+ "loss": 0.0788,
167
+ "step": 950
168
+ },
169
+ {
170
+ "epoch": 2.849002849002849,
171
+ "grad_norm": 1.330013632774353,
172
+ "learning_rate": 9.575680810639646e-06,
173
+ "loss": 0.0751,
174
+ "step": 1000
175
+ },
176
+ {
177
+ "epoch": 2.9914529914529915,
178
+ "grad_norm": 0.8843093514442444,
179
+ "learning_rate": 8.942368587713743e-06,
180
+ "loss": 0.0992,
181
+ "step": 1050
182
+ },
183
+ {
184
+ "epoch": 3.0,
185
+ "eval_accuracy": 0.0,
186
+ "eval_f1": 0.9034800222150778,
187
+ "eval_loss": 0.09727968275547028,
188
+ "eval_precision": 0.8917970535296625,
189
+ "eval_recall": 0.9157629923674507,
190
+ "eval_runtime": 2.246,
191
+ "eval_samples_per_second": 623.788,
192
+ "eval_steps_per_second": 19.591,
193
+ "step": 1053
194
+ },
195
+ {
196
+ "epoch": 3.133903133903134,
197
+ "grad_norm": 0.5358484983444214,
198
+ "learning_rate": 8.30905636478784e-06,
199
+ "loss": 0.0488,
200
+ "step": 1100
201
+ },
202
+ {
203
+ "epoch": 3.2763532763532766,
204
+ "grad_norm": 0.5537762641906738,
205
+ "learning_rate": 7.675744141861939e-06,
206
+ "loss": 0.0616,
207
+ "step": 1150
208
+ },
209
+ {
210
+ "epoch": 3.4188034188034186,
211
+ "grad_norm": 4.597471714019775,
212
+ "learning_rate": 7.0424319189360354e-06,
213
+ "loss": 0.0633,
214
+ "step": 1200
215
+ },
216
+ {
217
+ "epoch": 3.561253561253561,
218
+ "grad_norm": 1.4646154642105103,
219
+ "learning_rate": 6.4091196960101335e-06,
220
+ "loss": 0.0665,
221
+ "step": 1250
222
+ },
223
+ {
224
+ "epoch": 3.7037037037037037,
225
+ "grad_norm": 0.9090561270713806,
226
+ "learning_rate": 5.775807473084231e-06,
227
+ "loss": 0.0399,
228
+ "step": 1300
229
+ },
230
+ {
231
+ "epoch": 3.8461538461538463,
232
+ "grad_norm": 1.164624571800232,
233
+ "learning_rate": 5.142495250158328e-06,
234
+ "loss": 0.0377,
235
+ "step": 1350
236
+ },
237
+ {
238
+ "epoch": 3.9886039886039883,
239
+ "grad_norm": 12.009307861328125,
240
+ "learning_rate": 4.509183027232426e-06,
241
+ "loss": 0.0623,
242
+ "step": 1400
243
+ },
244
+ {
245
+ "epoch": 4.0,
246
+ "eval_accuracy": 0.0,
247
+ "eval_f1": 0.9103603502831294,
248
+ "eval_loss": 0.09384271502494812,
249
+ "eval_precision": 0.900490507971208,
250
+ "eval_recall": 0.9205799098457256,
251
+ "eval_runtime": 2.2095,
252
+ "eval_samples_per_second": 634.076,
253
+ "eval_steps_per_second": 19.914,
254
+ "step": 1404
255
+ }
256
+ ],
257
+ "logging_steps": 50,
258
+ "max_steps": 1755,
259
+ "num_input_tokens_seen": 0,
260
+ "num_train_epochs": 5,
261
+ "save_steps": 500,
262
+ "stateful_callbacks": {
263
+ "EarlyStoppingCallback": {
264
+ "args": {
265
+ "early_stopping_patience": 3,
266
+ "early_stopping_threshold": 0.001
267
+ },
268
+ "attributes": {
269
+ "early_stopping_patience_counter": 0
270
+ }
271
+ },
272
+ "TrainerControl": {
273
+ "args": {
274
+ "should_epoch_stop": false,
275
+ "should_evaluate": false,
276
+ "should_log": false,
277
+ "should_save": true,
278
+ "should_training_stop": false
279
+ },
280
+ "attributes": {}
281
+ }
282
+ },
283
+ "total_flos": 732827027512320.0,
284
+ "train_batch_size": 16,
285
+ "trial_name": null,
286
+ "trial_params": null
287
+ }
checkpoint-1404/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dc0d1f41610192cd2e3fb0ddf624cd2ba54ef1a162452d0ad275908ce535518
3
+ size 5841
checkpoint-1404/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1755/config.json ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation": "gelu",
3
+ "architectures": [
4
+ "DistilBertForTokenClassification"
5
+ ],
6
+ "attention_dropout": 0.1,
7
+ "dim": 768,
8
+ "dropout": 0.1,
9
+ "dtype": "float32",
10
+ "hidden_dim": 3072,
11
+ "id2label": {
12
+ "0": "O",
13
+ "1": "B-AMOUNT",
14
+ "2": "B-COUNTRY",
15
+ "3": "B-CREDENTIALS",
16
+ "4": "B-DATE",
17
+ "5": "B-DATETIME",
18
+ "6": "B-EMAIL",
19
+ "7": "B-EXPIRYDATE",
20
+ "8": "B-FIRSTNAME",
21
+ "9": "B-IPADDRESS",
22
+ "10": "B-LASTNAME",
23
+ "11": "B-LOCATION",
24
+ "12": "B-MACADDRESS",
25
+ "13": "B-NATIONALITY",
26
+ "14": "B-NUMBER",
27
+ "15": "B-ORGANIZATION",
28
+ "16": "B-PERCENT",
29
+ "17": "B-PHONE",
30
+ "18": "B-TIME",
31
+ "19": "B-UID",
32
+ "20": "B-ZIPCODE",
33
+ "21": "I-AMOUNT",
34
+ "22": "I-COUNTRY",
35
+ "23": "I-CREDENTIALS",
36
+ "24": "I-DATE",
37
+ "25": "I-DATETIME",
38
+ "26": "I-EMAIL",
39
+ "27": "I-EXPIRYDATE",
40
+ "28": "I-FIRSTNAME",
41
+ "29": "I-IPADDRESS",
42
+ "30": "I-LASTNAME",
43
+ "31": "I-LOCATION",
44
+ "32": "I-MACADDRESS",
45
+ "33": "I-NATIONALITY",
46
+ "34": "I-NUMBER",
47
+ "35": "I-ORGANIZATION",
48
+ "36": "I-PERCENT",
49
+ "37": "I-PHONE",
50
+ "38": "I-TIME",
51
+ "39": "I-UID",
52
+ "40": "I-ZIPCODE"
53
+ },
54
+ "initializer_range": 0.02,
55
+ "label2id": {
56
+ "B-AMOUNT": 1,
57
+ "B-COUNTRY": 2,
58
+ "B-CREDENTIALS": 3,
59
+ "B-DATE": 4,
60
+ "B-DATETIME": 5,
61
+ "B-EMAIL": 6,
62
+ "B-EXPIRYDATE": 7,
63
+ "B-FIRSTNAME": 8,
64
+ "B-IPADDRESS": 9,
65
+ "B-LASTNAME": 10,
66
+ "B-LOCATION": 11,
67
+ "B-MACADDRESS": 12,
68
+ "B-NATIONALITY": 13,
69
+ "B-NUMBER": 14,
70
+ "B-ORGANIZATION": 15,
71
+ "B-PERCENT": 16,
72
+ "B-PHONE": 17,
73
+ "B-TIME": 18,
74
+ "B-UID": 19,
75
+ "B-ZIPCODE": 20,
76
+ "I-AMOUNT": 21,
77
+ "I-COUNTRY": 22,
78
+ "I-CREDENTIALS": 23,
79
+ "I-DATE": 24,
80
+ "I-DATETIME": 25,
81
+ "I-EMAIL": 26,
82
+ "I-EXPIRYDATE": 27,
83
+ "I-FIRSTNAME": 28,
84
+ "I-IPADDRESS": 29,
85
+ "I-LASTNAME": 30,
86
+ "I-LOCATION": 31,
87
+ "I-MACADDRESS": 32,
88
+ "I-NATIONALITY": 33,
89
+ "I-NUMBER": 34,
90
+ "I-ORGANIZATION": 35,
91
+ "I-PERCENT": 36,
92
+ "I-PHONE": 37,
93
+ "I-TIME": 38,
94
+ "I-UID": 39,
95
+ "I-ZIPCODE": 40,
96
+ "O": 0
97
+ },
98
+ "max_position_embeddings": 512,
99
+ "model_type": "distilbert",
100
+ "n_heads": 12,
101
+ "n_layers": 6,
102
+ "output_past": true,
103
+ "pad_token_id": 0,
104
+ "qa_dropout": 0.1,
105
+ "seq_classif_dropout": 0.2,
106
+ "sinusoidal_pos_embds": false,
107
+ "tie_weights_": true,
108
+ "transformers_version": "4.56.2",
109
+ "vocab_size": 28996
110
+ }
checkpoint-1755/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:419fce98cde0a0c65acb33b98dd2e67820aec8342751d164cc0948369ef8dd59
3
+ size 260902108
checkpoint-1755/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d8419887f5c80bbfcedcb8f403f679ceb757f8da6bd6bec343de9a455b72c50
3
+ size 521866891
checkpoint-1755/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b62a08dbcdc9f7a18defa4779d3151c765fa336e6952e0c7f5dcc8bc9c10a06
3
+ size 14645
checkpoint-1755/scaler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc1ceae5ea465dc43ff24b6e9163cfed562a81984e23e3ce46d171bca92cf9a2
3
+ size 1383
checkpoint-1755/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bf29d5ef2a570d99a2f721fa7b09c44506a807f50b6cde70a5bdc2d2940068b
3
+ size 1465
checkpoint-1755/special_tokens_map.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
+ }
checkpoint-1755/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-1755/tokenizer_config.json ADDED
@@ -0,0 +1,60 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_lower_case": false,
47
+ "extra_special_tokens": {},
48
+ "mask_token": "[MASK]",
49
+ "max_length": 512,
50
+ "model_max_length": 512,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "stride": 0,
54
+ "strip_accents": null,
55
+ "tokenize_chinese_chars": true,
56
+ "tokenizer_class": "DistilBertTokenizer",
57
+ "truncation_side": "right",
58
+ "truncation_strategy": "longest_first",
59
+ "unk_token": "[UNK]"
60
+ }
checkpoint-1755/trainer_state.json ADDED
@@ -0,0 +1,348 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": 1755,
3
+ "best_metric": 0.9113534509294903,
4
+ "best_model_checkpoint": "/content/drive/MyDrive/AI ML/PIIMsaker/model_training/model11102025/model/checkpoint-1755",
5
+ "epoch": 5.0,
6
+ "eval_steps": 500,
7
+ "global_step": 1755,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.14245014245014245,
14
+ "grad_norm": 4.053828716278076,
15
+ "learning_rate": 5.568181818181818e-06,
16
+ "loss": 2.8837,
17
+ "step": 50
18
+ },
19
+ {
20
+ "epoch": 0.2849002849002849,
21
+ "grad_norm": 2.1837286949157715,
22
+ "learning_rate": 1.125e-05,
23
+ "loss": 1.505,
24
+ "step": 100
25
+ },
26
+ {
27
+ "epoch": 0.42735042735042733,
28
+ "grad_norm": 2.4568073749542236,
29
+ "learning_rate": 1.6931818181818182e-05,
30
+ "loss": 0.6653,
31
+ "step": 150
32
+ },
33
+ {
34
+ "epoch": 0.5698005698005698,
35
+ "grad_norm": 1.7492278814315796,
36
+ "learning_rate": 1.9708676377454084e-05,
37
+ "loss": 0.3883,
38
+ "step": 200
39
+ },
40
+ {
41
+ "epoch": 0.7122507122507122,
42
+ "grad_norm": 2.4913177490234375,
43
+ "learning_rate": 1.9075364154528183e-05,
44
+ "loss": 0.235,
45
+ "step": 250
46
+ },
47
+ {
48
+ "epoch": 0.8547008547008547,
49
+ "grad_norm": 1.5097750425338745,
50
+ "learning_rate": 1.8442051931602282e-05,
51
+ "loss": 0.2108,
52
+ "step": 300
53
+ },
54
+ {
55
+ "epoch": 0.9971509971509972,
56
+ "grad_norm": 1.7614439725875854,
57
+ "learning_rate": 1.7808739708676378e-05,
58
+ "loss": 0.1995,
59
+ "step": 350
60
+ },
61
+ {
62
+ "epoch": 1.0,
63
+ "eval_accuracy": 0.0,
64
+ "eval_f1": 0.849518825626749,
65
+ "eval_loss": 0.15223319828510284,
66
+ "eval_precision": 0.8316818335794274,
67
+ "eval_recall": 0.8700500342964214,
68
+ "eval_runtime": 2.0848,
69
+ "eval_samples_per_second": 672.015,
70
+ "eval_steps_per_second": 21.105,
71
+ "step": 351
72
+ },
73
+ {
74
+ "epoch": 1.1396011396011396,
75
+ "grad_norm": 3.9141924381256104,
76
+ "learning_rate": 1.7175427485750477e-05,
77
+ "loss": 0.1588,
78
+ "step": 400
79
+ },
80
+ {
81
+ "epoch": 1.282051282051282,
82
+ "grad_norm": 0.9784330129623413,
83
+ "learning_rate": 1.6542115262824576e-05,
84
+ "loss": 0.1272,
85
+ "step": 450
86
+ },
87
+ {
88
+ "epoch": 1.4245014245014245,
89
+ "grad_norm": 2.1048169136047363,
90
+ "learning_rate": 1.590880303989867e-05,
91
+ "loss": 0.1345,
92
+ "step": 500
93
+ },
94
+ {
95
+ "epoch": 1.566951566951567,
96
+ "grad_norm": 2.9420387744903564,
97
+ "learning_rate": 1.527549081697277e-05,
98
+ "loss": 0.1323,
99
+ "step": 550
100
+ },
101
+ {
102
+ "epoch": 1.7094017094017095,
103
+ "grad_norm": 2.0811967849731445,
104
+ "learning_rate": 1.4642178594046866e-05,
105
+ "loss": 0.1458,
106
+ "step": 600
107
+ },
108
+ {
109
+ "epoch": 1.8518518518518519,
110
+ "grad_norm": 6.859551906585693,
111
+ "learning_rate": 1.4008866371120963e-05,
112
+ "loss": 0.1069,
113
+ "step": 650
114
+ },
115
+ {
116
+ "epoch": 1.9943019943019942,
117
+ "grad_norm": 1.1304482221603394,
118
+ "learning_rate": 1.337555414819506e-05,
119
+ "loss": 0.0973,
120
+ "step": 700
121
+ },
122
+ {
123
+ "epoch": 2.0,
124
+ "eval_accuracy": 0.0,
125
+ "eval_f1": 0.8943449857600699,
126
+ "eval_loss": 0.10938618332147598,
127
+ "eval_precision": 0.880242372075773,
128
+ "eval_recall": 0.9092690274163624,
129
+ "eval_runtime": 2.116,
130
+ "eval_samples_per_second": 662.11,
131
+ "eval_steps_per_second": 20.794,
132
+ "step": 702
133
+ },
134
+ {
135
+ "epoch": 2.1367521367521367,
136
+ "grad_norm": 1.131783366203308,
137
+ "learning_rate": 1.274224192526916e-05,
138
+ "loss": 0.0717,
139
+ "step": 750
140
+ },
141
+ {
142
+ "epoch": 2.2792022792022792,
143
+ "grad_norm": 2.2972805500030518,
144
+ "learning_rate": 1.2108929702343256e-05,
145
+ "loss": 0.0852,
146
+ "step": 800
147
+ },
148
+ {
149
+ "epoch": 2.421652421652422,
150
+ "grad_norm": 0.8555024862289429,
151
+ "learning_rate": 1.1475617479417354e-05,
152
+ "loss": 0.0742,
153
+ "step": 850
154
+ },
155
+ {
156
+ "epoch": 2.564102564102564,
157
+ "grad_norm": 1.5859020948410034,
158
+ "learning_rate": 1.0842305256491453e-05,
159
+ "loss": 0.0824,
160
+ "step": 900
161
+ },
162
+ {
163
+ "epoch": 2.7065527065527064,
164
+ "grad_norm": 0.9457488059997559,
165
+ "learning_rate": 1.0208993033565548e-05,
166
+ "loss": 0.0788,
167
+ "step": 950
168
+ },
169
+ {
170
+ "epoch": 2.849002849002849,
171
+ "grad_norm": 1.330013632774353,
172
+ "learning_rate": 9.575680810639646e-06,
173
+ "loss": 0.0751,
174
+ "step": 1000
175
+ },
176
+ {
177
+ "epoch": 2.9914529914529915,
178
+ "grad_norm": 0.8843093514442444,
179
+ "learning_rate": 8.942368587713743e-06,
180
+ "loss": 0.0992,
181
+ "step": 1050
182
+ },
183
+ {
184
+ "epoch": 3.0,
185
+ "eval_accuracy": 0.0,
186
+ "eval_f1": 0.9034800222150778,
187
+ "eval_loss": 0.09727968275547028,
188
+ "eval_precision": 0.8917970535296625,
189
+ "eval_recall": 0.9157629923674507,
190
+ "eval_runtime": 2.246,
191
+ "eval_samples_per_second": 623.788,
192
+ "eval_steps_per_second": 19.591,
193
+ "step": 1053
194
+ },
195
+ {
196
+ "epoch": 3.133903133903134,
197
+ "grad_norm": 0.5358484983444214,
198
+ "learning_rate": 8.30905636478784e-06,
199
+ "loss": 0.0488,
200
+ "step": 1100
201
+ },
202
+ {
203
+ "epoch": 3.2763532763532766,
204
+ "grad_norm": 0.5537762641906738,
205
+ "learning_rate": 7.675744141861939e-06,
206
+ "loss": 0.0616,
207
+ "step": 1150
208
+ },
209
+ {
210
+ "epoch": 3.4188034188034186,
211
+ "grad_norm": 4.597471714019775,
212
+ "learning_rate": 7.0424319189360354e-06,
213
+ "loss": 0.0633,
214
+ "step": 1200
215
+ },
216
+ {
217
+ "epoch": 3.561253561253561,
218
+ "grad_norm": 1.4646154642105103,
219
+ "learning_rate": 6.4091196960101335e-06,
220
+ "loss": 0.0665,
221
+ "step": 1250
222
+ },
223
+ {
224
+ "epoch": 3.7037037037037037,
225
+ "grad_norm": 0.9090561270713806,
226
+ "learning_rate": 5.775807473084231e-06,
227
+ "loss": 0.0399,
228
+ "step": 1300
229
+ },
230
+ {
231
+ "epoch": 3.8461538461538463,
232
+ "grad_norm": 1.164624571800232,
233
+ "learning_rate": 5.142495250158328e-06,
234
+ "loss": 0.0377,
235
+ "step": 1350
236
+ },
237
+ {
238
+ "epoch": 3.9886039886039883,
239
+ "grad_norm": 12.009307861328125,
240
+ "learning_rate": 4.509183027232426e-06,
241
+ "loss": 0.0623,
242
+ "step": 1400
243
+ },
244
+ {
245
+ "epoch": 4.0,
246
+ "eval_accuracy": 0.0,
247
+ "eval_f1": 0.9103603502831294,
248
+ "eval_loss": 0.09384271502494812,
249
+ "eval_precision": 0.900490507971208,
250
+ "eval_recall": 0.9205799098457256,
251
+ "eval_runtime": 2.2095,
252
+ "eval_samples_per_second": 634.076,
253
+ "eval_steps_per_second": 19.914,
254
+ "step": 1404
255
+ },
256
+ {
257
+ "epoch": 4.131054131054131,
258
+ "grad_norm": 0.5174495577812195,
259
+ "learning_rate": 3.8758708043065234e-06,
260
+ "loss": 0.0408,
261
+ "step": 1450
262
+ },
263
+ {
264
+ "epoch": 4.273504273504273,
265
+ "grad_norm": 1.375473976135254,
266
+ "learning_rate": 3.242558581380621e-06,
267
+ "loss": 0.0418,
268
+ "step": 1500
269
+ },
270
+ {
271
+ "epoch": 4.415954415954416,
272
+ "grad_norm": 1.5857481956481934,
273
+ "learning_rate": 2.6092463584547184e-06,
274
+ "loss": 0.0464,
275
+ "step": 1550
276
+ },
277
+ {
278
+ "epoch": 4.5584045584045585,
279
+ "grad_norm": 0.6780904531478882,
280
+ "learning_rate": 1.975934135528816e-06,
281
+ "loss": 0.041,
282
+ "step": 1600
283
+ },
284
+ {
285
+ "epoch": 4.700854700854701,
286
+ "grad_norm": 2.867487907409668,
287
+ "learning_rate": 1.3426219126029133e-06,
288
+ "loss": 0.0292,
289
+ "step": 1650
290
+ },
291
+ {
292
+ "epoch": 4.843304843304844,
293
+ "grad_norm": 0.27440184354782104,
294
+ "learning_rate": 7.093096896770109e-07,
295
+ "loss": 0.0335,
296
+ "step": 1700
297
+ },
298
+ {
299
+ "epoch": 4.985754985754986,
300
+ "grad_norm": 0.47164642810821533,
301
+ "learning_rate": 7.59974667511083e-08,
302
+ "loss": 0.0367,
303
+ "step": 1750
304
+ },
305
+ {
306
+ "epoch": 5.0,
307
+ "eval_accuracy": 0.0,
308
+ "eval_f1": 0.9113534509294903,
309
+ "eval_loss": 0.08031181991100311,
310
+ "eval_precision": 0.90035757458181,
311
+ "eval_recall": 0.9227343452971066,
312
+ "eval_runtime": 2.7749,
313
+ "eval_samples_per_second": 504.892,
314
+ "eval_steps_per_second": 15.857,
315
+ "step": 1755
316
+ }
317
+ ],
318
+ "logging_steps": 50,
319
+ "max_steps": 1755,
320
+ "num_input_tokens_seen": 0,
321
+ "num_train_epochs": 5,
322
+ "save_steps": 500,
323
+ "stateful_callbacks": {
324
+ "EarlyStoppingCallback": {
325
+ "args": {
326
+ "early_stopping_patience": 3,
327
+ "early_stopping_threshold": 0.001
328
+ },
329
+ "attributes": {
330
+ "early_stopping_patience_counter": 1
331
+ }
332
+ },
333
+ "TrainerControl": {
334
+ "args": {
335
+ "should_epoch_stop": false,
336
+ "should_evaluate": false,
337
+ "should_log": false,
338
+ "should_save": true,
339
+ "should_training_stop": true
340
+ },
341
+ "attributes": {}
342
+ }
343
+ },
344
+ "total_flos": 916033784390400.0,
345
+ "train_batch_size": 16,
346
+ "trial_name": null,
347
+ "trial_params": null
348
+ }
checkpoint-1755/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dc0d1f41610192cd2e3fb0ddf624cd2ba54ef1a162452d0ad275908ce535518
3
+ size 5841
checkpoint-1755/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
config.json CHANGED
@@ -22,32 +22,34 @@
22
  "10": "B-LASTNAME",
23
  "11": "B-LOCATION",
24
  "12": "B-MACADDRESS",
25
- "13": "B-NUMBER",
26
- "14": "B-ORGANIZATION",
27
- "15": "B-PERCENT",
28
- "16": "B-PHONE",
29
- "17": "B-TIME",
30
- "18": "B-UID",
31
- "19": "B-ZIPCODE",
32
- "20": "I-AMOUNT",
33
- "21": "I-COUNTRY",
34
- "22": "I-CREDENTIALS",
35
- "23": "I-DATE",
36
- "24": "I-DATETIME",
37
- "25": "I-EMAIL",
38
- "26": "I-EXPIRYDATE",
39
- "27": "I-FIRSTNAME",
40
- "28": "I-IPADDRESS",
41
- "29": "I-LASTNAME",
42
- "30": "I-LOCATION",
43
- "31": "I-MACADDRESS",
44
- "32": "I-NUMBER",
45
- "33": "I-ORGANIZATION",
46
- "34": "I-PERCENT",
47
- "35": "I-PHONE",
48
- "36": "I-TIME",
49
- "37": "I-UID",
50
- "38": "I-ZIPCODE"
 
 
51
  },
52
  "initializer_range": 0.02,
53
  "label2id": {
@@ -63,32 +65,34 @@
63
  "B-LASTNAME": 10,
64
  "B-LOCATION": 11,
65
  "B-MACADDRESS": 12,
66
- "B-NUMBER": 13,
67
- "B-ORGANIZATION": 14,
68
- "B-PERCENT": 15,
69
- "B-PHONE": 16,
70
- "B-TIME": 17,
71
- "B-UID": 18,
72
- "B-ZIPCODE": 19,
73
- "I-AMOUNT": 20,
74
- "I-COUNTRY": 21,
75
- "I-CREDENTIALS": 22,
76
- "I-DATE": 23,
77
- "I-DATETIME": 24,
78
- "I-EMAIL": 25,
79
- "I-EXPIRYDATE": 26,
80
- "I-FIRSTNAME": 27,
81
- "I-IPADDRESS": 28,
82
- "I-LASTNAME": 29,
83
- "I-LOCATION": 30,
84
- "I-MACADDRESS": 31,
85
- "I-NUMBER": 32,
86
- "I-ORGANIZATION": 33,
87
- "I-PERCENT": 34,
88
- "I-PHONE": 35,
89
- "I-TIME": 36,
90
- "I-UID": 37,
91
- "I-ZIPCODE": 38,
 
 
92
  "O": 0
93
  },
94
  "max_position_embeddings": 512,
@@ -101,6 +105,6 @@
101
  "seq_classif_dropout": 0.2,
102
  "sinusoidal_pos_embds": false,
103
  "tie_weights_": true,
104
- "transformers_version": "4.56.1",
105
  "vocab_size": 28996
106
  }
 
22
  "10": "B-LASTNAME",
23
  "11": "B-LOCATION",
24
  "12": "B-MACADDRESS",
25
+ "13": "B-NATIONALITY",
26
+ "14": "B-NUMBER",
27
+ "15": "B-ORGANIZATION",
28
+ "16": "B-PERCENT",
29
+ "17": "B-PHONE",
30
+ "18": "B-TIME",
31
+ "19": "B-UID",
32
+ "20": "B-ZIPCODE",
33
+ "21": "I-AMOUNT",
34
+ "22": "I-COUNTRY",
35
+ "23": "I-CREDENTIALS",
36
+ "24": "I-DATE",
37
+ "25": "I-DATETIME",
38
+ "26": "I-EMAIL",
39
+ "27": "I-EXPIRYDATE",
40
+ "28": "I-FIRSTNAME",
41
+ "29": "I-IPADDRESS",
42
+ "30": "I-LASTNAME",
43
+ "31": "I-LOCATION",
44
+ "32": "I-MACADDRESS",
45
+ "33": "I-NATIONALITY",
46
+ "34": "I-NUMBER",
47
+ "35": "I-ORGANIZATION",
48
+ "36": "I-PERCENT",
49
+ "37": "I-PHONE",
50
+ "38": "I-TIME",
51
+ "39": "I-UID",
52
+ "40": "I-ZIPCODE"
53
  },
54
  "initializer_range": 0.02,
55
  "label2id": {
 
65
  "B-LASTNAME": 10,
66
  "B-LOCATION": 11,
67
  "B-MACADDRESS": 12,
68
+ "B-NATIONALITY": 13,
69
+ "B-NUMBER": 14,
70
+ "B-ORGANIZATION": 15,
71
+ "B-PERCENT": 16,
72
+ "B-PHONE": 17,
73
+ "B-TIME": 18,
74
+ "B-UID": 19,
75
+ "B-ZIPCODE": 20,
76
+ "I-AMOUNT": 21,
77
+ "I-COUNTRY": 22,
78
+ "I-CREDENTIALS": 23,
79
+ "I-DATE": 24,
80
+ "I-DATETIME": 25,
81
+ "I-EMAIL": 26,
82
+ "I-EXPIRYDATE": 27,
83
+ "I-FIRSTNAME": 28,
84
+ "I-IPADDRESS": 29,
85
+ "I-LASTNAME": 30,
86
+ "I-LOCATION": 31,
87
+ "I-MACADDRESS": 32,
88
+ "I-NATIONALITY": 33,
89
+ "I-NUMBER": 34,
90
+ "I-ORGANIZATION": 35,
91
+ "I-PERCENT": 36,
92
+ "I-PHONE": 37,
93
+ "I-TIME": 38,
94
+ "I-UID": 39,
95
+ "I-ZIPCODE": 40,
96
  "O": 0
97
  },
98
  "max_position_embeddings": 512,
 
105
  "seq_classif_dropout": 0.2,
106
  "sinusoidal_pos_embds": false,
107
  "tie_weights_": true,
108
+ "transformers_version": "4.56.2",
109
  "vocab_size": 28996
110
  }
label_mappings.json CHANGED
@@ -13,32 +13,34 @@
13
  "B-LASTNAME": 10,
14
  "B-LOCATION": 11,
15
  "B-MACADDRESS": 12,
16
- "B-NUMBER": 13,
17
- "B-ORGANIZATION": 14,
18
- "B-PERCENT": 15,
19
- "B-PHONE": 16,
20
- "B-TIME": 17,
21
- "B-UID": 18,
22
- "B-ZIPCODE": 19,
23
- "I-AMOUNT": 20,
24
- "I-COUNTRY": 21,
25
- "I-CREDENTIALS": 22,
26
- "I-DATE": 23,
27
- "I-DATETIME": 24,
28
- "I-EMAIL": 25,
29
- "I-EXPIRYDATE": 26,
30
- "I-FIRSTNAME": 27,
31
- "I-IPADDRESS": 28,
32
- "I-LASTNAME": 29,
33
- "I-LOCATION": 30,
34
- "I-MACADDRESS": 31,
35
- "I-NUMBER": 32,
36
- "I-ORGANIZATION": 33,
37
- "I-PERCENT": 34,
38
- "I-PHONE": 35,
39
- "I-TIME": 36,
40
- "I-UID": 37,
41
- "I-ZIPCODE": 38
 
 
42
  },
43
  "id2label": {
44
  "0": "O",
@@ -54,32 +56,34 @@
54
  "10": "B-LASTNAME",
55
  "11": "B-LOCATION",
56
  "12": "B-MACADDRESS",
57
- "13": "B-NUMBER",
58
- "14": "B-ORGANIZATION",
59
- "15": "B-PERCENT",
60
- "16": "B-PHONE",
61
- "17": "B-TIME",
62
- "18": "B-UID",
63
- "19": "B-ZIPCODE",
64
- "20": "I-AMOUNT",
65
- "21": "I-COUNTRY",
66
- "22": "I-CREDENTIALS",
67
- "23": "I-DATE",
68
- "24": "I-DATETIME",
69
- "25": "I-EMAIL",
70
- "26": "I-EXPIRYDATE",
71
- "27": "I-FIRSTNAME",
72
- "28": "I-IPADDRESS",
73
- "29": "I-LASTNAME",
74
- "30": "I-LOCATION",
75
- "31": "I-MACADDRESS",
76
- "32": "I-NUMBER",
77
- "33": "I-ORGANIZATION",
78
- "34": "I-PERCENT",
79
- "35": "I-PHONE",
80
- "36": "I-TIME",
81
- "37": "I-UID",
82
- "38": "I-ZIPCODE"
 
 
83
  },
84
- "num_labels": 39
85
  }
 
13
  "B-LASTNAME": 10,
14
  "B-LOCATION": 11,
15
  "B-MACADDRESS": 12,
16
+ "B-NATIONALITY": 13,
17
+ "B-NUMBER": 14,
18
+ "B-ORGANIZATION": 15,
19
+ "B-PERCENT": 16,
20
+ "B-PHONE": 17,
21
+ "B-TIME": 18,
22
+ "B-UID": 19,
23
+ "B-ZIPCODE": 20,
24
+ "I-AMOUNT": 21,
25
+ "I-COUNTRY": 22,
26
+ "I-CREDENTIALS": 23,
27
+ "I-DATE": 24,
28
+ "I-DATETIME": 25,
29
+ "I-EMAIL": 26,
30
+ "I-EXPIRYDATE": 27,
31
+ "I-FIRSTNAME": 28,
32
+ "I-IPADDRESS": 29,
33
+ "I-LASTNAME": 30,
34
+ "I-LOCATION": 31,
35
+ "I-MACADDRESS": 32,
36
+ "I-NATIONALITY": 33,
37
+ "I-NUMBER": 34,
38
+ "I-ORGANIZATION": 35,
39
+ "I-PERCENT": 36,
40
+ "I-PHONE": 37,
41
+ "I-TIME": 38,
42
+ "I-UID": 39,
43
+ "I-ZIPCODE": 40
44
  },
45
  "id2label": {
46
  "0": "O",
 
56
  "10": "B-LASTNAME",
57
  "11": "B-LOCATION",
58
  "12": "B-MACADDRESS",
59
+ "13": "B-NATIONALITY",
60
+ "14": "B-NUMBER",
61
+ "15": "B-ORGANIZATION",
62
+ "16": "B-PERCENT",
63
+ "17": "B-PHONE",
64
+ "18": "B-TIME",
65
+ "19": "B-UID",
66
+ "20": "B-ZIPCODE",
67
+ "21": "I-AMOUNT",
68
+ "22": "I-COUNTRY",
69
+ "23": "I-CREDENTIALS",
70
+ "24": "I-DATE",
71
+ "25": "I-DATETIME",
72
+ "26": "I-EMAIL",
73
+ "27": "I-EXPIRYDATE",
74
+ "28": "I-FIRSTNAME",
75
+ "29": "I-IPADDRESS",
76
+ "30": "I-LASTNAME",
77
+ "31": "I-LOCATION",
78
+ "32": "I-MACADDRESS",
79
+ "33": "I-NATIONALITY",
80
+ "34": "I-NUMBER",
81
+ "35": "I-ORGANIZATION",
82
+ "36": "I-PERCENT",
83
+ "37": "I-PHONE",
84
+ "38": "I-TIME",
85
+ "39": "I-UID",
86
+ "40": "I-ZIPCODE"
87
  },
88
+ "num_labels": 41
89
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2a326979a8b6b2139879e4ce8d53d1c6856227aaa183a0a11e03975158105141
3
- size 260895956
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:419fce98cde0a0c65acb33b98dd2e67820aec8342751d164cc0948369ef8dd59
3
+ size 260902108
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:607060dc1a385aeade053d5949a8e442300f240193b4ff8d5eb06938f5581f3d
3
  size 5841
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dc0d1f41610192cd2e3fb0ddf624cd2ba54ef1a162452d0ad275908ce535518
3
  size 5841