Downton commited on
Commit
a179149
·
verified ·
1 Parent(s): 4bd0e85

Upload 2 files

Browse files
ckpt/SchemaGuidedInstructBertModel.best.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ac07d34278eb676ee45045200bc92a85037f2e61bc4b7582d5dd0efb2f196a2
3
+ size 5212359486
task_params.yaml ADDED
@@ -0,0 +1,64 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ _config_info:
2
+ create_time: '2023-12-12 10:00:56'
3
+ use_default_base_config: true
4
+ config_filepath:
5
+ - /data/yangl/LDNet-main/conf/Pretrain_ld.yaml
6
+ task_type: SchemaGuidedInstructBertTask
7
+ task_name: LDNet_Pretrain
8
+ comment: ~~content as label, (start, end + 1) span
9
+ output_dir: LDNet_outputs
10
+ task_dir: LDNet_outputs/LDNet_Pretrain
11
+ dump_cache_dir: LDNet_outputs/LDNet_Pretrain/cache
12
+ regenerate_cache: false
13
+ data_dir: resources/Mirror/v1.4_sampled_v3/merged/all_excluded
14
+ train_filepath: resources/Mirror/v1.4_sampled_v3/merged/all_excluded/train.jsonl
15
+ dev_filepath: resources/Mirror/v1.4_sampled_v3/merged/all_excluded/dev.jsonl
16
+ test_filepath: resources/Mirror/v1.4_sampled_v3/merged/all_excluded/test.jsonl
17
+ random_seed: 1227
18
+ num_epochs: 3
19
+ num_steps: -1
20
+ warmup_proportion: 0.1
21
+ epoch_patience: -1
22
+ step_patience: -1
23
+ batch_size: 64
24
+ learning_rate: 2.0e-05
25
+ max_grad_norm: 1.0
26
+ skip_train: false
27
+ debug_mode: false
28
+ grad_accum_steps: 1
29
+ resumed_training_path: null
30
+ step_eval_interval: 5000
31
+ epoch_eval_interval: 1
32
+ eval_on_data:
33
+ - train
34
+ select_best_on_data: train
35
+ select_best_by_key: loss
36
+ best_metric_field: micro.f1
37
+ save_every_ckpt: true
38
+ save_best_ckpt: true
39
+ final_eval_on_test: false
40
+ main_process_logging: true
41
+ max_seq_len: 512
42
+ label_span: tag
43
+ mode: span
44
+ stream_mode: false
45
+ bce: true
46
+ kd: false
47
+ kd_file: None
48
+ fewshot: false
49
+ zeroshot: false
50
+ generate_logits: false
51
+ plm_dir: microsoft/deberta-v3-large
52
+ base_model_path: null
53
+ train_batch_size: 4
54
+ eval_batch_size: 4
55
+ other_learning_rate: 0.0001
56
+ weight_decay: 0.1
57
+ dropout: 0.3
58
+ use_rope: true
59
+ biaffine_size: 512
60
+ include_package:
61
+ - src.task
62
+ use_default_base_config: true
63
+ lddrop: false
64
+ droprate: 1.0