| _config_info: | |
| create_time: '2023-12-12 10:00:56' | |
| use_default_base_config: true | |
| config_filepath: | |
| - /data/yangl/LDNet-main/conf/Pretrain_ld.yaml | |
| task_type: SchemaGuidedInstructBertTask | |
| task_name: LDNet_Pretrain | |
| comment: ~~content as label, (start, end + 1) span | |
| output_dir: LDNet_outputs | |
| task_dir: LDNet_outputs/LDNet_Pretrain | |
| dump_cache_dir: LDNet_outputs/LDNet_Pretrain/cache | |
| regenerate_cache: false | |
| data_dir: resources/Mirror/v1.4_sampled_v3/merged/all_excluded | |
| train_filepath: resources/Mirror/v1.4_sampled_v3/merged/all_excluded/train.jsonl | |
| dev_filepath: resources/Mirror/v1.4_sampled_v3/merged/all_excluded/dev.jsonl | |
| test_filepath: resources/Mirror/v1.4_sampled_v3/merged/all_excluded/test.jsonl | |
| random_seed: 1227 | |
| num_epochs: 3 | |
| num_steps: -1 | |
| warmup_proportion: 0.1 | |
| epoch_patience: -1 | |
| step_patience: -1 | |
| batch_size: 64 | |
| learning_rate: 2.0e-05 | |
| max_grad_norm: 1.0 | |
| skip_train: false | |
| debug_mode: false | |
| grad_accum_steps: 1 | |
| resumed_training_path: null | |
| step_eval_interval: 5000 | |
| epoch_eval_interval: 1 | |
| eval_on_data: | |
| - train | |
| select_best_on_data: train | |
| select_best_by_key: loss | |
| best_metric_field: micro.f1 | |
| save_every_ckpt: true | |
| save_best_ckpt: true | |
| final_eval_on_test: false | |
| main_process_logging: true | |
| max_seq_len: 512 | |
| label_span: tag | |
| mode: span | |
| stream_mode: false | |
| bce: true | |
| kd: false | |
| kd_file: None | |
| fewshot: false | |
| zeroshot: false | |
| generate_logits: false | |
| plm_dir: microsoft/deberta-v3-large | |
| base_model_path: null | |
| train_batch_size: 4 | |
| eval_batch_size: 4 | |
| other_learning_rate: 0.0001 | |
| weight_decay: 0.1 | |
| dropout: 0.3 | |
| use_rope: true | |
| biaffine_size: 512 | |
| include_package: | |
| - src.task | |
| use_default_base_config: true | |
| lddrop: false | |
| droprate: 1.0 |