_config_info: create_time: '2023-12-12 10:00:56' use_default_base_config: true config_filepath: - /data/yangl/LDNet-main/conf/Pretrain_ld.yaml task_type: SchemaGuidedInstructBertTask task_name: LDNet_Pretrain comment: ~~content as label, (start, end + 1) span output_dir: LDNet_outputs task_dir: LDNet_outputs/LDNet_Pretrain dump_cache_dir: LDNet_outputs/LDNet_Pretrain/cache regenerate_cache: false data_dir: resources/Mirror/v1.4_sampled_v3/merged/all_excluded train_filepath: resources/Mirror/v1.4_sampled_v3/merged/all_excluded/train.jsonl dev_filepath: resources/Mirror/v1.4_sampled_v3/merged/all_excluded/dev.jsonl test_filepath: resources/Mirror/v1.4_sampled_v3/merged/all_excluded/test.jsonl random_seed: 1227 num_epochs: 3 num_steps: -1 warmup_proportion: 0.1 epoch_patience: -1 step_patience: -1 batch_size: 64 learning_rate: 2.0e-05 max_grad_norm: 1.0 skip_train: false debug_mode: false grad_accum_steps: 1 resumed_training_path: null step_eval_interval: 5000 epoch_eval_interval: 1 eval_on_data: - train select_best_on_data: train select_best_by_key: loss best_metric_field: micro.f1 save_every_ckpt: true save_best_ckpt: true final_eval_on_test: false main_process_logging: true max_seq_len: 512 label_span: tag mode: span stream_mode: false bce: true kd: false kd_file: None fewshot: false zeroshot: false generate_logits: false plm_dir: microsoft/deberta-v3-large base_model_path: null train_batch_size: 4 eval_batch_size: 4 other_learning_rate: 0.0001 weight_decay: 0.1 dropout: 0.3 use_rope: true biaffine_size: 512 include_package: - src.task use_default_base_config: true lddrop: false droprate: 1.0