litwell commited on
Commit
1c8db31
·
verified ·
1 Parent(s): 53e5620

Upload folder using huggingface_hub

Browse files
Files changed (33) hide show
  1. ww_training_128_4frame_v1.0_franka_4frame/checkpoints/trainstep_checkpoints/epoch=179-step=1800.ckpt +3 -0
  2. ww_training_128_4frame_v1.0_franka_4frame/configs/lightning.yaml +36 -0
  3. ww_training_128_4frame_v1.0_franka_4frame/configs/model.yaml +127 -0
  4. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_0:2025-05-04T16-59-18.txt +6 -0
  5. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_0:2025-05-04T17-00-59.txt +6 -0
  6. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_0:2025-05-04T17-03-42.txt +11 -0
  7. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_0:2025-05-04T17-07-39.txt +13 -0
  8. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_0:2025-05-04T17-10-25.txt +15 -0
  9. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_0:2025-05-04T17-12-30.txt +25 -0
  10. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_0:2025-05-04T17-17-13.txt +583 -0
  11. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_1:2025-05-04T16-59-18.txt +6 -0
  12. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_1:2025-05-04T17-00-59.txt +6 -0
  13. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_1:2025-05-04T17-03-42.txt +11 -0
  14. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_1:2025-05-04T17-07-39.txt +13 -0
  15. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_1:2025-05-04T17-10-25.txt +13 -0
  16. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_1:2025-05-04T17-12-30.txt +13 -0
  17. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_1:2025-05-04T17-17-13.txt +13 -0
  18. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_2:2025-05-04T16-59-18.txt +6 -0
  19. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_2:2025-05-04T17-00-59.txt +6 -0
  20. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_2:2025-05-04T17-03-42.txt +11 -0
  21. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_2:2025-05-04T17-07-39.txt +13 -0
  22. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_3:2025-05-04T16-59-18.txt +6 -0
  23. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_3:2025-05-04T17-00-59.txt +6 -0
  24. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_3:2025-05-04T17-03-42.txt +11 -0
  25. ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_3:2025-05-04T17-07-39.txt +13 -0
  26. ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_0/events.out.tfevents.1746349729.dgx-049.4189318.0 +3 -0
  27. ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_0/hparams.yaml +1 -0
  28. ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_1/events.out.tfevents.1746349895.dgx-049.43590.0 +3 -0
  29. ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_1/hparams.yaml +1 -0
  30. ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_2/events.out.tfevents.1746350023.dgx-049.51515.0 +3 -0
  31. ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_2/hparams.yaml +1 -0
  32. ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_3/events.out.tfevents.1746350306.dgx-049.141757.0 +3 -0
  33. ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_3/hparams.yaml +1 -0
ww_training_128_4frame_v1.0_franka_4frame/checkpoints/trainstep_checkpoints/epoch=179-step=1800.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f75b80baf1653611bff521b75889b36ad4d25a50ad81b2e30d9e5ec2874b45a
3
+ size 10436600940
ww_training_128_4frame_v1.0_franka_4frame/configs/lightning.yaml ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ lightning:
2
+ precision: 16
3
+ trainer:
4
+ benchmark: true
5
+ accumulate_grad_batches: 2
6
+ max_steps: 50000000000
7
+ log_every_n_steps: 50
8
+ val_check_interval: 0.5
9
+ gradient_clip_algorithm: norm
10
+ gradient_clip_val: 0.5
11
+ gpus: 2
12
+ num_nodes: 2
13
+ callbacks:
14
+ model_checkpoint:
15
+ target: pytorch_lightning.callbacks.ModelCheckpoint
16
+ params:
17
+ every_n_train_steps: 50
18
+ filename: '{epoch}-{step}'
19
+ save_weights_only: true
20
+ metrics_over_trainsteps_checkpoint:
21
+ target: pytorch_lightning.callbacks.ModelCheckpoint
22
+ params:
23
+ filename: '{epoch}-{step}'
24
+ save_weights_only: true
25
+ every_n_train_steps: 300
26
+ batch_logger:
27
+ target: callbacks.ImageLogger
28
+ params:
29
+ batch_frequency: 15
30
+ to_local: false
31
+ max_images: 8
32
+ log_images_kwargs:
33
+ ddim_steps: 50
34
+ unconditional_guidance_scale: 7.5
35
+ timestep_spacing: uniform_trailing
36
+ guidance_rescale: 0.7
ww_training_128_4frame_v1.0_franka_4frame/configs/model.yaml ADDED
@@ -0,0 +1,127 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ pretrained_checkpoint: checkpoints/Doubiiu/DynamiCrafter/model.ckpt
3
+ base_learning_rate: 1.0e-05
4
+ scale_lr: false
5
+ target: lvdm.models.ddpm3d.LatentVisualDiffusion
6
+ params:
7
+ rescale_betas_zero_snr: true
8
+ parameterization: v
9
+ linear_start: 0.00085
10
+ linear_end: 0.012
11
+ num_timesteps_cond: 1
12
+ log_every_t: 200
13
+ timesteps: 1000
14
+ first_stage_key: video
15
+ cond_stage_key: caption
16
+ cond_stage_trainable: false
17
+ image_proj_model_trainable: true
18
+ conditioning_key: hybrid
19
+ image_size:
20
+ - 16
21
+ - 16
22
+ channels: 4
23
+ scale_by_std: false
24
+ scale_factor: 0.18215
25
+ use_ema: false
26
+ uncond_prob: 0.05
27
+ uncond_type: empty_seq
28
+ rand_cond_frame: true
29
+ use_dynamic_rescale: true
30
+ base_scale: 0.7
31
+ fps_condition_type: fps
32
+ perframe_ae: true
33
+ unet_config:
34
+ target: lvdm.modules.networks.openaimodel3d.UNetModel
35
+ params:
36
+ in_channels: 8
37
+ out_channels: 4
38
+ model_channels: 320
39
+ attention_resolutions:
40
+ - 4
41
+ - 2
42
+ - 1
43
+ num_res_blocks: 2
44
+ channel_mult:
45
+ - 1
46
+ - 2
47
+ - 4
48
+ - 4
49
+ dropout: 0.1
50
+ num_head_channels: 64
51
+ transformer_depth: 1
52
+ context_dim: 1024
53
+ use_linear: true
54
+ use_checkpoint: true
55
+ temporal_conv: true
56
+ temporal_attention: true
57
+ temporal_selfatt_only: true
58
+ use_relative_position: false
59
+ use_causal_attention: false
60
+ temporal_length: 4
61
+ addition_attention: true
62
+ image_cross_attention: true
63
+ default_fs: 3
64
+ fs_condition: true
65
+ first_stage_config:
66
+ target: lvdm.models.autoencoder.AutoencoderKL
67
+ params:
68
+ embed_dim: 4
69
+ monitor: val/rec_loss
70
+ ddconfig:
71
+ double_z: true
72
+ z_channels: 4
73
+ resolution: 128
74
+ in_channels: 3
75
+ out_ch: 3
76
+ ch: 128
77
+ ch_mult:
78
+ - 1
79
+ - 2
80
+ - 4
81
+ - 4
82
+ num_res_blocks: 2
83
+ attn_resolutions: []
84
+ dropout: 0.0
85
+ lossconfig:
86
+ target: torch.nn.Identity
87
+ cond_stage_config:
88
+ target: lvdm.modules.encoders.condition.FrozenOpenCLIPEmbedder
89
+ params:
90
+ freeze: true
91
+ layer: penultimate
92
+ img_cond_stage_config:
93
+ target: lvdm.modules.encoders.condition.FrozenOpenCLIPImageEmbedderV2
94
+ params:
95
+ freeze: true
96
+ image_proj_stage_config:
97
+ target: lvdm.modules.encoders.resampler.Resampler
98
+ params:
99
+ dim: 1024
100
+ depth: 4
101
+ dim_head: 64
102
+ heads: 12
103
+ num_queries: 16
104
+ embedding_dim: 1280
105
+ output_dim: 1024
106
+ ff_mult: 4
107
+ video_length: 4
108
+ data:
109
+ target: utils_data.DataModuleFromConfig
110
+ params:
111
+ batch_size: 4
112
+ num_workers: 12
113
+ wrap: false
114
+ train:
115
+ target: lvdm.data.webvid.WebVid
116
+ params:
117
+ data_dir: /aifs4su/mmcode/worldm/open_x_embodiment/franka/4task_videos
118
+ meta_path: /aifs4su/mmcode/worldm/open_x_embodiment/franka/4task_franka.csv
119
+ video_length: 4
120
+ frame_stride: 4
121
+ load_raw_resolution: true
122
+ resolution:
123
+ - 128
124
+ - 128
125
+ spatial_transform: resize_center_crop
126
+ random_fs: false
127
+ --local-rank: 0
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_0:2025-05-04T16-59-18.txt ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ 2025-05-04 16:59:18,682-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 16:59:18,683-INFO: ***** Configing Model *****
3
+ 2025-05-04 16:59:20,029-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:00:07,548-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:00:28,963-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:00:28,964-INFO: ***** Configing Data *****
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_0:2025-05-04T17-00-59.txt ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ 2025-05-04 17:00:59,747-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:00:59,748-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:01:00,045-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:01:44,602-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:01:59,218-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:01:59,218-INFO: ***** Configing Data *****
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_0:2025-05-04T17-03-42.txt ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-05-04 17:03:42,340-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:03:42,341-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:03:42,712-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:04:27,116-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:04:41,966-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:04:41,966-INFO: ***** Configing Data *****
7
+ 2025-05-04 17:04:42,272-INFO: train, WebVid, 300
8
+ 2025-05-04 17:04:42,272-INFO: ***** Configing Trainer *****
9
+ 2025-05-04 17:04:42,273-INFO: Caution: Saving checkpoints every n train steps without deleting. This might require some free space.
10
+ 2025-05-04 17:04:42,340-INFO: ***** Running the Loop *****
11
+ 2025-05-04 17:04:42,340-INFO: <Training in DDPSharded Mode>
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_0:2025-05-04T17-07-39.txt ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-05-04 17:07:39,104-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:07:39,104-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:07:39,449-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:08:23,325-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:08:36,223-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:08:36,223-INFO: ***** Configing Data *****
7
+ 2025-05-04 17:08:36,608-INFO: train, WebVid, 300
8
+ 2025-05-04 17:08:36,608-INFO: ***** Configing Trainer *****
9
+ 2025-05-04 17:08:36,610-INFO: Caution: Saving checkpoints every n train steps without deleting. This might require some free space.
10
+ 2025-05-04 17:08:36,678-INFO: ***** Running the Loop *****
11
+ 2025-05-04 17:08:36,678-INFO: <Training in DDPSharded Mode>
12
+ 2025-05-04 17:08:45,749-INFO: @Training [1516] Full Paramters.
13
+ 2025-05-04 17:08:45,749-INFO: @Training [51] Paramters for Image_proj_model.
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_0:2025-05-04T17-10-25.txt ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-05-04 17:10:25,862-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:10:25,863-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:10:26,154-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:11:09,619-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:11:22,680-INFO: Running on 2=2x1 GPUs
6
+ 2025-05-04 17:11:22,681-INFO: ***** Configing Data *****
7
+ 2025-05-04 17:11:23,034-INFO: train, WebVid, 300
8
+ 2025-05-04 17:11:23,034-INFO: ***** Configing Trainer *****
9
+ 2025-05-04 17:11:23,036-INFO: Caution: Saving checkpoints every n train steps without deleting. This might require some free space.
10
+ 2025-05-04 17:11:23,111-INFO: ***** Running the Loop *****
11
+ 2025-05-04 17:11:23,111-INFO: <Training in DDPSharded Mode>
12
+ 2025-05-04 17:11:32,105-INFO: @Training [1516] Full Paramters.
13
+ 2025-05-04 17:11:32,105-INFO: @Training [51] Paramters for Image_proj_model.
14
+ 2025-05-04 17:12:05,683-INFO: Log [train] batch <ep0_idx14_rank0> to tensorboard ...
15
+ 2025-05-04 17:12:07,963-INFO: Finish!
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_0:2025-05-04T17-12-30.txt ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-05-04 17:12:31,042-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:12:31,042-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:12:31,374-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:13:17,362-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:13:32,576-INFO: Running on 2=2x1 GPUs
6
+ 2025-05-04 17:13:32,577-INFO: ***** Configing Data *****
7
+ 2025-05-04 17:13:32,983-INFO: train, WebVid, 300
8
+ 2025-05-04 17:13:32,983-INFO: ***** Configing Trainer *****
9
+ 2025-05-04 17:13:32,985-INFO: Caution: Saving checkpoints every n train steps without deleting. This might require some free space.
10
+ 2025-05-04 17:13:33,089-INFO: ***** Running the Loop *****
11
+ 2025-05-04 17:13:33,089-INFO: <Training in DDPSharded Mode>
12
+ 2025-05-04 17:13:39,091-INFO: @Training [1516] Full Paramters.
13
+ 2025-05-04 17:13:39,092-INFO: @Training [51] Paramters for Image_proj_model.
14
+ 2025-05-04 17:14:13,371-INFO: Log [train] batch <ep0_idx14_rank0> to tensorboard ...
15
+ 2025-05-04 17:14:15,193-INFO: Finish!
16
+ 2025-05-04 17:14:45,223-INFO: Log [train] batch <ep1_idx14_rank0> to tensorboard ...
17
+ 2025-05-04 17:14:45,473-INFO: Finish!
18
+ 2025-05-04 17:15:13,816-INFO: Log [train] batch <ep2_idx14_rank0> to tensorboard ...
19
+ 2025-05-04 17:15:14,005-INFO: Finish!
20
+ 2025-05-04 17:15:41,889-INFO: Log [train] batch <ep3_idx14_rank0> to tensorboard ...
21
+ 2025-05-04 17:15:42,108-INFO: Finish!
22
+ 2025-05-04 17:16:10,740-INFO: Log [train] batch <ep4_idx14_rank0> to tensorboard ...
23
+ 2025-05-04 17:16:10,956-INFO: Finish!
24
+ 2025-05-04 17:16:40,094-INFO: Log [train] batch <ep5_idx14_rank0> to tensorboard ...
25
+ 2025-05-04 17:16:40,363-INFO: Finish!
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_0:2025-05-04T17-17-13.txt ADDED
@@ -0,0 +1,583 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-05-04 17:17:13,844-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:17:13,844-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:17:14,271-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:17:57,623-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:18:13,647-INFO: Running on 2=2x1 GPUs
6
+ 2025-05-04 17:18:13,648-INFO: ***** Configing Data *****
7
+ 2025-05-04 17:18:14,022-INFO: train, WebVid, 300
8
+ 2025-05-04 17:18:14,022-INFO: ***** Configing Trainer *****
9
+ 2025-05-04 17:18:14,023-INFO: Caution: Saving checkpoints every n train steps without deleting. This might require some free space.
10
+ 2025-05-04 17:18:14,090-INFO: ***** Running the Loop *****
11
+ 2025-05-04 17:18:14,090-INFO: <Training in DDPSharded Mode>
12
+ 2025-05-04 17:18:22,554-INFO: @Training [1516] Full Paramters.
13
+ 2025-05-04 17:18:22,555-INFO: @Training [51] Paramters for Image_proj_model.
14
+ 2025-05-04 17:18:57,562-INFO: Log [train] batch <ep0_idx14_rank0> to tensorboard ...
15
+ 2025-05-04 17:18:58,573-INFO: Finish!
16
+ 2025-05-04 17:19:30,500-INFO: Log [train] batch <ep1_idx14_rank0> to tensorboard ...
17
+ 2025-05-04 17:19:30,771-INFO: Finish!
18
+ 2025-05-04 17:20:00,168-INFO: Log [train] batch <ep2_idx14_rank0> to tensorboard ...
19
+ 2025-05-04 17:20:00,360-INFO: Finish!
20
+ 2025-05-04 17:20:28,976-INFO: Log [train] batch <ep3_idx14_rank0> to tensorboard ...
21
+ 2025-05-04 17:20:29,178-INFO: Finish!
22
+ 2025-05-04 17:20:58,514-INFO: Log [train] batch <ep4_idx14_rank0> to tensorboard ...
23
+ 2025-05-04 17:20:58,738-INFO: Finish!
24
+ 2025-05-04 17:21:53,461-INFO: Log [train] batch <ep5_idx14_rank0> to tensorboard ...
25
+ 2025-05-04 17:21:53,684-INFO: Finish!
26
+ 2025-05-04 17:22:23,196-INFO: Log [train] batch <ep6_idx14_rank0> to tensorboard ...
27
+ 2025-05-04 17:22:23,426-INFO: Finish!
28
+ 2025-05-04 17:22:52,005-INFO: Log [train] batch <ep7_idx14_rank0> to tensorboard ...
29
+ 2025-05-04 17:22:52,235-INFO: Finish!
30
+ 2025-05-04 17:23:20,449-INFO: Log [train] batch <ep8_idx14_rank0> to tensorboard ...
31
+ 2025-05-04 17:23:20,659-INFO: Finish!
32
+ 2025-05-04 17:23:48,406-INFO: Log [train] batch <ep9_idx14_rank0> to tensorboard ...
33
+ 2025-05-04 17:23:48,628-INFO: Finish!
34
+ 2025-05-04 17:24:43,632-INFO: Log [train] batch <ep10_idx14_rank0> to tensorboard ...
35
+ 2025-05-04 17:24:43,849-INFO: Finish!
36
+ 2025-05-04 17:25:11,282-INFO: Log [train] batch <ep11_idx14_rank0> to tensorboard ...
37
+ 2025-05-04 17:25:11,538-INFO: Finish!
38
+ 2025-05-04 17:25:39,202-INFO: Log [train] batch <ep12_idx14_rank0> to tensorboard ...
39
+ 2025-05-04 17:25:39,428-INFO: Finish!
40
+ 2025-05-04 17:26:06,611-INFO: Log [train] batch <ep13_idx14_rank0> to tensorboard ...
41
+ 2025-05-04 17:26:06,806-INFO: Finish!
42
+ 2025-05-04 17:26:34,612-INFO: Log [train] batch <ep14_idx14_rank0> to tensorboard ...
43
+ 2025-05-04 17:26:34,836-INFO: Finish!
44
+ 2025-05-04 17:27:29,637-INFO: Log [train] batch <ep15_idx14_rank0> to tensorboard ...
45
+ 2025-05-04 17:27:29,855-INFO: Finish!
46
+ 2025-05-04 17:27:57,835-INFO: Log [train] batch <ep16_idx14_rank0> to tensorboard ...
47
+ 2025-05-04 17:27:58,086-INFO: Finish!
48
+ 2025-05-04 17:28:26,137-INFO: Log [train] batch <ep17_idx14_rank0> to tensorboard ...
49
+ 2025-05-04 17:28:26,379-INFO: Finish!
50
+ 2025-05-04 17:28:54,065-INFO: Log [train] batch <ep18_idx14_rank0> to tensorboard ...
51
+ 2025-05-04 17:28:54,256-INFO: Finish!
52
+ 2025-05-04 17:29:21,816-INFO: Log [train] batch <ep19_idx14_rank0> to tensorboard ...
53
+ 2025-05-04 17:29:22,039-INFO: Finish!
54
+ 2025-05-04 17:30:15,918-INFO: Log [train] batch <ep20_idx14_rank0> to tensorboard ...
55
+ 2025-05-04 17:30:16,121-INFO: Finish!
56
+ 2025-05-04 17:30:44,608-INFO: Log [train] batch <ep21_idx14_rank0> to tensorboard ...
57
+ 2025-05-04 17:30:44,808-INFO: Finish!
58
+ 2025-05-04 17:31:13,114-INFO: Log [train] batch <ep22_idx14_rank0> to tensorboard ...
59
+ 2025-05-04 17:31:13,340-INFO: Finish!
60
+ 2025-05-04 17:31:40,250-INFO: Log [train] batch <ep23_idx14_rank0> to tensorboard ...
61
+ 2025-05-04 17:31:40,464-INFO: Finish!
62
+ 2025-05-04 17:32:06,482-INFO: Log [train] batch <ep24_idx14_rank0> to tensorboard ...
63
+ 2025-05-04 17:32:06,685-INFO: Finish!
64
+ 2025-05-04 17:32:58,297-INFO: Log [train] batch <ep25_idx14_rank0> to tensorboard ...
65
+ 2025-05-04 17:32:58,493-INFO: Finish!
66
+ 2025-05-04 17:33:25,289-INFO: Log [train] batch <ep26_idx14_rank0> to tensorboard ...
67
+ 2025-05-04 17:33:25,479-INFO: Finish!
68
+ 2025-05-04 17:33:52,200-INFO: Log [train] batch <ep27_idx14_rank0> to tensorboard ...
69
+ 2025-05-04 17:33:52,411-INFO: Finish!
70
+ 2025-05-04 17:34:19,592-INFO: Log [train] batch <ep28_idx14_rank0> to tensorboard ...
71
+ 2025-05-04 17:34:19,827-INFO: Finish!
72
+ 2025-05-04 17:34:45,584-INFO: Log [train] batch <ep29_idx14_rank0> to tensorboard ...
73
+ 2025-05-04 17:34:45,763-INFO: Finish!
74
+ 2025-05-04 17:35:58,335-INFO: Log [train] batch <ep30_idx14_rank0> to tensorboard ...
75
+ 2025-05-04 17:35:58,531-INFO: Finish!
76
+ 2025-05-04 17:36:23,293-INFO: Log [train] batch <ep31_idx14_rank0> to tensorboard ...
77
+ 2025-05-04 17:36:23,448-INFO: Finish!
78
+ 2025-05-04 17:36:49,402-INFO: Log [train] batch <ep32_idx14_rank0> to tensorboard ...
79
+ 2025-05-04 17:36:49,602-INFO: Finish!
80
+ 2025-05-04 17:37:14,317-INFO: Log [train] batch <ep33_idx14_rank0> to tensorboard ...
81
+ 2025-05-04 17:37:14,489-INFO: Finish!
82
+ 2025-05-04 17:37:39,296-INFO: Log [train] batch <ep34_idx14_rank0> to tensorboard ...
83
+ 2025-05-04 17:37:39,442-INFO: Finish!
84
+ 2025-05-04 17:38:27,956-INFO: Log [train] batch <ep35_idx14_rank0> to tensorboard ...
85
+ 2025-05-04 17:38:28,178-INFO: Finish!
86
+ 2025-05-04 17:38:53,477-INFO: Log [train] batch <ep36_idx14_rank0> to tensorboard ...
87
+ 2025-05-04 17:38:53,656-INFO: Finish!
88
+ 2025-05-04 17:39:18,438-INFO: Log [train] batch <ep37_idx14_rank0> to tensorboard ...
89
+ 2025-05-04 17:39:18,646-INFO: Finish!
90
+ 2025-05-04 17:39:43,817-INFO: Log [train] batch <ep38_idx14_rank0> to tensorboard ...
91
+ 2025-05-04 17:39:44,027-INFO: Finish!
92
+ 2025-05-04 17:40:09,296-INFO: Log [train] batch <ep39_idx14_rank0> to tensorboard ...
93
+ 2025-05-04 17:40:09,487-INFO: Finish!
94
+ 2025-05-04 17:40:58,464-INFO: Log [train] batch <ep40_idx14_rank0> to tensorboard ...
95
+ 2025-05-04 17:40:58,662-INFO: Finish!
96
+ 2025-05-04 17:41:23,693-INFO: Log [train] batch <ep41_idx14_rank0> to tensorboard ...
97
+ 2025-05-04 17:41:23,893-INFO: Finish!
98
+ 2025-05-04 17:41:48,863-INFO: Log [train] batch <ep42_idx14_rank0> to tensorboard ...
99
+ 2025-05-04 17:41:49,059-INFO: Finish!
100
+ 2025-05-04 17:42:14,121-INFO: Log [train] batch <ep43_idx14_rank0> to tensorboard ...
101
+ 2025-05-04 17:42:14,332-INFO: Finish!
102
+ 2025-05-04 17:42:39,386-INFO: Log [train] batch <ep44_idx14_rank0> to tensorboard ...
103
+ 2025-05-04 17:42:39,603-INFO: Finish!
104
+ 2025-05-04 17:43:29,011-INFO: Log [train] batch <ep45_idx14_rank0> to tensorboard ...
105
+ 2025-05-04 17:43:29,220-INFO: Finish!
106
+ 2025-05-04 17:43:54,165-INFO: Log [train] batch <ep46_idx14_rank0> to tensorboard ...
107
+ 2025-05-04 17:43:54,374-INFO: Finish!
108
+ 2025-05-04 17:44:19,669-INFO: Log [train] batch <ep47_idx14_rank0> to tensorboard ...
109
+ 2025-05-04 17:44:19,870-INFO: Finish!
110
+ 2025-05-04 17:44:44,992-INFO: Log [train] batch <ep48_idx14_rank0> to tensorboard ...
111
+ 2025-05-04 17:44:45,196-INFO: Finish!
112
+ 2025-05-04 17:45:10,210-INFO: Log [train] batch <ep49_idx14_rank0> to tensorboard ...
113
+ 2025-05-04 17:45:10,431-INFO: Finish!
114
+ 2025-05-04 17:45:58,683-INFO: Log [train] batch <ep50_idx14_rank0> to tensorboard ...
115
+ 2025-05-04 17:45:58,888-INFO: Finish!
116
+ 2025-05-04 17:46:23,723-INFO: Log [train] batch <ep51_idx14_rank0> to tensorboard ...
117
+ 2025-05-04 17:46:23,921-INFO: Finish!
118
+ 2025-05-04 17:46:48,656-INFO: Log [train] batch <ep52_idx14_rank0> to tensorboard ...
119
+ 2025-05-04 17:46:48,850-INFO: Finish!
120
+ 2025-05-04 17:47:13,638-INFO: Log [train] batch <ep53_idx14_rank0> to tensorboard ...
121
+ 2025-05-04 17:47:13,841-INFO: Finish!
122
+ 2025-05-04 17:47:38,920-INFO: Log [train] batch <ep54_idx14_rank0> to tensorboard ...
123
+ 2025-05-04 17:47:39,128-INFO: Finish!
124
+ 2025-05-04 17:48:27,728-INFO: Log [train] batch <ep55_idx14_rank0> to tensorboard ...
125
+ 2025-05-04 17:48:27,921-INFO: Finish!
126
+ 2025-05-04 17:48:52,595-INFO: Log [train] batch <ep56_idx14_rank0> to tensorboard ...
127
+ 2025-05-04 17:48:52,736-INFO: Finish!
128
+ 2025-05-04 17:49:17,691-INFO: Log [train] batch <ep57_idx14_rank0> to tensorboard ...
129
+ 2025-05-04 17:49:17,890-INFO: Finish!
130
+ 2025-05-04 17:49:42,405-INFO: Log [train] batch <ep58_idx14_rank0> to tensorboard ...
131
+ 2025-05-04 17:49:42,580-INFO: Finish!
132
+ 2025-05-04 17:50:08,055-INFO: Log [train] batch <ep59_idx14_rank0> to tensorboard ...
133
+ 2025-05-04 17:50:08,253-INFO: Finish!
134
+ 2025-05-04 17:51:17,137-INFO: Log [train] batch <ep60_idx14_rank0> to tensorboard ...
135
+ 2025-05-04 17:51:17,316-INFO: Finish!
136
+ 2025-05-04 17:51:40,388-INFO: Log [train] batch <ep61_idx14_rank0> to tensorboard ...
137
+ 2025-05-04 17:51:40,587-INFO: Finish!
138
+ 2025-05-04 17:52:03,914-INFO: Log [train] batch <ep62_idx14_rank0> to tensorboard ...
139
+ 2025-05-04 17:52:04,092-INFO: Finish!
140
+ 2025-05-04 17:52:27,711-INFO: Log [train] batch <ep63_idx14_rank0> to tensorboard ...
141
+ 2025-05-04 17:52:27,870-INFO: Finish!
142
+ 2025-05-04 17:52:51,491-INFO: Log [train] batch <ep64_idx14_rank0> to tensorboard ...
143
+ 2025-05-04 17:52:51,692-INFO: Finish!
144
+ 2025-05-04 17:53:37,483-INFO: Log [train] batch <ep65_idx14_rank0> to tensorboard ...
145
+ 2025-05-04 17:53:37,631-INFO: Finish!
146
+ 2025-05-04 17:54:01,442-INFO: Log [train] batch <ep66_idx14_rank0> to tensorboard ...
147
+ 2025-05-04 17:54:01,619-INFO: Finish!
148
+ 2025-05-04 17:54:25,123-INFO: Log [train] batch <ep67_idx14_rank0> to tensorboard ...
149
+ 2025-05-04 17:54:25,317-INFO: Finish!
150
+ 2025-05-04 17:54:48,809-INFO: Log [train] batch <ep68_idx14_rank0> to tensorboard ...
151
+ 2025-05-04 17:54:49,054-INFO: Finish!
152
+ 2025-05-04 17:55:12,231-INFO: Log [train] batch <ep69_idx14_rank0> to tensorboard ...
153
+ 2025-05-04 17:55:12,416-INFO: Finish!
154
+ 2025-05-04 17:55:55,799-INFO: Log [train] batch <ep70_idx14_rank0> to tensorboard ...
155
+ 2025-05-04 17:55:55,978-INFO: Finish!
156
+ 2025-05-04 17:56:18,360-INFO: Log [train] batch <ep71_idx14_rank0> to tensorboard ...
157
+ 2025-05-04 17:56:18,519-INFO: Finish!
158
+ 2025-05-04 17:56:39,974-INFO: Log [train] batch <ep72_idx14_rank0> to tensorboard ...
159
+ 2025-05-04 17:56:40,131-INFO: Finish!
160
+ 2025-05-04 17:57:00,994-INFO: Log [train] batch <ep73_idx14_rank0> to tensorboard ...
161
+ 2025-05-04 17:57:01,174-INFO: Finish!
162
+ 2025-05-04 17:57:22,513-INFO: Log [train] batch <ep74_idx14_rank0> to tensorboard ...
163
+ 2025-05-04 17:57:22,645-INFO: Finish!
164
+ 2025-05-04 17:58:05,368-INFO: Log [train] batch <ep75_idx14_rank0> to tensorboard ...
165
+ 2025-05-04 17:58:05,559-INFO: Finish!
166
+ 2025-05-04 17:58:27,062-INFO: Log [train] batch <ep76_idx14_rank0> to tensorboard ...
167
+ 2025-05-04 17:58:27,234-INFO: Finish!
168
+ 2025-05-04 17:58:48,756-INFO: Log [train] batch <ep77_idx14_rank0> to tensorboard ...
169
+ 2025-05-04 17:58:48,901-INFO: Finish!
170
+ 2025-05-04 17:59:10,755-INFO: Log [train] batch <ep78_idx14_rank0> to tensorboard ...
171
+ 2025-05-04 17:59:10,883-INFO: Finish!
172
+ 2025-05-04 17:59:32,441-INFO: Log [train] batch <ep79_idx14_rank0> to tensorboard ...
173
+ 2025-05-04 17:59:32,590-INFO: Finish!
174
+ 2025-05-04 18:00:14,761-INFO: Log [train] batch <ep80_idx14_rank0> to tensorboard ...
175
+ 2025-05-04 18:00:14,904-INFO: Finish!
176
+ 2025-05-04 18:00:37,114-INFO: Log [train] batch <ep81_idx14_rank0> to tensorboard ...
177
+ 2025-05-04 18:00:37,280-INFO: Finish!
178
+ 2025-05-04 18:00:57,583-INFO: Log [train] batch <ep82_idx14_rank0> to tensorboard ...
179
+ 2025-05-04 18:00:57,733-INFO: Finish!
180
+ 2025-05-04 18:01:17,570-INFO: Log [train] batch <ep83_idx14_rank0> to tensorboard ...
181
+ 2025-05-04 18:01:17,707-INFO: Finish!
182
+ 2025-05-04 18:01:37,393-INFO: Log [train] batch <ep84_idx14_rank0> to tensorboard ...
183
+ 2025-05-04 18:01:37,538-INFO: Finish!
184
+ 2025-05-04 18:02:16,773-INFO: Log [train] batch <ep85_idx14_rank0> to tensorboard ...
185
+ 2025-05-04 18:02:16,919-INFO: Finish!
186
+ 2025-05-04 18:02:36,981-INFO: Log [train] batch <ep86_idx14_rank0> to tensorboard ...
187
+ 2025-05-04 18:02:37,111-INFO: Finish!
188
+ 2025-05-04 18:02:56,999-INFO: Log [train] batch <ep87_idx14_rank0> to tensorboard ...
189
+ 2025-05-04 18:02:57,142-INFO: Finish!
190
+ 2025-05-04 18:03:17,221-INFO: Log [train] batch <ep88_idx14_rank0> to tensorboard ...
191
+ 2025-05-04 18:03:17,364-INFO: Finish!
192
+ 2025-05-04 18:03:37,275-INFO: Log [train] batch <ep89_idx14_rank0> to tensorboard ...
193
+ 2025-05-04 18:03:37,423-INFO: Finish!
194
+ 2025-05-04 18:04:35,954-INFO: Log [train] batch <ep90_idx14_rank0> to tensorboard ...
195
+ 2025-05-04 18:04:36,080-INFO: Finish!
196
+ 2025-05-04 18:04:56,644-INFO: Log [train] batch <ep91_idx14_rank0> to tensorboard ...
197
+ 2025-05-04 18:04:56,777-INFO: Finish!
198
+ 2025-05-04 18:05:16,439-INFO: Log [train] batch <ep92_idx14_rank0> to tensorboard ...
199
+ 2025-05-04 18:05:16,569-INFO: Finish!
200
+ 2025-05-04 18:05:36,164-INFO: Log [train] batch <ep93_idx14_rank0> to tensorboard ...
201
+ 2025-05-04 18:05:36,327-INFO: Finish!
202
+ 2025-05-04 18:05:56,223-INFO: Log [train] batch <ep94_idx14_rank0> to tensorboard ...
203
+ 2025-05-04 18:05:56,356-INFO: Finish!
204
+ 2025-05-04 18:06:35,044-INFO: Log [train] batch <ep95_idx14_rank0> to tensorboard ...
205
+ 2025-05-04 18:06:35,202-INFO: Finish!
206
+ 2025-05-04 18:06:55,374-INFO: Log [train] batch <ep96_idx14_rank0> to tensorboard ...
207
+ 2025-05-04 18:06:55,557-INFO: Finish!
208
+ 2025-05-04 18:07:15,298-INFO: Log [train] batch <ep97_idx14_rank0> to tensorboard ...
209
+ 2025-05-04 18:07:15,445-INFO: Finish!
210
+ 2025-05-04 18:07:35,433-INFO: Log [train] batch <ep98_idx14_rank0> to tensorboard ...
211
+ 2025-05-04 18:07:35,578-INFO: Finish!
212
+ 2025-05-04 18:07:55,318-INFO: Log [train] batch <ep99_idx14_rank0> to tensorboard ...
213
+ 2025-05-04 18:07:55,448-INFO: Finish!
214
+ 2025-05-04 18:08:34,994-INFO: Log [train] batch <ep100_idx14_rank0> to tensorboard ...
215
+ 2025-05-04 18:08:35,126-INFO: Finish!
216
+ 2025-05-04 18:08:54,896-INFO: Log [train] batch <ep101_idx14_rank0> to tensorboard ...
217
+ 2025-05-04 18:08:55,036-INFO: Finish!
218
+ 2025-05-04 18:09:14,550-INFO: Log [train] batch <ep102_idx14_rank0> to tensorboard ...
219
+ 2025-05-04 18:09:14,689-INFO: Finish!
220
+ 2025-05-04 18:09:34,856-INFO: Log [train] batch <ep103_idx14_rank0> to tensorboard ...
221
+ 2025-05-04 18:09:35,007-INFO: Finish!
222
+ 2025-05-04 18:09:53,846-INFO: Log [train] batch <ep104_idx14_rank0> to tensorboard ...
223
+ 2025-05-04 18:09:53,995-INFO: Finish!
224
+ 2025-05-04 18:10:40,434-INFO: Log [train] batch <ep105_idx14_rank0> to tensorboard ...
225
+ 2025-05-04 18:10:40,567-INFO: Finish!
226
+ 2025-05-04 18:11:09,947-INFO: Log [train] batch <ep106_idx14_rank0> to tensorboard ...
227
+ 2025-05-04 18:11:10,254-INFO: Finish!
228
+ 2025-05-04 18:11:38,955-INFO: Log [train] batch <ep107_idx14_rank0> to tensorboard ...
229
+ 2025-05-04 18:11:39,142-INFO: Finish!
230
+ 2025-05-04 18:12:08,332-INFO: Log [train] batch <ep108_idx14_rank0> to tensorboard ...
231
+ 2025-05-04 18:12:08,557-INFO: Finish!
232
+ 2025-05-04 18:12:37,597-INFO: Log [train] batch <ep109_idx14_rank0> to tensorboard ...
233
+ 2025-05-04 18:12:37,802-INFO: Finish!
234
+ 2025-05-04 18:13:33,740-INFO: Log [train] batch <ep110_idx14_rank0> to tensorboard ...
235
+ 2025-05-04 18:13:33,958-INFO: Finish!
236
+ 2025-05-04 18:14:02,815-INFO: Log [train] batch <ep111_idx14_rank0> to tensorboard ...
237
+ 2025-05-04 18:14:03,051-INFO: Finish!
238
+ 2025-05-04 18:14:30,614-INFO: Log [train] batch <ep112_idx14_rank0> to tensorboard ...
239
+ 2025-05-04 18:14:30,804-INFO: Finish!
240
+ 2025-05-04 18:14:59,990-INFO: Log [train] batch <ep113_idx14_rank0> to tensorboard ...
241
+ 2025-05-04 18:15:00,187-INFO: Finish!
242
+ 2025-05-04 18:15:28,177-INFO: Log [train] batch <ep114_idx14_rank0> to tensorboard ...
243
+ 2025-05-04 18:15:28,383-INFO: Finish!
244
+ 2025-05-04 18:16:24,559-INFO: Log [train] batch <ep115_idx14_rank0> to tensorboard ...
245
+ 2025-05-04 18:16:24,774-INFO: Finish!
246
+ 2025-05-04 18:16:53,026-INFO: Log [train] batch <ep116_idx14_rank0> to tensorboard ...
247
+ 2025-05-04 18:16:53,254-INFO: Finish!
248
+ 2025-05-04 18:17:22,089-INFO: Log [train] batch <ep117_idx14_rank0> to tensorboard ...
249
+ 2025-05-04 18:17:22,281-INFO: Finish!
250
+ 2025-05-04 18:17:51,150-INFO: Log [train] batch <ep118_idx14_rank0> to tensorboard ...
251
+ 2025-05-04 18:17:51,362-INFO: Finish!
252
+ 2025-05-04 18:18:20,173-INFO: Log [train] batch <ep119_idx14_rank0> to tensorboard ...
253
+ 2025-05-04 18:18:20,358-INFO: Finish!
254
+ 2025-05-04 18:19:45,900-INFO: Log [train] batch <ep120_idx14_rank0> to tensorboard ...
255
+ 2025-05-04 18:19:46,106-INFO: Finish!
256
+ 2025-05-04 18:20:14,980-INFO: Log [train] batch <ep121_idx14_rank0> to tensorboard ...
257
+ 2025-05-04 18:20:15,180-INFO: Finish!
258
+ 2025-05-04 18:20:43,246-INFO: Log [train] batch <ep122_idx14_rank0> to tensorboard ...
259
+ 2025-05-04 18:20:43,448-INFO: Finish!
260
+ 2025-05-04 18:21:12,726-INFO: Log [train] batch <ep123_idx14_rank0> to tensorboard ...
261
+ 2025-05-04 18:21:12,937-INFO: Finish!
262
+ 2025-05-04 18:21:41,950-INFO: Log [train] batch <ep124_idx14_rank0> to tensorboard ...
263
+ 2025-05-04 18:21:42,168-INFO: Finish!
264
+ 2025-05-04 18:22:39,962-INFO: Log [train] batch <ep125_idx14_rank0> to tensorboard ...
265
+ 2025-05-04 18:22:40,163-INFO: Finish!
266
+ 2025-05-04 18:23:08,663-INFO: Log [train] batch <ep126_idx14_rank0> to tensorboard ...
267
+ 2025-05-04 18:23:08,900-INFO: Finish!
268
+ 2025-05-04 18:23:38,072-INFO: Log [train] batch <ep127_idx14_rank0> to tensorboard ...
269
+ 2025-05-04 18:23:38,297-INFO: Finish!
270
+ 2025-05-04 18:24:07,830-INFO: Log [train] batch <ep128_idx14_rank0> to tensorboard ...
271
+ 2025-05-04 18:24:08,048-INFO: Finish!
272
+ 2025-05-04 18:24:37,396-INFO: Log [train] batch <ep129_idx14_rank0> to tensorboard ...
273
+ 2025-05-04 18:24:37,653-INFO: Finish!
274
+ 2025-05-04 18:25:34,619-INFO: Log [train] batch <ep130_idx14_rank0> to tensorboard ...
275
+ 2025-05-04 18:25:34,790-INFO: Finish!
276
+ 2025-05-04 18:26:03,983-INFO: Log [train] batch <ep131_idx14_rank0> to tensorboard ...
277
+ 2025-05-04 18:26:04,176-INFO: Finish!
278
+ 2025-05-04 18:26:32,496-INFO: Log [train] batch <ep132_idx14_rank0> to tensorboard ...
279
+ 2025-05-04 18:26:32,692-INFO: Finish!
280
+ 2025-05-04 18:27:01,848-INFO: Log [train] batch <ep133_idx14_rank0> to tensorboard ...
281
+ 2025-05-04 18:27:02,058-INFO: Finish!
282
+ 2025-05-04 18:27:30,564-INFO: Log [train] batch <ep134_idx14_rank0> to tensorboard ...
283
+ 2025-05-04 18:27:30,775-INFO: Finish!
284
+ 2025-05-04 18:28:27,735-INFO: Log [train] batch <ep135_idx14_rank0> to tensorboard ...
285
+ 2025-05-04 18:28:27,934-INFO: Finish!
286
+ 2025-05-04 18:28:55,953-INFO: Log [train] batch <ep136_idx14_rank0> to tensorboard ...
287
+ 2025-05-04 18:28:56,131-INFO: Finish!
288
+ 2025-05-04 18:29:24,425-INFO: Log [train] batch <ep137_idx14_rank0> to tensorboard ...
289
+ 2025-05-04 18:29:24,625-INFO: Finish!
290
+ 2025-05-04 18:29:53,926-INFO: Log [train] batch <ep138_idx14_rank0> to tensorboard ...
291
+ 2025-05-04 18:29:54,151-INFO: Finish!
292
+ 2025-05-04 18:30:22,425-INFO: Log [train] batch <ep139_idx14_rank0> to tensorboard ...
293
+ 2025-05-04 18:30:22,636-INFO: Finish!
294
+ 2025-05-04 18:31:19,827-INFO: Log [train] batch <ep140_idx14_rank0> to tensorboard ...
295
+ 2025-05-04 18:31:20,047-INFO: Finish!
296
+ 2025-05-04 18:31:48,394-INFO: Log [train] batch <ep141_idx14_rank0> to tensorboard ...
297
+ 2025-05-04 18:31:48,583-INFO: Finish!
298
+ 2025-05-04 18:32:17,548-INFO: Log [train] batch <ep142_idx14_rank0> to tensorboard ...
299
+ 2025-05-04 18:32:17,746-INFO: Finish!
300
+ 2025-05-04 18:32:46,681-INFO: Log [train] batch <ep143_idx14_rank0> to tensorboard ...
301
+ 2025-05-04 18:32:46,945-INFO: Finish!
302
+ 2025-05-04 18:33:15,957-INFO: Log [train] batch <ep144_idx14_rank0> to tensorboard ...
303
+ 2025-05-04 18:33:16,165-INFO: Finish!
304
+ 2025-05-04 18:34:13,916-INFO: Log [train] batch <ep145_idx14_rank0> to tensorboard ...
305
+ 2025-05-04 18:34:14,117-INFO: Finish!
306
+ 2025-05-04 18:34:43,118-INFO: Log [train] batch <ep146_idx14_rank0> to tensorboard ...
307
+ 2025-05-04 18:34:43,309-INFO: Finish!
308
+ 2025-05-04 18:35:11,886-INFO: Log [train] batch <ep147_idx14_rank0> to tensorboard ...
309
+ 2025-05-04 18:35:12,176-INFO: Finish!
310
+ 2025-05-04 18:35:40,859-INFO: Log [train] batch <ep148_idx14_rank0> to tensorboard ...
311
+ 2025-05-04 18:35:41,070-INFO: Finish!
312
+ 2025-05-04 18:36:10,104-INFO: Log [train] batch <ep149_idx14_rank0> to tensorboard ...
313
+ 2025-05-04 18:36:10,284-INFO: Finish!
314
+ 2025-05-04 18:37:37,976-INFO: Log [train] batch <ep150_idx14_rank0> to tensorboard ...
315
+ 2025-05-04 18:37:38,186-INFO: Finish!
316
+ 2025-05-04 18:38:07,264-INFO: Log [train] batch <ep151_idx14_rank0> to tensorboard ...
317
+ 2025-05-04 18:38:07,438-INFO: Finish!
318
+ 2025-05-04 18:38:36,838-INFO: Log [train] batch <ep152_idx14_rank0> to tensorboard ...
319
+ 2025-05-04 18:38:37,055-INFO: Finish!
320
+ 2025-05-04 18:39:05,824-INFO: Log [train] batch <ep153_idx14_rank0> to tensorboard ...
321
+ 2025-05-04 18:39:06,034-INFO: Finish!
322
+ 2025-05-04 18:39:34,350-INFO: Log [train] batch <ep154_idx14_rank0> to tensorboard ...
323
+ 2025-05-04 18:39:34,576-INFO: Finish!
324
+ 2025-05-04 18:40:29,428-INFO: Log [train] batch <ep155_idx14_rank0> to tensorboard ...
325
+ 2025-05-04 18:40:29,623-INFO: Finish!
326
+ 2025-05-04 18:40:58,475-INFO: Log [train] batch <ep156_idx14_rank0> to tensorboard ...
327
+ 2025-05-04 18:40:58,628-INFO: Finish!
328
+ 2025-05-04 18:41:27,075-INFO: Log [train] batch <ep157_idx14_rank0> to tensorboard ...
329
+ 2025-05-04 18:41:27,264-INFO: Finish!
330
+ 2025-05-04 18:41:55,668-INFO: Log [train] batch <ep158_idx14_rank0> to tensorboard ...
331
+ 2025-05-04 18:41:55,860-INFO: Finish!
332
+ 2025-05-04 18:42:24,449-INFO: Log [train] batch <ep159_idx14_rank0> to tensorboard ...
333
+ 2025-05-04 18:42:24,626-INFO: Finish!
334
+ 2025-05-04 18:43:20,719-INFO: Log [train] batch <ep160_idx14_rank0> to tensorboard ...
335
+ 2025-05-04 18:43:20,904-INFO: Finish!
336
+ 2025-05-04 18:43:49,971-INFO: Log [train] batch <ep161_idx14_rank0> to tensorboard ...
337
+ 2025-05-04 18:43:50,188-INFO: Finish!
338
+ 2025-05-04 18:44:19,122-INFO: Log [train] batch <ep162_idx14_rank0> to tensorboard ...
339
+ 2025-05-04 18:44:19,327-INFO: Finish!
340
+ 2025-05-04 18:44:48,015-INFO: Log [train] batch <ep163_idx14_rank0> to tensorboard ...
341
+ 2025-05-04 18:44:48,210-INFO: Finish!
342
+ 2025-05-04 18:45:17,127-INFO: Log [train] batch <ep164_idx14_rank0> to tensorboard ...
343
+ 2025-05-04 18:45:17,325-INFO: Finish!
344
+ 2025-05-04 18:46:13,698-INFO: Log [train] batch <ep165_idx14_rank0> to tensorboard ...
345
+ 2025-05-04 18:46:13,927-INFO: Finish!
346
+ 2025-05-04 18:46:42,259-INFO: Log [train] batch <ep166_idx14_rank0> to tensorboard ...
347
+ 2025-05-04 18:46:42,473-INFO: Finish!
348
+ 2025-05-04 18:47:11,576-INFO: Log [train] batch <ep167_idx14_rank0> to tensorboard ...
349
+ 2025-05-04 18:47:11,768-INFO: Finish!
350
+ 2025-05-04 18:47:41,134-INFO: Log [train] batch <ep168_idx14_rank0> to tensorboard ...
351
+ 2025-05-04 18:47:41,374-INFO: Finish!
352
+ 2025-05-04 18:48:09,737-INFO: Log [train] batch <ep169_idx14_rank0> to tensorboard ...
353
+ 2025-05-04 18:48:09,927-INFO: Finish!
354
+ 2025-05-04 18:49:06,074-INFO: Log [train] batch <ep170_idx14_rank0> to tensorboard ...
355
+ 2025-05-04 18:49:06,284-INFO: Finish!
356
+ 2025-05-04 18:49:34,999-INFO: Log [train] batch <ep171_idx14_rank0> to tensorboard ...
357
+ 2025-05-04 18:49:35,237-INFO: Finish!
358
+ 2025-05-04 18:50:03,863-INFO: Log [train] batch <ep172_idx14_rank0> to tensorboard ...
359
+ 2025-05-04 18:50:04,079-INFO: Finish!
360
+ 2025-05-04 18:50:32,651-INFO: Log [train] batch <ep173_idx14_rank0> to tensorboard ...
361
+ 2025-05-04 18:50:32,897-INFO: Finish!
362
+ 2025-05-04 18:51:01,363-INFO: Log [train] batch <ep174_idx14_rank0> to tensorboard ...
363
+ 2025-05-04 18:51:01,573-INFO: Finish!
364
+ 2025-05-04 18:51:57,107-INFO: Log [train] batch <ep175_idx14_rank0> to tensorboard ...
365
+ 2025-05-04 18:51:57,320-INFO: Finish!
366
+ 2025-05-04 18:52:27,230-INFO: Log [train] batch <ep176_idx14_rank0> to tensorboard ...
367
+ 2025-05-04 18:52:27,442-INFO: Finish!
368
+ 2025-05-04 18:52:55,908-INFO: Log [train] batch <ep177_idx14_rank0> to tensorboard ...
369
+ 2025-05-04 18:52:56,099-INFO: Finish!
370
+ 2025-05-04 18:53:24,700-INFO: Log [train] batch <ep178_idx14_rank0> to tensorboard ...
371
+ 2025-05-04 18:53:24,887-INFO: Finish!
372
+ 2025-05-04 18:53:53,534-INFO: Log [train] batch <ep179_idx14_rank0> to tensorboard ...
373
+ 2025-05-04 18:53:53,741-INFO: Finish!
374
+ 2025-05-04 18:55:15,802-INFO: Log [train] batch <ep180_idx14_rank0> to tensorboard ...
375
+ 2025-05-04 18:55:16,023-INFO: Finish!
376
+ 2025-05-04 18:55:44,512-INFO: Log [train] batch <ep181_idx14_rank0> to tensorboard ...
377
+ 2025-05-04 18:55:44,732-INFO: Finish!
378
+ 2025-05-04 18:56:13,348-INFO: Log [train] batch <ep182_idx14_rank0> to tensorboard ...
379
+ 2025-05-04 18:56:13,562-INFO: Finish!
380
+ 2025-05-04 18:56:41,627-INFO: Log [train] batch <ep183_idx14_rank0> to tensorboard ...
381
+ 2025-05-04 18:56:41,864-INFO: Finish!
382
+ 2025-05-04 18:57:10,317-INFO: Log [train] batch <ep184_idx14_rank0> to tensorboard ...
383
+ 2025-05-04 18:57:10,521-INFO: Finish!
384
+ 2025-05-04 18:58:06,780-INFO: Log [train] batch <ep185_idx14_rank0> to tensorboard ...
385
+ 2025-05-04 18:58:07,029-INFO: Finish!
386
+ 2025-05-04 18:58:35,414-INFO: Log [train] batch <ep186_idx14_rank0> to tensorboard ...
387
+ 2025-05-04 18:58:35,611-INFO: Finish!
388
+ 2025-05-04 18:59:04,723-INFO: Log [train] batch <ep187_idx14_rank0> to tensorboard ...
389
+ 2025-05-04 18:59:04,994-INFO: Finish!
390
+ 2025-05-04 18:59:33,378-INFO: Log [train] batch <ep188_idx14_rank0> to tensorboard ...
391
+ 2025-05-04 18:59:33,611-INFO: Finish!
392
+ 2025-05-04 19:00:02,612-INFO: Log [train] batch <ep189_idx14_rank0> to tensorboard ...
393
+ 2025-05-04 19:00:02,829-INFO: Finish!
394
+ 2025-05-04 19:00:59,209-INFO: Log [train] batch <ep190_idx14_rank0> to tensorboard ...
395
+ 2025-05-04 19:00:59,460-INFO: Finish!
396
+ 2025-05-04 19:01:27,989-INFO: Log [train] batch <ep191_idx14_rank0> to tensorboard ...
397
+ 2025-05-04 19:01:28,178-INFO: Finish!
398
+ 2025-05-04 19:01:56,941-INFO: Log [train] batch <ep192_idx14_rank0> to tensorboard ...
399
+ 2025-05-04 19:01:57,209-INFO: Finish!
400
+ 2025-05-04 19:02:27,256-INFO: Log [train] batch <ep193_idx14_rank0> to tensorboard ...
401
+ 2025-05-04 19:02:27,470-INFO: Finish!
402
+ 2025-05-04 19:02:56,313-INFO: Log [train] batch <ep194_idx14_rank0> to tensorboard ...
403
+ 2025-05-04 19:02:56,587-INFO: Finish!
404
+ 2025-05-04 19:03:52,671-INFO: Log [train] batch <ep195_idx14_rank0> to tensorboard ...
405
+ 2025-05-04 19:03:52,894-INFO: Finish!
406
+ 2025-05-04 19:04:21,012-INFO: Log [train] batch <ep196_idx14_rank0> to tensorboard ...
407
+ 2025-05-04 19:04:21,221-INFO: Finish!
408
+ 2025-05-04 19:04:50,561-INFO: Log [train] batch <ep197_idx14_rank0> to tensorboard ...
409
+ 2025-05-04 19:04:50,771-INFO: Finish!
410
+ 2025-05-04 19:05:20,738-INFO: Log [train] batch <ep198_idx14_rank0> to tensorboard ...
411
+ 2025-05-04 19:05:20,947-INFO: Finish!
412
+ 2025-05-04 19:05:51,283-INFO: Log [train] batch <ep199_idx14_rank0> to tensorboard ...
413
+ 2025-05-04 19:05:51,504-INFO: Finish!
414
+ 2025-05-04 19:06:47,933-INFO: Log [train] batch <ep200_idx14_rank0> to tensorboard ...
415
+ 2025-05-04 19:06:48,112-INFO: Finish!
416
+ 2025-05-04 19:07:17,728-INFO: Log [train] batch <ep201_idx14_rank0> to tensorboard ...
417
+ 2025-05-04 19:07:17,939-INFO: Finish!
418
+ 2025-05-04 19:07:47,987-INFO: Log [train] batch <ep202_idx14_rank0> to tensorboard ...
419
+ 2025-05-04 19:07:48,217-INFO: Finish!
420
+ 2025-05-04 19:08:17,560-INFO: Log [train] batch <ep203_idx14_rank0> to tensorboard ...
421
+ 2025-05-04 19:08:17,769-INFO: Finish!
422
+ 2025-05-04 19:08:47,656-INFO: Log [train] batch <ep204_idx14_rank0> to tensorboard ...
423
+ 2025-05-04 19:08:47,905-INFO: Finish!
424
+ 2025-05-04 19:09:46,139-INFO: Log [train] batch <ep205_idx14_rank0> to tensorboard ...
425
+ 2025-05-04 19:09:46,367-INFO: Finish!
426
+ 2025-05-04 19:10:16,868-INFO: Log [train] batch <ep206_idx14_rank0> to tensorboard ...
427
+ 2025-05-04 19:10:17,086-INFO: Finish!
428
+ 2025-05-04 19:10:47,133-INFO: Log [train] batch <ep207_idx14_rank0> to tensorboard ...
429
+ 2025-05-04 19:10:47,336-INFO: Finish!
430
+ 2025-05-04 19:11:16,789-INFO: Log [train] batch <ep208_idx14_rank0> to tensorboard ...
431
+ 2025-05-04 19:11:16,997-INFO: Finish!
432
+ 2025-05-04 19:11:47,245-INFO: Log [train] batch <ep209_idx14_rank0> to tensorboard ...
433
+ 2025-05-04 19:11:47,465-INFO: Finish!
434
+ 2025-05-04 19:13:09,651-INFO: Log [train] batch <ep210_idx14_rank0> to tensorboard ...
435
+ 2025-05-04 19:13:09,838-INFO: Finish!
436
+ 2025-05-04 19:13:38,806-INFO: Log [train] batch <ep211_idx14_rank0> to tensorboard ...
437
+ 2025-05-04 19:13:39,020-INFO: Finish!
438
+ 2025-05-04 19:14:07,262-INFO: Log [train] batch <ep212_idx14_rank0> to tensorboard ...
439
+ 2025-05-04 19:14:07,458-INFO: Finish!
440
+ 2025-05-04 19:14:34,644-INFO: Log [train] batch <ep213_idx14_rank0> to tensorboard ...
441
+ 2025-05-04 19:14:34,857-INFO: Finish!
442
+ 2025-05-04 19:15:02,109-INFO: Log [train] batch <ep214_idx14_rank0> to tensorboard ...
443
+ 2025-05-04 19:15:02,302-INFO: Finish!
444
+ 2025-05-04 19:15:55,189-INFO: Log [train] batch <ep215_idx14_rank0> to tensorboard ...
445
+ 2025-05-04 19:15:55,367-INFO: Finish!
446
+ 2025-05-04 19:16:23,047-INFO: Log [train] batch <ep216_idx14_rank0> to tensorboard ...
447
+ 2025-05-04 19:16:23,270-INFO: Finish!
448
+ 2025-05-04 19:16:50,630-INFO: Log [train] batch <ep217_idx14_rank0> to tensorboard ...
449
+ 2025-05-04 19:16:50,843-INFO: Finish!
450
+ 2025-05-04 19:17:18,244-INFO: Log [train] batch <ep218_idx14_rank0> to tensorboard ...
451
+ 2025-05-04 19:17:18,442-INFO: Finish!
452
+ 2025-05-04 19:17:45,930-INFO: Log [train] batch <ep219_idx14_rank0> to tensorboard ...
453
+ 2025-05-04 19:17:46,152-INFO: Finish!
454
+ 2025-05-04 19:18:39,484-INFO: Log [train] batch <ep220_idx14_rank0> to tensorboard ...
455
+ 2025-05-04 19:18:39,671-INFO: Finish!
456
+ 2025-05-04 19:19:06,855-INFO: Log [train] batch <ep221_idx14_rank0> to tensorboard ...
457
+ 2025-05-04 19:19:07,053-INFO: Finish!
458
+ 2025-05-04 19:19:33,798-INFO: Log [train] batch <ep222_idx14_rank0> to tensorboard ...
459
+ 2025-05-04 19:19:33,991-INFO: Finish!
460
+ 2025-05-04 19:19:59,370-INFO: Log [train] batch <ep223_idx14_rank0> to tensorboard ...
461
+ 2025-05-04 19:19:59,565-INFO: Finish!
462
+ 2025-05-04 19:20:25,825-INFO: Log [train] batch <ep224_idx14_rank0> to tensorboard ...
463
+ 2025-05-04 19:20:26,022-INFO: Finish!
464
+ 2025-05-04 19:21:16,287-INFO: Log [train] batch <ep225_idx14_rank0> to tensorboard ...
465
+ 2025-05-04 19:21:16,487-INFO: Finish!
466
+ 2025-05-04 19:21:42,664-INFO: Log [train] batch <ep226_idx14_rank0> to tensorboard ...
467
+ 2025-05-04 19:21:42,839-INFO: Finish!
468
+ 2025-05-04 19:22:09,067-INFO: Log [train] batch <ep227_idx14_rank0> to tensorboard ...
469
+ 2025-05-04 19:22:09,257-INFO: Finish!
470
+ 2025-05-04 19:22:35,439-INFO: Log [train] batch <ep228_idx14_rank0> to tensorboard ...
471
+ 2025-05-04 19:22:35,617-INFO: Finish!
472
+ 2025-05-04 19:23:02,418-INFO: Log [train] batch <ep229_idx14_rank0> to tensorboard ...
473
+ 2025-05-04 19:23:02,633-INFO: Finish!
474
+ 2025-05-04 19:23:52,955-INFO: Log [train] batch <ep230_idx14_rank0> to tensorboard ...
475
+ 2025-05-04 19:23:53,149-INFO: Finish!
476
+ 2025-05-04 19:24:18,578-INFO: Log [train] batch <ep231_idx14_rank0> to tensorboard ...
477
+ 2025-05-04 19:24:18,780-INFO: Finish!
478
+ 2025-05-04 19:24:44,876-INFO: Log [train] batch <ep232_idx14_rank0> to tensorboard ...
479
+ 2025-05-04 19:24:45,128-INFO: Finish!
480
+ 2025-05-04 19:25:10,492-INFO: Log [train] batch <ep233_idx14_rank0> to tensorboard ...
481
+ 2025-05-04 19:25:10,674-INFO: Finish!
482
+ 2025-05-04 19:25:36,700-INFO: Log [train] batch <ep234_idx14_rank0> to tensorboard ...
483
+ 2025-05-04 19:25:36,879-INFO: Finish!
484
+ 2025-05-04 19:26:26,799-INFO: Log [train] batch <ep235_idx14_rank0> to tensorboard ...
485
+ 2025-05-04 19:26:27,004-INFO: Finish!
486
+ 2025-05-04 19:26:52,425-INFO: Log [train] batch <ep236_idx14_rank0> to tensorboard ...
487
+ 2025-05-04 19:26:52,612-INFO: Finish!
488
+ 2025-05-04 19:27:18,372-INFO: Log [train] batch <ep237_idx14_rank0> to tensorboard ...
489
+ 2025-05-04 19:27:18,542-INFO: Finish!
490
+ 2025-05-04 19:27:43,878-INFO: Log [train] batch <ep238_idx14_rank0> to tensorboard ...
491
+ 2025-05-04 19:27:44,048-INFO: Finish!
492
+ 2025-05-04 19:28:09,611-INFO: Log [train] batch <ep239_idx14_rank0> to tensorboard ...
493
+ 2025-05-04 19:28:09,822-INFO: Finish!
494
+ 2025-05-04 19:29:25,788-INFO: Log [train] batch <ep240_idx14_rank0> to tensorboard ...
495
+ 2025-05-04 19:29:25,965-INFO: Finish!
496
+ 2025-05-04 19:29:51,770-INFO: Log [train] batch <ep241_idx14_rank0> to tensorboard ...
497
+ 2025-05-04 19:29:51,959-INFO: Finish!
498
+ 2025-05-04 19:30:18,058-INFO: Log [train] batch <ep242_idx14_rank0> to tensorboard ...
499
+ 2025-05-04 19:30:18,251-INFO: Finish!
500
+ 2025-05-04 19:30:42,496-INFO: Log [train] batch <ep243_idx14_rank0> to tensorboard ...
501
+ 2025-05-04 19:30:42,668-INFO: Finish!
502
+ 2025-05-04 19:31:07,347-INFO: Log [train] batch <ep244_idx14_rank0> to tensorboard ...
503
+ 2025-05-04 19:31:07,506-INFO: Finish!
504
+ 2025-05-04 19:31:53,934-INFO: Log [train] batch <ep245_idx14_rank0> to tensorboard ...
505
+ 2025-05-04 19:31:54,108-INFO: Finish!
506
+ 2025-05-04 19:32:18,757-INFO: Log [train] batch <ep246_idx14_rank0> to tensorboard ...
507
+ 2025-05-04 19:32:18,940-INFO: Finish!
508
+ 2025-05-04 19:32:43,319-INFO: Log [train] batch <ep247_idx14_rank0> to tensorboard ...
509
+ 2025-05-04 19:32:43,522-INFO: Finish!
510
+ 2025-05-04 19:33:08,199-INFO: Log [train] batch <ep248_idx14_rank0> to tensorboard ...
511
+ 2025-05-04 19:33:08,354-INFO: Finish!
512
+ 2025-05-04 19:33:32,362-INFO: Log [train] batch <ep249_idx14_rank0> to tensorboard ...
513
+ 2025-05-04 19:33:32,536-INFO: Finish!
514
+ 2025-05-04 19:34:19,610-INFO: Log [train] batch <ep250_idx14_rank0> to tensorboard ...
515
+ 2025-05-04 19:34:19,784-INFO: Finish!
516
+ 2025-05-04 19:34:44,125-INFO: Log [train] batch <ep251_idx14_rank0> to tensorboard ...
517
+ 2025-05-04 19:34:44,286-INFO: Finish!
518
+ 2025-05-04 19:35:08,469-INFO: Log [train] batch <ep252_idx14_rank0> to tensorboard ...
519
+ 2025-05-04 19:35:08,643-INFO: Finish!
520
+ 2025-05-04 19:35:32,796-INFO: Log [train] batch <ep253_idx14_rank0> to tensorboard ...
521
+ 2025-05-04 19:35:32,978-INFO: Finish!
522
+ 2025-05-04 19:35:57,078-INFO: Log [train] batch <ep254_idx14_rank0> to tensorboard ...
523
+ 2025-05-04 19:35:57,251-INFO: Finish!
524
+ 2025-05-04 19:36:41,075-INFO: Log [train] batch <ep255_idx14_rank0> to tensorboard ...
525
+ 2025-05-04 19:36:41,242-INFO: Finish!
526
+ 2025-05-04 19:37:04,964-INFO: Log [train] batch <ep256_idx14_rank0> to tensorboard ...
527
+ 2025-05-04 19:37:05,118-INFO: Finish!
528
+ 2025-05-04 19:37:27,704-INFO: Log [train] batch <ep257_idx14_rank0> to tensorboard ...
529
+ 2025-05-04 19:37:27,857-INFO: Finish!
530
+ 2025-05-04 19:37:50,333-INFO: Log [train] batch <ep258_idx14_rank0> to tensorboard ...
531
+ 2025-05-04 19:37:50,487-INFO: Finish!
532
+ 2025-05-04 19:38:13,011-INFO: Log [train] batch <ep259_idx14_rank0> to tensorboard ...
533
+ 2025-05-04 19:38:13,204-INFO: Finish!
534
+ 2025-05-04 19:38:56,763-INFO: Log [train] batch <ep260_idx14_rank0> to tensorboard ...
535
+ 2025-05-04 19:38:56,920-INFO: Finish!
536
+ 2025-05-04 19:39:17,737-INFO: Log [train] batch <ep261_idx14_rank0> to tensorboard ...
537
+ 2025-05-04 19:39:17,901-INFO: Finish!
538
+ 2025-05-04 19:39:38,577-INFO: Log [train] batch <ep262_idx14_rank0> to tensorboard ...
539
+ 2025-05-04 19:39:38,705-INFO: Finish!
540
+ 2025-05-04 19:39:59,211-INFO: Log [train] batch <ep263_idx14_rank0> to tensorboard ...
541
+ 2025-05-04 19:39:59,341-INFO: Finish!
542
+ 2025-05-04 19:40:20,098-INFO: Log [train] batch <ep264_idx14_rank0> to tensorboard ...
543
+ 2025-05-04 19:40:20,225-INFO: Finish!
544
+ 2025-05-04 19:40:59,800-INFO: Log [train] batch <ep265_idx14_rank0> to tensorboard ...
545
+ 2025-05-04 19:40:59,936-INFO: Finish!
546
+ 2025-05-04 19:41:21,424-INFO: Log [train] batch <ep266_idx14_rank0> to tensorboard ...
547
+ 2025-05-04 19:41:21,550-INFO: Finish!
548
+ 2025-05-04 19:41:42,055-INFO: Log [train] batch <ep267_idx14_rank0> to tensorboard ...
549
+ 2025-05-04 19:41:42,195-INFO: Finish!
550
+ 2025-05-04 19:42:02,328-INFO: Log [train] batch <ep268_idx14_rank0> to tensorboard ...
551
+ 2025-05-04 19:42:02,462-INFO: Finish!
552
+ 2025-05-04 19:42:22,506-INFO: Log [train] batch <ep269_idx14_rank0> to tensorboard ...
553
+ 2025-05-04 19:42:22,681-INFO: Finish!
554
+ 2025-05-04 19:43:21,215-INFO: Log [train] batch <ep270_idx14_rank0> to tensorboard ...
555
+ 2025-05-04 19:43:21,352-INFO: Finish!
556
+ 2025-05-04 19:43:41,403-INFO: Log [train] batch <ep271_idx14_rank0> to tensorboard ...
557
+ 2025-05-04 19:43:41,563-INFO: Finish!
558
+ 2025-05-04 19:44:02,276-INFO: Log [train] batch <ep272_idx14_rank0> to tensorboard ...
559
+ 2025-05-04 19:44:02,448-INFO: Finish!
560
+ 2025-05-04 19:44:22,968-INFO: Log [train] batch <ep273_idx14_rank0> to tensorboard ...
561
+ 2025-05-04 19:44:23,093-INFO: Finish!
562
+ 2025-05-04 19:44:43,565-INFO: Log [train] batch <ep274_idx14_rank0> to tensorboard ...
563
+ 2025-05-04 19:44:43,697-INFO: Finish!
564
+ 2025-05-04 19:45:24,825-INFO: Log [train] batch <ep275_idx14_rank0> to tensorboard ...
565
+ 2025-05-04 19:45:24,985-INFO: Finish!
566
+ 2025-05-04 19:45:46,096-INFO: Log [train] batch <ep276_idx14_rank0> to tensorboard ...
567
+ 2025-05-04 19:45:46,250-INFO: Finish!
568
+ 2025-05-04 19:46:06,843-INFO: Log [train] batch <ep277_idx14_rank0> to tensorboard ...
569
+ 2025-05-04 19:46:06,986-INFO: Finish!
570
+ 2025-05-04 19:46:27,887-INFO: Log [train] batch <ep278_idx14_rank0> to tensorboard ...
571
+ 2025-05-04 19:46:28,022-INFO: Finish!
572
+ 2025-05-04 19:46:49,421-INFO: Log [train] batch <ep279_idx14_rank0> to tensorboard ...
573
+ 2025-05-04 19:46:49,581-INFO: Finish!
574
+ 2025-05-04 19:47:30,203-INFO: Log [train] batch <ep280_idx14_rank0> to tensorboard ...
575
+ 2025-05-04 19:47:30,345-INFO: Finish!
576
+ 2025-05-04 19:48:00,887-INFO: Log [train] batch <ep281_idx14_rank0> to tensorboard ...
577
+ 2025-05-04 19:48:01,015-INFO: Finish!
578
+ 2025-05-04 19:48:44,292-INFO: Log [train] batch <ep282_idx14_rank0> to tensorboard ...
579
+ 2025-05-04 19:48:44,421-INFO: Finish!
580
+ 2025-05-04 19:49:22,154-INFO: Log [train] batch <ep283_idx14_rank0> to tensorboard ...
581
+ 2025-05-04 19:49:22,356-INFO: Finish!
582
+ 2025-05-04 19:49:47,044-INFO: Log [train] batch <ep284_idx14_rank0> to tensorboard ...
583
+ 2025-05-04 19:49:47,297-INFO: Finish!
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_1:2025-05-04T16-59-18.txt ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ 2025-05-04 16:59:18,669-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 16:59:18,669-INFO: ***** Configing Model *****
3
+ 2025-05-04 16:59:20,030-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:00:09,199-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:00:29,290-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:00:29,290-INFO: ***** Configing Data *****
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_1:2025-05-04T17-00-59.txt ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ 2025-05-04 17:00:59,650-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:00:59,651-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:00:59,967-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:01:45,676-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:01:59,655-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:01:59,655-INFO: ***** Configing Data *****
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_1:2025-05-04T17-03-42.txt ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-05-04 17:03:42,340-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:03:42,340-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:03:42,715-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:04:28,409-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:04:43,040-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:04:43,041-INFO: ***** Configing Data *****
7
+ 2025-05-04 17:04:43,432-INFO: train, WebVid, 300
8
+ 2025-05-04 17:04:43,432-INFO: ***** Configing Trainer *****
9
+ 2025-05-04 17:04:43,434-INFO: Caution: Saving checkpoints every n train steps without deleting. This might require some free space.
10
+ 2025-05-04 17:04:43,530-INFO: ***** Running the Loop *****
11
+ 2025-05-04 17:04:43,530-INFO: <Training in DDPSharded Mode>
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_1:2025-05-04T17-07-39.txt ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-05-04 17:07:39,241-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:07:39,241-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:07:39,551-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:08:24,805-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:08:38,121-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:08:38,122-INFO: ***** Configing Data *****
7
+ 2025-05-04 17:08:38,438-INFO: train, WebVid, 300
8
+ 2025-05-04 17:08:38,438-INFO: ***** Configing Trainer *****
9
+ 2025-05-04 17:08:38,440-INFO: Caution: Saving checkpoints every n train steps without deleting. This might require some free space.
10
+ 2025-05-04 17:08:38,533-INFO: ***** Running the Loop *****
11
+ 2025-05-04 17:08:38,534-INFO: <Training in DDPSharded Mode>
12
+ 2025-05-04 17:08:45,415-INFO: @Training [1516] Full Paramters.
13
+ 2025-05-04 17:08:45,416-INFO: @Training [51] Paramters for Image_proj_model.
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_1:2025-05-04T17-10-25.txt ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-05-04 17:10:25,759-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:10:25,760-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:10:26,077-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:11:09,684-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:11:23,860-INFO: Running on 2=2x1 GPUs
6
+ 2025-05-04 17:11:23,861-INFO: ***** Configing Data *****
7
+ 2025-05-04 17:11:24,203-INFO: train, WebVid, 300
8
+ 2025-05-04 17:11:24,204-INFO: ***** Configing Trainer *****
9
+ 2025-05-04 17:11:24,206-INFO: Caution: Saving checkpoints every n train steps without deleting. This might require some free space.
10
+ 2025-05-04 17:11:24,269-INFO: ***** Running the Loop *****
11
+ 2025-05-04 17:11:24,269-INFO: <Training in DDPSharded Mode>
12
+ 2025-05-04 17:11:29,755-INFO: @Training [1516] Full Paramters.
13
+ 2025-05-04 17:11:29,756-INFO: @Training [51] Paramters for Image_proj_model.
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_1:2025-05-04T17-12-30.txt ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-05-04 17:12:30,927-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:12:30,928-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:12:31,297-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:13:16,358-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:13:30,518-INFO: Running on 2=2x1 GPUs
6
+ 2025-05-04 17:13:30,519-INFO: ***** Configing Data *****
7
+ 2025-05-04 17:13:30,977-INFO: train, WebVid, 300
8
+ 2025-05-04 17:13:30,977-INFO: ***** Configing Trainer *****
9
+ 2025-05-04 17:13:30,979-INFO: Caution: Saving checkpoints every n train steps without deleting. This might require some free space.
10
+ 2025-05-04 17:13:31,057-INFO: ***** Running the Loop *****
11
+ 2025-05-04 17:13:31,057-INFO: <Training in DDPSharded Mode>
12
+ 2025-05-04 17:13:38,255-INFO: @Training [1516] Full Paramters.
13
+ 2025-05-04 17:13:38,256-INFO: @Training [51] Paramters for Image_proj_model.
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_1:2025-05-04T17-17-13.txt ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-05-04 17:17:13,848-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:17:13,848-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:17:14,272-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:17:59,165-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:18:14,191-INFO: Running on 2=2x1 GPUs
6
+ 2025-05-04 17:18:14,191-INFO: ***** Configing Data *****
7
+ 2025-05-04 17:18:14,523-INFO: train, WebVid, 300
8
+ 2025-05-04 17:18:14,523-INFO: ***** Configing Trainer *****
9
+ 2025-05-04 17:18:14,524-INFO: Caution: Saving checkpoints every n train steps without deleting. This might require some free space.
10
+ 2025-05-04 17:18:14,599-INFO: ***** Running the Loop *****
11
+ 2025-05-04 17:18:14,599-INFO: <Training in DDPSharded Mode>
12
+ 2025-05-04 17:18:22,567-INFO: @Training [1516] Full Paramters.
13
+ 2025-05-04 17:18:22,568-INFO: @Training [51] Paramters for Image_proj_model.
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_2:2025-05-04T16-59-18.txt ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ 2025-05-04 16:59:18,755-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 16:59:18,756-INFO: ***** Configing Model *****
3
+ 2025-05-04 16:59:20,030-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:00:08,938-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:00:29,355-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:00:29,355-INFO: ***** Configing Data *****
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_2:2025-05-04T17-00-59.txt ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ 2025-05-04 17:00:59,834-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:00:59,834-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:01:00,244-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:01:46,501-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:02:00,113-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:02:00,113-INFO: ***** Configing Data *****
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_2:2025-05-04T17-03-42.txt ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-05-04 17:03:42,382-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:03:42,382-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:03:42,713-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:04:28,601-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:04:40,805-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:04:40,805-INFO: ***** Configing Data *****
7
+ 2025-05-04 17:04:41,159-INFO: train, WebVid, 300
8
+ 2025-05-04 17:04:41,159-INFO: ***** Configing Trainer *****
9
+ 2025-05-04 17:04:41,162-INFO: Caution: Saving checkpoints every n train steps without deleting. This might require some free space.
10
+ 2025-05-04 17:04:41,259-INFO: ***** Running the Loop *****
11
+ 2025-05-04 17:04:41,259-INFO: <Training in DDPSharded Mode>
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_2:2025-05-04T17-07-39.txt ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-05-04 17:07:39,130-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:07:39,131-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:07:39,473-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:08:24,522-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:08:37,012-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:08:37,013-INFO: ***** Configing Data *****
7
+ 2025-05-04 17:08:37,401-INFO: train, WebVid, 300
8
+ 2025-05-04 17:08:37,401-INFO: ***** Configing Trainer *****
9
+ 2025-05-04 17:08:37,403-INFO: Caution: Saving checkpoints every n train steps without deleting. This might require some free space.
10
+ 2025-05-04 17:08:37,484-INFO: ***** Running the Loop *****
11
+ 2025-05-04 17:08:37,484-INFO: <Training in DDPSharded Mode>
12
+ 2025-05-04 17:08:45,322-INFO: @Training [1516] Full Paramters.
13
+ 2025-05-04 17:08:45,323-INFO: @Training [51] Paramters for Image_proj_model.
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_3:2025-05-04T16-59-18.txt ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ 2025-05-04 16:59:18,669-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 16:59:18,670-INFO: ***** Configing Model *****
3
+ 2025-05-04 16:59:20,029-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:00:08,808-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:00:29,636-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:00:29,637-INFO: ***** Configing Data *****
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_3:2025-05-04T17-00-59.txt ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ 2025-05-04 17:00:59,850-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:00:59,850-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:01:00,228-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:01:46,214-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:02:00,326-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:02:00,327-INFO: ***** Configing Data *****
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_3:2025-05-04T17-03-42.txt ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-05-04 17:03:42,328-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:03:42,329-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:03:42,713-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:04:28,671-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:04:43,756-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:04:43,756-INFO: ***** Configing Data *****
7
+ 2025-05-04 17:04:44,102-INFO: train, WebVid, 300
8
+ 2025-05-04 17:04:44,102-INFO: ***** Configing Trainer *****
9
+ 2025-05-04 17:04:44,103-INFO: Caution: Saving checkpoints every n train steps without deleting. This might require some free space.
10
+ 2025-05-04 17:04:44,167-INFO: ***** Running the Loop *****
11
+ 2025-05-04 17:04:44,167-INFO: <Training in DDPSharded Mode>
ww_training_128_4frame_v1.0_franka_4frame/loginfo/log_3:2025-05-04T17-07-39.txt ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ 2025-05-04 17:07:39,191-INFO: @lightning version: 1.9.3 [>=1.8 required]
2
+ 2025-05-04 17:07:39,192-INFO: ***** Configing Model *****
3
+ 2025-05-04 17:07:39,496-INFO: LatentVisualDiffusion: Running in v-prediction mode
4
+ 2025-05-04 17:08:24,531-INFO: >>> Load weights from pretrained checkpoint
5
+ 2025-05-04 17:08:36,562-INFO: Running on 4=4x1 GPUs
6
+ 2025-05-04 17:08:36,563-INFO: ***** Configing Data *****
7
+ 2025-05-04 17:08:36,886-INFO: train, WebVid, 300
8
+ 2025-05-04 17:08:36,886-INFO: ***** Configing Trainer *****
9
+ 2025-05-04 17:08:36,888-INFO: Caution: Saving checkpoints every n train steps without deleting. This might require some free space.
10
+ 2025-05-04 17:08:36,948-INFO: ***** Running the Loop *****
11
+ 2025-05-04 17:08:36,948-INFO: <Training in DDPSharded Mode>
12
+ 2025-05-04 17:08:45,527-INFO: @Training [1516] Full Paramters.
13
+ 2025-05-04 17:08:45,528-INFO: @Training [51] Paramters for Image_proj_model.
ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_0/events.out.tfevents.1746349729.dgx-049.4189318.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92262918fbf807560f8f91ba41cc5ec0463baa48547f648f2bb5bbc208099ce9
3
+ size 1816
ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_0/hparams.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ {}
ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_1/events.out.tfevents.1746349895.dgx-049.43590.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4225eac5dc7f209bed9b4231d863d51e363c22f6d8d8717d6135953075abe381
3
+ size 118650
ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_1/hparams.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ {}
ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_2/events.out.tfevents.1746350023.dgx-049.51515.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51ef94dc2030da235e58b87d1238ea603f9e909fa27f1aab5a83bc80fdf3c6f5
3
+ size 696395
ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_2/hparams.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ {}
ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_3/events.out.tfevents.1746350306.dgx-049.141757.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ace6c12a402c04c4b1cf5429c804d81a8be6c312c23ed40e74d095c594dad97c
3
+ size 33115227
ww_training_128_4frame_v1.0_franka_4frame/tensorboard/version_3/hparams.yaml ADDED
@@ -0,0 +1 @@
 
 
1
+ {}