Junyi42 commited on
Commit
9f328a9
·
verified ·
1 Parent(s): 32a25e1

Upload checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins

Browse files
checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/wandb/offline-run-20260125_150523-checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins-run0/files/output.log CHANGED
@@ -1,3 +1,173 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  wandb: Detected [huggingface_hub.inference] in use.
2
  wandb: Use W&B Weave for improved LLM call tracing. Install Weave with `pip install weave` then add `import weave` to the top of your script.
3
  wandb: For more information, check out the docs at: https://weave-docs.wandb.ai/
@@ -1094,197 +1264,6 @@ wandb: For more information, check out the docs at: https://weave-docs.wandb.ai/
1094
  [2026-01-25 16:03:40] (step=0001083) Train Loss mse: 0.0000, Train Loss ce: 0.2555, Train Steps/Sec: 0.39,
1095
  [2026-01-25 16:03:42] (step=0001084) Train Loss mse: 0.0000, Train Loss ce: 0.2940, Train Steps/Sec: 0.42,
1096
  [2026-01-25 16:03:45] (step=0001085) Train Loss mse: 0.0000, Train Loss ce: 0.2703, Train Steps/Sec: 0.33,
1097
- FullyShardedDataParallel(
1098
- (_fsdp_wrapped_module): Bagel(
1099
- (language_model): Qwen2ForCausalLM(
1100
- (model): Qwen2Model(
1101
- (embed_tokens): Embedding(152064, 3584)
1102
- (layers): ModuleList(
1103
- (0-27): 28 x FullyShardedDataParallel(
1104
- (_fsdp_wrapped_module): CheckpointWrapper(
1105
- (_checkpoint_wrapped_module): Qwen2MoTDecoderLayer(
1106
- (self_attn): PackedAttentionMoT(
1107
- (q_proj): Linear(in_features=3584, out_features=3584, bias=True)
1108
- (k_proj): Linear(in_features=3584, out_features=512, bias=True)
1109
- (v_proj): Linear(in_features=3584, out_features=512, bias=True)
1110
- (o_proj): Linear(in_features=3584, out_features=3584, bias=False)
1111
- (q_norm): Qwen2RMSNorm((128,), eps=1e-06)
1112
- (k_norm): Qwen2RMSNorm((128,), eps=1e-06)
1113
- (q_norm_moe_gen): Qwen2RMSNorm((128,), eps=1e-06)
1114
- (k_norm_moe_gen): Qwen2RMSNorm((128,), eps=1e-06)
1115
- (q_proj_moe_gen): Linear(in_features=3584, out_features=3584, bias=True)
1116
- (k_proj_moe_gen): Linear(in_features=3584, out_features=512, bias=True)
1117
- (v_proj_moe_gen): Linear(in_features=3584, out_features=512, bias=True)
1118
- (o_proj_moe_gen): Linear(in_features=3584, out_features=3584, bias=False)
1119
- )
1120
- (mlp): Qwen2MLP(
1121
- (gate_proj): Linear(in_features=3584, out_features=18944, bias=False)
1122
- (up_proj): Linear(in_features=3584, out_features=18944, bias=False)
1123
- (down_proj): Linear(in_features=18944, out_features=3584, bias=False)
1124
- (act_fn): SiLU()
1125
- )
1126
- (mlp_moe_gen): Qwen2MLP(
1127
- (gate_proj): Linear(in_features=3584, out_features=18944, bias=False)
1128
- (up_proj): Linear(in_features=3584, out_features=18944, bias=False)
1129
- (down_proj): Linear(in_features=18944, out_features=3584, bias=False)
1130
- (act_fn): SiLU()
1131
- )
1132
- (input_layernorm): Qwen2RMSNorm((3584,), eps=1e-06)
1133
- (input_layernorm_moe_gen): Qwen2RMSNorm((3584,), eps=1e-06)
1134
- (post_attention_layernorm): Qwen2RMSNorm((3584,), eps=1e-06)
1135
- (post_attention_layernorm_moe_gen): Qwen2RMSNorm((3584,), eps=1e-06)
1136
- )
1137
- )
1138
- )
1139
- )
1140
- (norm): Qwen2RMSNorm((3584,), eps=1e-06)
1141
- (norm_moe_gen): Qwen2RMSNorm((3584,), eps=1e-06)
1142
- (rotary_emb): Qwen2RotaryEmbedding()
1143
- )
1144
- (lm_head): Linear(in_features=3584, out_features=152064, bias=False)
1145
- )
1146
- (vit_model): SiglipVisionModel(
1147
- (vision_model): FullyShardedDataParallel(
1148
- (_fsdp_wrapped_module): SiglipVisionTransformer(
1149
- (embeddings): SiglipVisionEmbeddings(
1150
- (position_embedding): Embedding(4900, 1152)
1151
- (patch_embedding): Linear(in_features=588, out_features=1152, bias=True)
1152
- )
1153
- (encoder): SiglipEncoder(
1154
- (layers): ModuleList(
1155
- (0-25): 26 x FullyShardedDataParallel(
1156
- (_fsdp_wrapped_module): CheckpointWrapper(
1157
- (_checkpoint_wrapped_module): SiglipEncoderLayer(
1158
- (self_attn): SiglipFlashAttention2(
1159
- (k_proj): Linear(in_features=1152, out_features=1152, bias=True)
1160
- (v_proj): Linear(in_features=1152, out_features=1152, bias=True)
1161
- (q_proj): Linear(in_features=1152, out_features=1152, bias=True)
1162
- (out_proj): Linear(in_features=1152, out_features=1152, bias=True)
1163
- )
1164
- (layer_norm1): LayerNorm((1152,), eps=1e-06, elementwise_affine=True)
1165
- (mlp): SiglipMLP(
1166
- (activation_fn): PytorchGELUTanh()
1167
- (fc1): Linear(in_features=1152, out_features=4304, bias=True)
1168
- (fc2): Linear(in_features=4304, out_features=1152, bias=True)
1169
- )
1170
- (layer_norm2): LayerNorm((1152,), eps=1e-06, elementwise_affine=True)
1171
- )
1172
- )
1173
- )
1174
- )
1175
- )
1176
- (post_layernorm): LayerNorm((1152,), eps=1e-06, elementwise_affine=True)
1177
- )
1178
- )
1179
- )
1180
- (connector): FullyShardedDataParallel(
1181
- (_fsdp_wrapped_module): CheckpointWrapper(
1182
- (_checkpoint_wrapped_module): MLPconnector(
1183
- (activation_fn): PytorchGELUTanh()
1184
- (fc1): Linear(in_features=1152, out_features=3584, bias=True)
1185
- (fc2): Linear(in_features=3584, out_features=3584, bias=True)
1186
- )
1187
- )
1188
- )
1189
- (vit_pos_embed): FullyShardedDataParallel(
1190
- (_fsdp_wrapped_module): PositionEmbedding()
1191
- )
1192
- )
1193
- )
1194
- _flat_param True
1195
- language_model.model.layers.0._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1196
- language_model.model.layers.1._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1197
- language_model.model.layers.2._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1198
- language_model.model.layers.3._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1199
- language_model.model.layers.4._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1200
- language_model.model.layers.5._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1201
- language_model.model.layers.6._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1202
- language_model.model.layers.7._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1203
- language_model.model.layers.8._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1204
- language_model.model.layers.9._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1205
- language_model.model.layers.10._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1206
- language_model.model.layers.11._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1207
- language_model.model.layers.12._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1208
- language_model.model.layers.13._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1209
- language_model.model.layers.14._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1210
- language_model.model.layers.15._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1211
- language_model.model.layers.16._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1212
- language_model.model.layers.17._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1213
- language_model.model.layers.18._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1214
- language_model.model.layers.19._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1215
- language_model.model.layers.20._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1216
- language_model.model.layers.21._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1217
- language_model.model.layers.22._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1218
- language_model.model.layers.23._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1219
- language_model.model.layers.24._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1220
- language_model.model.layers.25._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1221
- language_model.model.layers.26._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1222
- language_model.model.layers.27._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1223
- vit_model.vision_model._fsdp_wrapped_module._flat_param True
1224
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.0._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1225
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.1._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1226
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.2._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1227
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.3._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1228
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.4._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1229
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.5._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1230
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.6._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1231
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.7._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1232
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.8._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1233
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.9._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1234
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.10._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1235
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.11._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1236
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.12._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1237
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.13._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1238
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.14._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1239
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.15._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1240
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.16._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1241
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.17._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1242
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.18._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1243
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.19._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1244
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.20._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1245
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.21._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1246
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.22._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1247
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.23._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1248
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.24._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1249
- vit_model.vision_model._fsdp_wrapped_module.encoder.layers.25._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1250
- connector._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
1251
- vit_pos_embed._fsdp_wrapped_module._flat_param False
1252
- Preparing Dataset vlm_gym_colorization_celoss_no_mse/vlm_gym_colorization_train
1253
- base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step0
1254
- Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
1255
- [eval debug] first 3 batch fingerprints:
1256
- fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1257
- fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1258
- fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1259
- ce_avg: 0.8449353575706482, mse_avg: 0.0
1260
- base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step500
1261
- Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
1262
- [eval debug] first 3 batch fingerprints:
1263
- fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1264
- fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1265
- fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1266
- ce_avg: 0.2859387993812561, mse_avg: 0.0
1267
- base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step1000
1268
- Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
1269
- [eval debug] first 3 batch fingerprints:
1270
- fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1271
- fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1272
- fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1273
- ce_avg: 0.46788886189460754, mse_avg: 0.0
1274
- base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step1500
1275
- Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
1276
- [eval debug] first 3 batch fingerprints:
1277
- fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1278
- fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1279
- fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1280
- ce_avg: 0.6231057643890381, mse_avg: 0.0
1281
- base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step2000
1282
- Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
1283
- [eval debug] first 3 batch fingerprints:
1284
- fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1285
- fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1286
- fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1287
- ce_avg: 0.6934272646903992, mse_avg: 0.0
1288
  [2026-01-25 16:03:48] (step=0001086) Train Loss mse: 0.0000, Train Loss ce: 0.2461, Train Steps/Sec: 0.38,
1289
  [2026-01-25 16:03:51] (step=0001087) Train Loss mse: 0.0000, Train Loss ce: 0.2672, Train Steps/Sec: 0.40,
1290
  [2026-01-25 16:03:53] (step=0001088) Train Loss mse: 0.0000, Train Loss ce: 0.2795, Train Steps/Sec: 0.34,
@@ -1316,6 +1295,27 @@ ce_avg: 0.6934272646903992, mse_avg: 0.0
1316
  [2026-01-25 16:05:05] (step=0001114) Train Loss mse: 0.0000, Train Loss ce: 0.2708, Train Steps/Sec: 0.40,
1317
  [2026-01-25 16:05:08] (step=0001115) Train Loss mse: 0.0000, Train Loss ce: 0.2779, Train Steps/Sec: 0.37,
1318
  [2026-01-25 16:05:11] (step=0001116) Train Loss mse: 0.0000, Train Loss ce: 0.2583, Train Steps/Sec: 0.39,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1319
  [2026-01-25 16:05:14] (step=0001117) Train Loss mse: 0.0000, Train Loss ce: 0.2621, Train Steps/Sec: 0.33,
1320
  [2026-01-25 16:05:16] (step=0001118) Train Loss mse: 0.0000, Train Loss ce: 0.2592, Train Steps/Sec: 0.35,
1321
  [2026-01-25 16:05:19] (step=0001119) Train Loss mse: 0.0000, Train Loss ce: 0.2544, Train Steps/Sec: 0.35,
@@ -2782,27 +2782,6 @@ ce_avg: 0.6934272646903992, mse_avg: 0.0
2782
  [2026-01-25 17:13:45] (step=0002580) Train Loss mse: 0.0000, Train Loss ce: 0.2436, Train Steps/Sec: 0.31,
2783
  [2026-01-25 17:13:48] (step=0002581) Train Loss mse: 0.0000, Train Loss ce: 0.2514, Train Steps/Sec: 0.42,
2784
  [2026-01-25 17:13:51] (step=0002582) Train Loss mse: 0.0000, Train Loss ce: 0.2485, Train Steps/Sec: 0.29,
2785
- base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step2500
2786
- Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
2787
- [eval debug] first 3 batch fingerprints:
2788
- fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2789
- fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2790
- fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2791
- ce_avg: 0.719638466835022, mse_avg: 0.0
2792
- base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step3000
2793
- Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
2794
- [eval debug] first 3 batch fingerprints:
2795
- fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2796
- fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2797
- fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2798
- ce_avg: 0.7213400602340698, mse_avg: 0.0
2799
- base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step3500
2800
- Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
2801
- [eval debug] first 3 batch fingerprints:
2802
- fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2803
- fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2804
- fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2805
- ce_avg: 0.6960676312446594, mse_avg: 0.0
2806
  [2026-01-25 17:13:54] (step=0002583) Train Loss mse: 0.0000, Train Loss ce: 0.2590, Train Steps/Sec: 0.37,
2807
  [2026-01-25 17:13:56] (step=0002584) Train Loss mse: 0.0000, Train Loss ce: 0.2530, Train Steps/Sec: 0.40,
2808
  [2026-01-25 17:14:00] (step=0002585) Train Loss mse: 0.0000, Train Loss ce: 0.2502, Train Steps/Sec: 0.32,
@@ -2842,6 +2821,27 @@ ce_avg: 0.6960676312446594, mse_avg: 0.0
2842
  [2026-01-25 17:15:34] (step=0002619) Train Loss mse: 0.0000, Train Loss ce: 0.2615, Train Steps/Sec: 0.39,
2843
  [2026-01-25 17:15:37] (step=0002620) Train Loss mse: 0.0000, Train Loss ce: 0.2592, Train Steps/Sec: 0.34,
2844
  [2026-01-25 17:15:39] (step=0002621) Train Loss mse: 0.0000, Train Loss ce: 0.2630, Train Steps/Sec: 0.38,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2845
  [2026-01-25 17:15:42] (step=0002622) Train Loss mse: 0.0000, Train Loss ce: 0.2424, Train Steps/Sec: 0.38,
2846
  [2026-01-25 17:15:45] (step=0002623) Train Loss mse: 0.0000, Train Loss ce: 0.2457, Train Steps/Sec: 0.37,
2847
  [2026-01-25 17:15:48] (step=0002624) Train Loss mse: 0.0000, Train Loss ce: 0.2660, Train Steps/Sec: 0.36,
@@ -3800,20 +3800,6 @@ ce_avg: 0.6960676312446594, mse_avg: 0.0
3800
  [2026-01-25 18:00:35] (step=0003577) Train Loss mse: 0.0000, Train Loss ce: 0.2664, Train Steps/Sec: 0.29,
3801
  [2026-01-25 18:00:38] (step=0003578) Train Loss mse: 0.0000, Train Loss ce: 0.2234, Train Steps/Sec: 0.35,
3802
  [2026-01-25 18:00:41] (step=0003579) Train Loss mse: 0.0000, Train Loss ce: 0.2551, Train Steps/Sec: 0.36,
3803
- base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step4000
3804
- Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
3805
- [eval debug] first 3 batch fingerprints:
3806
- fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
3807
- fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
3808
- fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
3809
- ce_avg: 0.7003546953201294, mse_avg: 0.0
3810
- base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step4500
3811
- Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
3812
- [eval debug] first 3 batch fingerprints:
3813
- fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
3814
- fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
3815
- fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
3816
- ce_avg: 0.6831608414649963, mse_avg: 0.0
3817
  [2026-01-25 18:00:44] (step=0003580) Train Loss mse: 0.0000, Train Loss ce: 0.2514, Train Steps/Sec: 0.37,
3818
  [2026-01-25 18:00:46] (step=0003581) Train Loss mse: 0.0000, Train Loss ce: 0.2396, Train Steps/Sec: 0.39,
3819
  [2026-01-25 18:00:49] (step=0003582) Train Loss mse: 0.0000, Train Loss ce: 0.2210, Train Steps/Sec: 0.40,
@@ -3880,6 +3866,20 @@ ce_avg: 0.6831608414649963, mse_avg: 0.0
3880
  [2026-01-25 18:03:40] (step=0003643) Train Loss mse: 0.0000, Train Loss ce: 0.2529, Train Steps/Sec: 0.37,
3881
  [2026-01-25 18:03:43] (step=0003644) Train Loss mse: 0.0000, Train Loss ce: 0.2446, Train Steps/Sec: 0.35,
3882
  [2026-01-25 18:03:46] (step=0003645) Train Loss mse: 0.0000, Train Loss ce: 0.2420, Train Steps/Sec: 0.35,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3883
  [2026-01-25 18:03:48] (step=0003646) Train Loss mse: 0.0000, Train Loss ce: 0.2638, Train Steps/Sec: 0.40,
3884
  [2026-01-25 18:03:51] (step=0003647) Train Loss mse: 0.0000, Train Loss ce: 0.2499, Train Steps/Sec: 0.38,
3885
  [2026-01-25 18:03:54] (step=0003648) Train Loss mse: 0.0000, Train Loss ce: 0.2567, Train Steps/Sec: 0.36,
 
1
+ FullyShardedDataParallel(
2
+ (_fsdp_wrapped_module): Bagel(
3
+ (language_model): Qwen2ForCausalLM(
4
+ (model): Qwen2Model(
5
+ (embed_tokens): Embedding(152064, 3584)
6
+ (layers): ModuleList(
7
+ (0-27): 28 x FullyShardedDataParallel(
8
+ (_fsdp_wrapped_module): CheckpointWrapper(
9
+ (_checkpoint_wrapped_module): Qwen2MoTDecoderLayer(
10
+ (self_attn): PackedAttentionMoT(
11
+ (q_proj): Linear(in_features=3584, out_features=3584, bias=True)
12
+ (k_proj): Linear(in_features=3584, out_features=512, bias=True)
13
+ (v_proj): Linear(in_features=3584, out_features=512, bias=True)
14
+ (o_proj): Linear(in_features=3584, out_features=3584, bias=False)
15
+ (q_norm): Qwen2RMSNorm((128,), eps=1e-06)
16
+ (k_norm): Qwen2RMSNorm((128,), eps=1e-06)
17
+ (q_norm_moe_gen): Qwen2RMSNorm((128,), eps=1e-06)
18
+ (k_norm_moe_gen): Qwen2RMSNorm((128,), eps=1e-06)
19
+ (q_proj_moe_gen): Linear(in_features=3584, out_features=3584, bias=True)
20
+ (k_proj_moe_gen): Linear(in_features=3584, out_features=512, bias=True)
21
+ (v_proj_moe_gen): Linear(in_features=3584, out_features=512, bias=True)
22
+ (o_proj_moe_gen): Linear(in_features=3584, out_features=3584, bias=False)
23
+ )
24
+ (mlp): Qwen2MLP(
25
+ (gate_proj): Linear(in_features=3584, out_features=18944, bias=False)
26
+ (up_proj): Linear(in_features=3584, out_features=18944, bias=False)
27
+ (down_proj): Linear(in_features=18944, out_features=3584, bias=False)
28
+ (act_fn): SiLU()
29
+ )
30
+ (mlp_moe_gen): Qwen2MLP(
31
+ (gate_proj): Linear(in_features=3584, out_features=18944, bias=False)
32
+ (up_proj): Linear(in_features=3584, out_features=18944, bias=False)
33
+ (down_proj): Linear(in_features=18944, out_features=3584, bias=False)
34
+ (act_fn): SiLU()
35
+ )
36
+ (input_layernorm): Qwen2RMSNorm((3584,), eps=1e-06)
37
+ (input_layernorm_moe_gen): Qwen2RMSNorm((3584,), eps=1e-06)
38
+ (post_attention_layernorm): Qwen2RMSNorm((3584,), eps=1e-06)
39
+ (post_attention_layernorm_moe_gen): Qwen2RMSNorm((3584,), eps=1e-06)
40
+ )
41
+ )
42
+ )
43
+ )
44
+ (norm): Qwen2RMSNorm((3584,), eps=1e-06)
45
+ (norm_moe_gen): Qwen2RMSNorm((3584,), eps=1e-06)
46
+ (rotary_emb): Qwen2RotaryEmbedding()
47
+ )
48
+ (lm_head): Linear(in_features=3584, out_features=152064, bias=False)
49
+ )
50
+ (vit_model): SiglipVisionModel(
51
+ (vision_model): FullyShardedDataParallel(
52
+ (_fsdp_wrapped_module): SiglipVisionTransformer(
53
+ (embeddings): SiglipVisionEmbeddings(
54
+ (position_embedding): Embedding(4900, 1152)
55
+ (patch_embedding): Linear(in_features=588, out_features=1152, bias=True)
56
+ )
57
+ (encoder): SiglipEncoder(
58
+ (layers): ModuleList(
59
+ (0-25): 26 x FullyShardedDataParallel(
60
+ (_fsdp_wrapped_module): CheckpointWrapper(
61
+ (_checkpoint_wrapped_module): SiglipEncoderLayer(
62
+ (self_attn): SiglipFlashAttention2(
63
+ (k_proj): Linear(in_features=1152, out_features=1152, bias=True)
64
+ (v_proj): Linear(in_features=1152, out_features=1152, bias=True)
65
+ (q_proj): Linear(in_features=1152, out_features=1152, bias=True)
66
+ (out_proj): Linear(in_features=1152, out_features=1152, bias=True)
67
+ )
68
+ (layer_norm1): LayerNorm((1152,), eps=1e-06, elementwise_affine=True)
69
+ (mlp): SiglipMLP(
70
+ (activation_fn): PytorchGELUTanh()
71
+ (fc1): Linear(in_features=1152, out_features=4304, bias=True)
72
+ (fc2): Linear(in_features=4304, out_features=1152, bias=True)
73
+ )
74
+ (layer_norm2): LayerNorm((1152,), eps=1e-06, elementwise_affine=True)
75
+ )
76
+ )
77
+ )
78
+ )
79
+ )
80
+ (post_layernorm): LayerNorm((1152,), eps=1e-06, elementwise_affine=True)
81
+ )
82
+ )
83
+ )
84
+ (connector): FullyShardedDataParallel(
85
+ (_fsdp_wrapped_module): CheckpointWrapper(
86
+ (_checkpoint_wrapped_module): MLPconnector(
87
+ (activation_fn): PytorchGELUTanh()
88
+ (fc1): Linear(in_features=1152, out_features=3584, bias=True)
89
+ (fc2): Linear(in_features=3584, out_features=3584, bias=True)
90
+ )
91
+ )
92
+ )
93
+ (vit_pos_embed): FullyShardedDataParallel(
94
+ (_fsdp_wrapped_module): PositionEmbedding()
95
+ )
96
+ )
97
+ )
98
+ _flat_param True
99
+ language_model.model.layers.0._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
100
+ language_model.model.layers.1._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
101
+ language_model.model.layers.2._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
102
+ language_model.model.layers.3._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
103
+ language_model.model.layers.4._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
104
+ language_model.model.layers.5._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
105
+ language_model.model.layers.6._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
106
+ language_model.model.layers.7._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
107
+ language_model.model.layers.8._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
108
+ language_model.model.layers.9._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
109
+ language_model.model.layers.10._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
110
+ language_model.model.layers.11._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
111
+ language_model.model.layers.12._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
112
+ language_model.model.layers.13._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
113
+ language_model.model.layers.14._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
114
+ language_model.model.layers.15._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
115
+ language_model.model.layers.16._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
116
+ language_model.model.layers.17._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
117
+ language_model.model.layers.18._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
118
+ language_model.model.layers.19._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
119
+ language_model.model.layers.20._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
120
+ language_model.model.layers.21._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
121
+ language_model.model.layers.22._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
122
+ language_model.model.layers.23._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
123
+ language_model.model.layers.24._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
124
+ language_model.model.layers.25._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
125
+ language_model.model.layers.26._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
126
+ language_model.model.layers.27._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
127
+ vit_model.vision_model._fsdp_wrapped_module._flat_param True
128
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.0._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
129
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.1._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
130
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.2._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
131
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.3._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
132
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.4._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
133
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.5._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
134
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.6._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
135
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.7._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
136
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.8._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
137
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.9._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
138
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.10._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
139
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.11._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
140
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.12._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
141
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.13._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
142
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.14._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
143
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.15._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
144
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.16._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
145
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.17._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
146
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.18._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
147
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.19._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
148
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.20._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
149
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.21._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
150
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.22._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
151
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.23._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
152
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.24._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
153
+ vit_model.vision_model._fsdp_wrapped_module.encoder.layers.25._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
154
+ connector._fsdp_wrapped_module._checkpoint_wrapped_module._flat_param True
155
+ vit_pos_embed._fsdp_wrapped_module._flat_param False
156
+ Preparing Dataset vlm_gym_colorization_celoss_no_mse/vlm_gym_colorization_train
157
+ base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step0
158
+ Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
159
+ [eval debug] first 3 batch fingerprints:
160
+ fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
161
+ fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
162
+ fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
163
+ ce_avg: 0.8449353575706482, mse_avg: 0.0
164
+ base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step500
165
+ Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
166
+ [eval debug] first 3 batch fingerprints:
167
+ fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
168
+ fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
169
+ fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
170
+ ce_avg: 0.2859387993812561, mse_avg: 0.0
171
  wandb: Detected [huggingface_hub.inference] in use.
172
  wandb: Use W&B Weave for improved LLM call tracing. Install Weave with `pip install weave` then add `import weave` to the top of your script.
173
  wandb: For more information, check out the docs at: https://weave-docs.wandb.ai/
 
1264
  [2026-01-25 16:03:40] (step=0001083) Train Loss mse: 0.0000, Train Loss ce: 0.2555, Train Steps/Sec: 0.39,
1265
  [2026-01-25 16:03:42] (step=0001084) Train Loss mse: 0.0000, Train Loss ce: 0.2940, Train Steps/Sec: 0.42,
1266
  [2026-01-25 16:03:45] (step=0001085) Train Loss mse: 0.0000, Train Loss ce: 0.2703, Train Steps/Sec: 0.33,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1267
  [2026-01-25 16:03:48] (step=0001086) Train Loss mse: 0.0000, Train Loss ce: 0.2461, Train Steps/Sec: 0.38,
1268
  [2026-01-25 16:03:51] (step=0001087) Train Loss mse: 0.0000, Train Loss ce: 0.2672, Train Steps/Sec: 0.40,
1269
  [2026-01-25 16:03:53] (step=0001088) Train Loss mse: 0.0000, Train Loss ce: 0.2795, Train Steps/Sec: 0.34,
 
1295
  [2026-01-25 16:05:05] (step=0001114) Train Loss mse: 0.0000, Train Loss ce: 0.2708, Train Steps/Sec: 0.40,
1296
  [2026-01-25 16:05:08] (step=0001115) Train Loss mse: 0.0000, Train Loss ce: 0.2779, Train Steps/Sec: 0.37,
1297
  [2026-01-25 16:05:11] (step=0001116) Train Loss mse: 0.0000, Train Loss ce: 0.2583, Train Steps/Sec: 0.39,
1298
+ base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step1000
1299
+ Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
1300
+ [eval debug] first 3 batch fingerprints:
1301
+ fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1302
+ fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1303
+ fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1304
+ ce_avg: 0.46788886189460754, mse_avg: 0.0
1305
+ base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step1500
1306
+ Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
1307
+ [eval debug] first 3 batch fingerprints:
1308
+ fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1309
+ fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1310
+ fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1311
+ ce_avg: 0.6231057643890381, mse_avg: 0.0
1312
+ base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step2000
1313
+ Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
1314
+ [eval debug] first 3 batch fingerprints:
1315
+ fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1316
+ fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1317
+ fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
1318
+ ce_avg: 0.6934272646903992, mse_avg: 0.0
1319
  [2026-01-25 16:05:14] (step=0001117) Train Loss mse: 0.0000, Train Loss ce: 0.2621, Train Steps/Sec: 0.33,
1320
  [2026-01-25 16:05:16] (step=0001118) Train Loss mse: 0.0000, Train Loss ce: 0.2592, Train Steps/Sec: 0.35,
1321
  [2026-01-25 16:05:19] (step=0001119) Train Loss mse: 0.0000, Train Loss ce: 0.2544, Train Steps/Sec: 0.35,
 
2782
  [2026-01-25 17:13:45] (step=0002580) Train Loss mse: 0.0000, Train Loss ce: 0.2436, Train Steps/Sec: 0.31,
2783
  [2026-01-25 17:13:48] (step=0002581) Train Loss mse: 0.0000, Train Loss ce: 0.2514, Train Steps/Sec: 0.42,
2784
  [2026-01-25 17:13:51] (step=0002582) Train Loss mse: 0.0000, Train Loss ce: 0.2485, Train Steps/Sec: 0.29,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2785
  [2026-01-25 17:13:54] (step=0002583) Train Loss mse: 0.0000, Train Loss ce: 0.2590, Train Steps/Sec: 0.37,
2786
  [2026-01-25 17:13:56] (step=0002584) Train Loss mse: 0.0000, Train Loss ce: 0.2530, Train Steps/Sec: 0.40,
2787
  [2026-01-25 17:14:00] (step=0002585) Train Loss mse: 0.0000, Train Loss ce: 0.2502, Train Steps/Sec: 0.32,
 
2821
  [2026-01-25 17:15:34] (step=0002619) Train Loss mse: 0.0000, Train Loss ce: 0.2615, Train Steps/Sec: 0.39,
2822
  [2026-01-25 17:15:37] (step=0002620) Train Loss mse: 0.0000, Train Loss ce: 0.2592, Train Steps/Sec: 0.34,
2823
  [2026-01-25 17:15:39] (step=0002621) Train Loss mse: 0.0000, Train Loss ce: 0.2630, Train Steps/Sec: 0.38,
2824
+ base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step2500
2825
+ Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
2826
+ [eval debug] first 3 batch fingerprints:
2827
+ fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2828
+ fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2829
+ fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2830
+ ce_avg: 0.719638466835022, mse_avg: 0.0
2831
+ base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step3000
2832
+ Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
2833
+ [eval debug] first 3 batch fingerprints:
2834
+ fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2835
+ fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2836
+ fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2837
+ ce_avg: 0.7213400602340698, mse_avg: 0.0
2838
+ base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step3500
2839
+ Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
2840
+ [eval debug] first 3 batch fingerprints:
2841
+ fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2842
+ fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2843
+ fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
2844
+ ce_avg: 0.6960676312446594, mse_avg: 0.0
2845
  [2026-01-25 17:15:42] (step=0002622) Train Loss mse: 0.0000, Train Loss ce: 0.2424, Train Steps/Sec: 0.38,
2846
  [2026-01-25 17:15:45] (step=0002623) Train Loss mse: 0.0000, Train Loss ce: 0.2457, Train Steps/Sec: 0.37,
2847
  [2026-01-25 17:15:48] (step=0002624) Train Loss mse: 0.0000, Train Loss ce: 0.2660, Train Steps/Sec: 0.36,
 
3800
  [2026-01-25 18:00:35] (step=0003577) Train Loss mse: 0.0000, Train Loss ce: 0.2664, Train Steps/Sec: 0.29,
3801
  [2026-01-25 18:00:38] (step=0003578) Train Loss mse: 0.0000, Train Loss ce: 0.2234, Train Steps/Sec: 0.35,
3802
  [2026-01-25 18:00:41] (step=0003579) Train Loss mse: 0.0000, Train Loss ce: 0.2551, Train Steps/Sec: 0.36,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3803
  [2026-01-25 18:00:44] (step=0003580) Train Loss mse: 0.0000, Train Loss ce: 0.2514, Train Steps/Sec: 0.37,
3804
  [2026-01-25 18:00:46] (step=0003581) Train Loss mse: 0.0000, Train Loss ce: 0.2396, Train Steps/Sec: 0.39,
3805
  [2026-01-25 18:00:49] (step=0003582) Train Loss mse: 0.0000, Train Loss ce: 0.2210, Train Steps/Sec: 0.40,
 
3866
  [2026-01-25 18:03:40] (step=0003643) Train Loss mse: 0.0000, Train Loss ce: 0.2529, Train Steps/Sec: 0.37,
3867
  [2026-01-25 18:03:43] (step=0003644) Train Loss mse: 0.0000, Train Loss ce: 0.2446, Train Steps/Sec: 0.35,
3868
  [2026-01-25 18:03:46] (step=0003645) Train Loss mse: 0.0000, Train Loss ce: 0.2420, Train Steps/Sec: 0.35,
3869
+ base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step4000
3870
+ Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
3871
+ [eval debug] first 3 batch fingerprints:
3872
+ fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
3873
+ fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
3874
+ fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
3875
+ ce_avg: 0.7003546953201294, mse_avg: 0.0
3876
+ base_dir is /dev/shm/models/checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins/eval_used_rows, step_tag is checkpoints_vlm_gym_colorization_one_image_lr2e_5_ce_no_mse_ins_step4500
3877
+ Preparing Dataset vlm_gym_colorization_celoss_no_mse_evalonce/vlm_gym_colorization_val
3878
+ [eval debug] first 3 batch fingerprints:
3879
+ fp[0]: [{'data_indexes': [0], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
3880
+ fp[1]: [{'data_indexes': [8], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
3881
+ fp[2]: [{'data_indexes': [16], 'worker_id': 0, 'dataset_name': 'vlm_gym_colorization_celoss_no_mse_evalonce'}]
3882
+ ce_avg: 0.6831608414649963, mse_avg: 0.0
3883
  [2026-01-25 18:03:48] (step=0003646) Train Loss mse: 0.0000, Train Loss ce: 0.2638, Train Steps/Sec: 0.40,
3884
  [2026-01-25 18:03:51] (step=0003647) Train Loss mse: 0.0000, Train Loss ce: 0.2499, Train Steps/Sec: 0.38,
3885
  [2026-01-25 18:03:54] (step=0003648) Train Loss mse: 0.0000, Train Loss ce: 0.2567, Train Steps/Sec: 0.36,