Upload folder using huggingface_hub
Browse files- attnserver.run_attnserver.slurm.sh.343188.out.log +30 -0
- attnserver.run_attnserver.slurm.sh.343195.out.log +323 -0
- attnserver.run_attnserver.slurm.sh.343196.err.log +2 -2
- attnserver.run_attnserver.slurm.sh.343196.out.log +0 -0
- attnserver.run_attnserver.slurm.sh.343197.err.log +0 -0
- attnserver.run_attnserver.slurm.sh.343197.out.log +0 -0
- attnserver.run_attnserver.slurm.sh.343198.err.log +0 -0
- attnserver.run_attnserver.slurm.sh.343198.out.log +0 -0
- attnserver.run_attnserver.slurm.sh.343199.err.log +0 -0
- attnserver.run_attnserver.slurm.sh.343199.out.log +0 -0
- attnserver.run_attnserver.slurm.sh.343200.err.log +186 -0
- attnserver.run_attnserver.slurm.sh.343200.out.log +37 -0
- attnserver.run_attnserver.slurm.sh.343202.out.log +1162 -0
attnserver.run_attnserver.slurm.sh.343188.out.log
CHANGED
|
@@ -122183,3 +122183,33 @@ batch tensor after cp: labels torch.Size([1, 16384])
|
|
| 122183 |
batch tensor after cp: loss_mask torch.Size([1, 16384])
|
| 122184 |
batch tensor after cp: attention_mask torch.Size([1, 1, 16384, 131072])
|
| 122185 |
batch tensor after cp: position_ids torch.Size([1, 16384])
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 122183 |
batch tensor after cp: loss_mask torch.Size([1, 16384])
|
| 122184 |
batch tensor after cp: attention_mask torch.Size([1, 1, 16384, 131072])
|
| 122185 |
batch tensor after cp: position_ids torch.Size([1, 16384])
|
| 122186 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 122187 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 122188 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 122189 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 122190 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 122191 |
+
batch tensor after cp: tokens torch.Size([1, 16384])
|
| 122192 |
+
batch tensor after cp: labels torch.Size([1, 16384])
|
| 122193 |
+
batch tensor after cp: loss_mask torch.Size([1, 16384])
|
| 122194 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 16384, 131072])
|
| 122195 |
+
batch tensor after cp: position_ids torch.Size([1, 16384])
|
| 122196 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 122197 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 122198 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 122199 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 122200 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 122201 |
+
batch tensor after cp: tokens torch.Size([1, 16384])
|
| 122202 |
+
batch tensor after cp: labels torch.Size([1, 16384])
|
| 122203 |
+
batch tensor after cp: loss_mask torch.Size([1, 16384])
|
| 122204 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 16384, 131072])
|
| 122205 |
+
batch tensor after cp: position_ids torch.Size([1, 16384])
|
| 122206 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 122207 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 122208 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 122209 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 122210 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 122211 |
+
batch tensor after cp: tokens torch.Size([1, 16384])
|
| 122212 |
+
batch tensor after cp: labels torch.Size([1, 16384])
|
| 122213 |
+
batch tensor after cp: loss_mask torch.Size([1, 16384])
|
| 122214 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 16384, 131072])
|
| 122215 |
+
batch tensor after cp: position_ids torch.Size([1, 16384])
|
attnserver.run_attnserver.slurm.sh.343195.out.log
CHANGED
|
@@ -66428,3 +66428,326 @@ batch tensor after cp: labels torch.Size([1, 32768])
|
|
| 66428 |
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66429 |
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66430 |
batch tensor after cp: position_ids torch.Size([1, 32768])
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 66428 |
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66429 |
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66430 |
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66431 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66432 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66433 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66434 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66435 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66436 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66437 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66438 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66439 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66440 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66441 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66442 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66443 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66444 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66445 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66446 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66447 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66448 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66449 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66450 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66451 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66452 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66453 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66454 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66455 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66456 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66457 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66458 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66459 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66460 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66461 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66462 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66463 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66464 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66465 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66466 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66467 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66468 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66469 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66470 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66471 |
+
Start exporting trace 2
|
| 66472 |
+
Done exporting trace 2
|
| 66473 |
+
[2025-06-21 21:07:56] iteration 3/ 10 | consumed samples: 3 | elapsed time per iteration (ms): 174501.8 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 1073741824.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
| 66474 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66475 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66476 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66477 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66478 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66479 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66480 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66481 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66482 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66483 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66484 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66485 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66486 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66487 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66488 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66489 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66490 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66491 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66492 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66493 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66494 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66495 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66496 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66497 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66498 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66499 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66500 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66501 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66502 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66503 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66504 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66505 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66506 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66507 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66508 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66509 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66510 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66511 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66512 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66513 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66514 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66515 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66516 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66517 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66518 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66519 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66520 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66521 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66522 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66523 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66524 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66525 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66526 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66527 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66528 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66529 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66530 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66531 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66532 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66533 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66534 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66535 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66536 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66537 |
+
batch tensor:batch tensor: attention_mask tokens torch.Size([1, 1, 131072, 131072])
|
| 66538 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66539 |
+
torch.Size([1, 131072])
|
| 66540 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66541 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66542 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66543 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66544 |
+
batch tensor after cp: tokensbatch tensor after cp: tokens torch.Size([1, 32768])torch.Size([1, 32768])
|
| 66545 |
+
|
| 66546 |
+
batch tensor after cp:batch tensor after cp: labelslabels torch.Size([1, 32768])torch.Size([1, 32768])
|
| 66547 |
+
|
| 66548 |
+
batch tensor after cp:batch tensor after cp: loss_maskloss_mask torch.Size([1, 32768])torch.Size([1, 32768])
|
| 66549 |
+
|
| 66550 |
+
batch tensor after cp:batch tensor after cp: attention_maskattention_mask torch.Size([1, 1, 32768, 131072])
|
| 66551 |
+
torch.Size([1, 1, 32768, 131072])batch tensor after cp:
|
| 66552 |
+
position_idsbatch tensor after cp: position_idstorch.Size([1, 32768])
|
| 66553 |
+
torch.Size([1, 32768])
|
| 66554 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66555 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66556 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66557 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66558 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66559 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66560 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66561 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66562 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66563 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66564 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66565 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66566 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66567 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66568 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66569 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66570 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66571 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66572 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66573 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66574 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66575 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66576 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66577 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66578 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66579 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66580 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66581 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66582 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66583 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66584 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66585 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66586 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66587 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66588 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66589 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66590 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66591 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66592 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66593 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66594 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66595 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66596 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66597 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66598 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66599 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66600 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66601 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66602 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66603 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66604 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66605 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66606 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66607 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66608 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66609 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66610 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66611 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66612 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66613 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66614 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66615 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66616 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66617 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66618 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66619 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66620 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66621 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66622 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66623 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66624 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66625 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66626 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66627 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66628 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66629 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66630 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66631 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66632 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66633 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66634 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66635 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66636 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66637 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66638 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66639 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66640 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66641 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66642 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66643 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66644 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66645 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66646 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66647 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66648 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66649 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66650 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66651 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66652 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66653 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66654 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66655 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66656 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66657 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66658 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66659 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66660 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66661 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66662 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66663 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66664 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66665 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66666 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66667 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66668 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66669 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66670 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66671 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66672 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66673 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66674 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66675 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66676 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66677 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66678 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66679 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66680 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66681 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66682 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66683 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66684 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66685 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66686 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66687 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66688 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66689 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66690 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66691 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66692 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66693 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66694 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66695 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66696 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66697 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66698 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66699 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66700 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66701 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66702 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66703 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66704 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66705 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66706 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66707 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66708 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66709 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66710 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66711 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66712 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66713 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66714 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66715 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66716 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66717 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66718 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66719 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66720 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66721 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66722 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66723 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66724 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66725 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66726 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66727 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66728 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66729 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66730 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66731 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66732 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66733 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66734 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66735 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66736 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66737 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66738 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66739 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66740 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66741 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66742 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66743 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
| 66744 |
+
batch tensor: tokens torch.Size([1, 131072])
|
| 66745 |
+
batch tensor: labels torch.Size([1, 131072])
|
| 66746 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
| 66747 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
| 66748 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
| 66749 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
| 66750 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
| 66751 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
| 66752 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
| 66753 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
attnserver.run_attnserver.slurm.sh.343196.err.log
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c78b4349417cacd0f3c66f837e9bed0c1241276553a9acecd79541d563f8e013
|
| 3 |
+
size 30470641
|
attnserver.run_attnserver.slurm.sh.343196.out.log
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
attnserver.run_attnserver.slurm.sh.343197.err.log
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
attnserver.run_attnserver.slurm.sh.343197.out.log
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
attnserver.run_attnserver.slurm.sh.343198.err.log
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
attnserver.run_attnserver.slurm.sh.343198.out.log
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
attnserver.run_attnserver.slurm.sh.343199.err.log
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
attnserver.run_attnserver.slurm.sh.343199.out.log
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
attnserver.run_attnserver.slurm.sh.343200.err.log
ADDED
|
@@ -0,0 +1,186 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
+ source /mnt/weka/home/hao.zhang/conda/miniconda/bin/activate
|
| 2 |
+
++ _CONDA_ROOT=/mnt/weka/home/hao.zhang/conda/miniconda
|
| 3 |
+
++ . /mnt/weka/home/hao.zhang/conda/miniconda/etc/profile.d/conda.sh
|
| 4 |
+
+++ export CONDA_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda
|
| 5 |
+
+++ CONDA_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda
|
| 6 |
+
+++ export _CE_M=
|
| 7 |
+
+++ _CE_M=
|
| 8 |
+
+++ export _CE_CONDA=
|
| 9 |
+
+++ _CE_CONDA=
|
| 10 |
+
+++ export CONDA_PYTHON_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/python
|
| 11 |
+
+++ CONDA_PYTHON_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/python
|
| 12 |
+
+++ '[' -z x ']'
|
| 13 |
+
++ conda activate
|
| 14 |
+
++ local cmd=activate
|
| 15 |
+
++ case "$cmd" in
|
| 16 |
+
++ __conda_activate activate
|
| 17 |
+
++ '[' -n '' ']'
|
| 18 |
+
++ local ask_conda
|
| 19 |
+
+++ PS1=
|
| 20 |
+
+++ __conda_exe shell.posix activate
|
| 21 |
+
+++ '[' -n '' ']'
|
| 22 |
+
+++ /mnt/weka/home/hao.zhang/conda/miniconda/bin/conda shell.posix activate
|
| 23 |
+
++ ask_conda='unset _CE_M
|
| 24 |
+
unset _CE_CONDA
|
| 25 |
+
PS1='\''(base) '\''
|
| 26 |
+
export PATH='\''/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin'\''
|
| 27 |
+
export CONDA_SHLVL='\''1'\''
|
| 28 |
+
export CONDA_PROMPT_MODIFIER='\''(base) '\''
|
| 29 |
+
export CONDA_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda'\''
|
| 30 |
+
export CONDA_PYTHON_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/python'\'''
|
| 31 |
+
++ eval 'unset _CE_M
|
| 32 |
+
unset _CE_CONDA
|
| 33 |
+
PS1='\''(base) '\''
|
| 34 |
+
export PATH='\''/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin'\''
|
| 35 |
+
export CONDA_SHLVL='\''1'\''
|
| 36 |
+
export CONDA_PROMPT_MODIFIER='\''(base) '\''
|
| 37 |
+
export CONDA_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda'\''
|
| 38 |
+
export CONDA_PYTHON_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/python'\'''
|
| 39 |
+
+++ unset _CE_M
|
| 40 |
+
+++ unset _CE_CONDA
|
| 41 |
+
+++ PS1='(base) '
|
| 42 |
+
+++ export PATH=/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin
|
| 43 |
+
+++ PATH=/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin
|
| 44 |
+
+++ export CONDA_SHLVL=1
|
| 45 |
+
+++ CONDA_SHLVL=1
|
| 46 |
+
+++ export 'CONDA_PROMPT_MODIFIER=(base) '
|
| 47 |
+
+++ CONDA_PROMPT_MODIFIER='(base) '
|
| 48 |
+
+++ export CONDA_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda
|
| 49 |
+
+++ CONDA_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda
|
| 50 |
+
+++ export CONDA_PYTHON_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/python
|
| 51 |
+
+++ CONDA_PYTHON_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/python
|
| 52 |
+
++ __conda_hashr
|
| 53 |
+
++ '[' -n '' ']'
|
| 54 |
+
++ '[' -n '' ']'
|
| 55 |
+
++ hash -r
|
| 56 |
+
+ conda activate junda-attnserver
|
| 57 |
+
+ local cmd=activate
|
| 58 |
+
+ case "$cmd" in
|
| 59 |
+
+ __conda_activate activate junda-attnserver
|
| 60 |
+
+ '[' -n '' ']'
|
| 61 |
+
+ local ask_conda
|
| 62 |
+
++ PS1='(base) '
|
| 63 |
+
++ __conda_exe shell.posix activate junda-attnserver
|
| 64 |
+
++ '[' -n '' ']'
|
| 65 |
+
++ /mnt/weka/home/hao.zhang/conda/miniconda/bin/conda shell.posix activate junda-attnserver
|
| 66 |
+
+ ask_conda='unset _CE_M
|
| 67 |
+
unset _CE_CONDA
|
| 68 |
+
PS1='\''(junda-attnserver) '\''
|
| 69 |
+
export PATH='\''/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin'\''
|
| 70 |
+
export CONDA_PREFIX='\''/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver'\''
|
| 71 |
+
export CONDA_SHLVL='\''2'\''
|
| 72 |
+
export CONDA_DEFAULT_ENV='\''junda-attnserver'\''
|
| 73 |
+
export CONDA_PROMPT_MODIFIER='\''(junda-attnserver) '\''
|
| 74 |
+
export CONDA_PREFIX_1='\''/mnt/weka/home/hao.zhang/conda/miniconda'\''
|
| 75 |
+
export CONDA_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda'\''
|
| 76 |
+
export CONDA_PYTHON_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/python'\'''
|
| 77 |
+
+ eval 'unset _CE_M
|
| 78 |
+
unset _CE_CONDA
|
| 79 |
+
PS1='\''(junda-attnserver) '\''
|
| 80 |
+
export PATH='\''/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin'\''
|
| 81 |
+
export CONDA_PREFIX='\''/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver'\''
|
| 82 |
+
export CONDA_SHLVL='\''2'\''
|
| 83 |
+
export CONDA_DEFAULT_ENV='\''junda-attnserver'\''
|
| 84 |
+
export CONDA_PROMPT_MODIFIER='\''(junda-attnserver) '\''
|
| 85 |
+
export CONDA_PREFIX_1='\''/mnt/weka/home/hao.zhang/conda/miniconda'\''
|
| 86 |
+
export CONDA_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda'\''
|
| 87 |
+
export CONDA_PYTHON_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/python'\'''
|
| 88 |
+
++ unset _CE_M
|
| 89 |
+
++ unset _CE_CONDA
|
| 90 |
+
++ PS1='(junda-attnserver) '
|
| 91 |
+
++ export PATH=/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin
|
| 92 |
+
++ PATH=/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin
|
| 93 |
+
++ export CONDA_PREFIX=/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver
|
| 94 |
+
++ CONDA_PREFIX=/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver
|
| 95 |
+
++ export CONDA_SHLVL=2
|
| 96 |
+
++ CONDA_SHLVL=2
|
| 97 |
+
++ export CONDA_DEFAULT_ENV=junda-attnserver
|
| 98 |
+
++ CONDA_DEFAULT_ENV=junda-attnserver
|
| 99 |
+
++ export 'CONDA_PROMPT_MODIFIER=(junda-attnserver) '
|
| 100 |
+
++ CONDA_PROMPT_MODIFIER='(junda-attnserver) '
|
| 101 |
+
++ export CONDA_PREFIX_1=/mnt/weka/home/hao.zhang/conda/miniconda
|
| 102 |
+
++ CONDA_PREFIX_1=/mnt/weka/home/hao.zhang/conda/miniconda
|
| 103 |
+
++ export CONDA_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda
|
| 104 |
+
++ CONDA_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda
|
| 105 |
+
++ export CONDA_PYTHON_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/python
|
| 106 |
+
++ CONDA_PYTHON_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/python
|
| 107 |
+
+ __conda_hashr
|
| 108 |
+
+ '[' -n '' ']'
|
| 109 |
+
+ '[' -n '' ']'
|
| 110 |
+
+ hash -r
|
| 111 |
+
+ export CHROME_TRACE_PREFIX=/mnt/sharefs/users/hao.zhang/junda/megatron-prof-data--unstable-v5
|
| 112 |
+
+ CHROME_TRACE_PREFIX=/mnt/sharefs/users/hao.zhang/junda/megatron-prof-data--unstable-v5
|
| 113 |
+
+ mkdir -p /mnt/sharefs/users/hao.zhang/junda/megatron-prof-data--unstable-v5
|
| 114 |
+
+ export PROF_TP_SIZE=8
|
| 115 |
+
+ PROF_TP_SIZE=8
|
| 116 |
+
+ export PROF_CP_SIZE=4
|
| 117 |
+
+ PROF_CP_SIZE=4
|
| 118 |
+
+ export PROF_BS=32
|
| 119 |
+
+ PROF_BS=32
|
| 120 |
+
+ for ctx_length in 1024 2048 4096 8192 12288 16384 24576 32768 40960 49152 65536 81920 98304 131072
|
| 121 |
+
+ export PROF_CTX_LENGTH=1024
|
| 122 |
+
+ PROF_CTX_LENGTH=1024
|
| 123 |
+
+ name='/mnt/sharefs/users/hao.zhang/junda/megatron-prof-data--unstable-v5/mytrace.L1024*tp8.cp4.bs32.json'
|
| 124 |
+
+ '[' -f '/mnt/sharefs/users/hao.zhang/junda/megatron-prof-data--unstable-v5/mytrace.L1024*tp8.cp4.bs32.json' ']'
|
| 125 |
+
+ echo 'Running ctx_length=1024, TP_SIZE=8, CP_SIZE=4, BATCH_SIZE=32'
|
| 126 |
+
+ srun bash ./attnserver.sh
|
| 127 |
+
+ which python3
|
| 128 |
+
+ python3 -m torch.distributed.launch --nproc_per_node 8 --nnodes 4 --node_rank 3 --rdzv_id 343200 --rdzv_backend c10d --rdzv_endpoint fs-mbz-gpu-852:29500 ./pretrain_gpt_profile.py --tensor-model-parallel-size 8 --context-parallel-size 4 --num-layers 2 --hidden-size 4096 --num-attention-heads 64 --group-query-attention --num-query-groups 16 --seq-length 1024 --max-position-embeddings 1024 --micro-batch-size 1 --global-batch-size 1 --lr 0.0005 --train-iters 10 --lr-decay-iters 150000 --lr-decay-style cosine --lr-warmup-iters 2 --weight-decay .1 --adam-beta2 .999 --fp16 --log-interval 1 --save-interval 16 --eval-interval 16 --eval-iters 1 --vocab-file vocab.json --merge-file merges.txt --save gpt-checkpoint --load gpt-checkpoint --logging-level 0 --mock-data --tensorboard-dir tensorboard-logs/
|
| 129 |
+
+ which python3
|
| 130 |
+
+ python3 -m torch.distributed.launch --nproc_per_node 8 --nnodes 4 --node_rank 2 --rdzv_id 343200 --rdzv_backend c10d --rdzv_endpoint fs-mbz-gpu-852:29500 ./pretrain_gpt_profile.py --tensor-model-parallel-size 8 --context-parallel-size 4 --num-layers 2 --hidden-size 4096 --num-attention-heads 64 --group-query-attention --num-query-groups 16 --seq-length 1024 --max-position-embeddings 1024 --micro-batch-size 1 --global-batch-size 1 --lr 0.0005 --train-iters 10 --lr-decay-iters 150000 --lr-decay-style cosine --lr-warmup-iters 2 --weight-decay .1 --adam-beta2 .999 --fp16 --log-interval 1 --save-interval 16 --eval-interval 16 --eval-iters 1 --vocab-file vocab.json --merge-file merges.txt --save gpt-checkpoint --load gpt-checkpoint --logging-level 0 --mock-data --tensorboard-dir tensorboard-logs/
|
| 131 |
+
+ which python3
|
| 132 |
+
+ python3 -m torch.distributed.launch --nproc_per_node 8 --nnodes 4 --node_rank 0 --rdzv_id 343200 --rdzv_backend c10d --rdzv_endpoint fs-mbz-gpu-852:29500 ./pretrain_gpt_profile.py --tensor-model-parallel-size 8 --context-parallel-size 4 --num-layers 2 --hidden-size 4096 --num-attention-heads 64 --group-query-attention --num-query-groups 16 --seq-length 1024 --max-position-embeddings 1024 --micro-batch-size 1 --global-batch-size 1 --lr 0.0005 --train-iters 10 --lr-decay-iters 150000 --lr-decay-style cosine --lr-warmup-iters 2 --weight-decay .1 --adam-beta2 .999 --fp16 --log-interval 1 --save-interval 16 --eval-interval 16 --eval-iters 1 --vocab-file vocab.json --merge-file merges.txt --save gpt-checkpoint --load gpt-checkpoint --logging-level 0 --mock-data --tensorboard-dir tensorboard-logs/
|
| 133 |
+
+ which python3
|
| 134 |
+
+ python3 -m torch.distributed.launch --nproc_per_node 8 --nnodes 4 --node_rank 1 --rdzv_id 343200 --rdzv_backend c10d --rdzv_endpoint fs-mbz-gpu-852:29500 ./pretrain_gpt_profile.py --tensor-model-parallel-size 8 --context-parallel-size 4 --num-layers 2 --hidden-size 4096 --num-attention-heads 64 --group-query-attention --num-query-groups 16 --seq-length 1024 --max-position-embeddings 1024 --micro-batch-size 1 --global-batch-size 1 --lr 0.0005 --train-iters 10 --lr-decay-iters 150000 --lr-decay-style cosine --lr-warmup-iters 2 --weight-decay .1 --adam-beta2 .999 --fp16 --log-interval 1 --save-interval 16 --eval-interval 16 --eval-iters 1 --vocab-file vocab.json --merge-file merges.txt --save gpt-checkpoint --load gpt-checkpoint --logging-level 0 --mock-data --tensorboard-dir tensorboard-logs/
|
| 135 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/lib/python3.12/site-packages/torch/distributed/launch.py:207: FutureWarning: The module torch.distributed.launch is deprecated
|
| 136 |
+
and will be removed in future. Use torchrun.
|
| 137 |
+
Note that --use-env is set by default in torchrun.
|
| 138 |
+
If your script expects `--local-rank` argument to be set, please
|
| 139 |
+
change it to read from `os.environ['LOCAL_RANK']` instead. See
|
| 140 |
+
https://pytorch.org/docs/stable/distributed.html#launch-utility for
|
| 141 |
+
further instructions
|
| 142 |
+
|
| 143 |
+
main()
|
| 144 |
+
W0621 21:08:49.911000 121562 site-packages/torch/distributed/run.py:766]
|
| 145 |
+
W0621 21:08:49.911000 121562 site-packages/torch/distributed/run.py:766] *****************************************
|
| 146 |
+
W0621 21:08:49.911000 121562 site-packages/torch/distributed/run.py:766] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
|
| 147 |
+
W0621 21:08:49.911000 121562 site-packages/torch/distributed/run.py:766] *****************************************
|
| 148 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/lib/python3.12/site-packages/torch/distributed/launch.py:207: FutureWarning: The module torch.distributed.launch is deprecated
|
| 149 |
+
and will be removed in future. Use torchrun.
|
| 150 |
+
Note that --use-env is set by default in torchrun.
|
| 151 |
+
If your script expects `--local-rank` argument to be set, please
|
| 152 |
+
change it to read from `os.environ['LOCAL_RANK']` instead. See
|
| 153 |
+
https://pytorch.org/docs/stable/distributed.html#launch-utility for
|
| 154 |
+
further instructions
|
| 155 |
+
|
| 156 |
+
main()
|
| 157 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/lib/python3.12/site-packages/torch/distributed/launch.py:207: FutureWarning: The module torch.distributed.launch is deprecated
|
| 158 |
+
and will be removed in future. Use torchrun.
|
| 159 |
+
Note that --use-env is set by default in torchrun.
|
| 160 |
+
If your script expects `--local-rank` argument to be set, please
|
| 161 |
+
change it to read from `os.environ['LOCAL_RANK']` instead. See
|
| 162 |
+
https://pytorch.org/docs/stable/distributed.html#launch-utility for
|
| 163 |
+
further instructions
|
| 164 |
+
|
| 165 |
+
main()
|
| 166 |
+
W0621 21:08:49.997000 2041342 site-packages/torch/distributed/run.py:766]
|
| 167 |
+
W0621 21:08:49.997000 2041342 site-packages/torch/distributed/run.py:766] *****************************************
|
| 168 |
+
W0621 21:08:49.997000 2041342 site-packages/torch/distributed/run.py:766] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
|
| 169 |
+
W0621 21:08:49.997000 2041342 site-packages/torch/distributed/run.py:766] *****************************************
|
| 170 |
+
W0621 21:08:49.997000 3410371 site-packages/torch/distributed/run.py:766]
|
| 171 |
+
W0621 21:08:49.997000 3410371 site-packages/torch/distributed/run.py:766] *****************************************
|
| 172 |
+
W0621 21:08:49.997000 3410371 site-packages/torch/distributed/run.py:766] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
|
| 173 |
+
W0621 21:08:49.997000 3410371 site-packages/torch/distributed/run.py:766] *****************************************
|
| 174 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/lib/python3.12/site-packages/torch/distributed/launch.py:207: FutureWarning: The module torch.distributed.launch is deprecated
|
| 175 |
+
and will be removed in future. Use torchrun.
|
| 176 |
+
Note that --use-env is set by default in torchrun.
|
| 177 |
+
If your script expects `--local-rank` argument to be set, please
|
| 178 |
+
change it to read from `os.environ['LOCAL_RANK']` instead. See
|
| 179 |
+
https://pytorch.org/docs/stable/distributed.html#launch-utility for
|
| 180 |
+
further instructions
|
| 181 |
+
|
| 182 |
+
main()
|
| 183 |
+
W0621 21:08:50.019000 3341756 site-packages/torch/distributed/run.py:766]
|
| 184 |
+
W0621 21:08:50.019000 3341756 site-packages/torch/distributed/run.py:766] *****************************************
|
| 185 |
+
W0621 21:08:50.019000 3341756 site-packages/torch/distributed/run.py:766] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
|
| 186 |
+
W0621 21:08:50.019000 3341756 site-packages/torch/distributed/run.py:766] *****************************************
|
attnserver.run_attnserver.slurm.sh.343200.out.log
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
Running ctx_length=1024, TP_SIZE=8, CP_SIZE=4, BATCH_SIZE=32
|
| 2 |
+
Cleaning up checkpoint directory: gpt-checkpoint
|
| 3 |
+
--------------------------------
|
| 4 |
+
CTX_LENGTH: 1024
|
| 5 |
+
TP_SIZE: 8
|
| 6 |
+
CP_SIZE: 4
|
| 7 |
+
CHECKPOINT_PATH: gpt-checkpoint
|
| 8 |
+
PWD: /mnt/weka/home/hao.zhang/junda/attnserver-megatron
|
| 9 |
+
--------------------------------
|
| 10 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin/python3
|
| 11 |
+
Cleaning up checkpoint directory: gpt-checkpoint
|
| 12 |
+
--------------------------------
|
| 13 |
+
CTX_LENGTH: 1024
|
| 14 |
+
TP_SIZE: 8
|
| 15 |
+
CP_SIZE: 4
|
| 16 |
+
CHECKPOINT_PATH: gpt-checkpoint
|
| 17 |
+
PWD: /mnt/weka/home/hao.zhang/junda/attnserver-megatron
|
| 18 |
+
--------------------------------
|
| 19 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin/python3
|
| 20 |
+
Cleaning up checkpoint directory: gpt-checkpoint
|
| 21 |
+
--------------------------------
|
| 22 |
+
CTX_LENGTH: 1024
|
| 23 |
+
TP_SIZE: 8
|
| 24 |
+
CP_SIZE: 4
|
| 25 |
+
CHECKPOINT_PATH: gpt-checkpoint
|
| 26 |
+
PWD: /mnt/weka/home/hao.zhang/junda/attnserver-megatron
|
| 27 |
+
--------------------------------
|
| 28 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin/python3
|
| 29 |
+
Cleaning up checkpoint directory: gpt-checkpoint
|
| 30 |
+
--------------------------------
|
| 31 |
+
CTX_LENGTH: 1024
|
| 32 |
+
TP_SIZE: 8
|
| 33 |
+
CP_SIZE: 4
|
| 34 |
+
CHECKPOINT_PATH: gpt-checkpoint
|
| 35 |
+
PWD: /mnt/weka/home/hao.zhang/junda/attnserver-megatron
|
| 36 |
+
--------------------------------
|
| 37 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin/python3
|
attnserver.run_attnserver.slurm.sh.343202.out.log
CHANGED
|
@@ -22196,3 +22196,1165 @@ batch tensor after cp: labels torch.Size([2, 40960])
|
|
| 22196 |
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22197 |
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22198 |
batch tensor after cp: position_ids torch.Size([2, 40960])
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 22196 |
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22197 |
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22198 |
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22199 |
+
Start exporting trace 0
|
| 22200 |
+
Done exporting trace 0
|
| 22201 |
+
Number of parameters in transformer block in billions: 0.35
|
| 22202 |
+
Number of parameters in embedding layers in billions: 0.21
|
| 22203 |
+
Total number of parameters in billions: 0.56
|
| 22204 |
+
Number of parameters in most loaded shard in billions: 0.0703
|
| 22205 |
+
Theoretical memory footprints: weight and optimizer=1206.09 MB
|
| 22206 |
+
[2025-06-21 21:07:57] iteration 1/ 10 | consumed samples: 1 | elapsed time per iteration (ms): 29934.7 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 4294967296.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
| 22207 |
+
[Rank 5] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36108.0 | max reserved: 36108.0
|
| 22208 |
+
[Rank 1] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36108.0 | max reserved: 36108.0
|
| 22209 |
+
[Rank 7] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36108.0 | max reserved: 36108.0
|
| 22210 |
+
[Rank 14] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36148.0 | max reserved: 36148.0
|
| 22211 |
+
[Rank 12] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36148.0 | max reserved: 36148.0
|
| 22212 |
+
[Rank 11] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36148.0 | max reserved: 36148.0
|
| 22213 |
+
[Rank 10] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36148.0 | max reserved: 36148.0
|
| 22214 |
+
[Rank 3] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36108.0 | max reserved: 36108.0
|
| 22215 |
+
[Rank 15] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36148.0 | max reserved: 36148.0
|
| 22216 |
+
[Rank 8] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 35508.0 | max reserved: 35508.0
|
| 22217 |
+
[Rank 2] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36108.0 | max reserved: 36108.0
|
| 22218 |
+
[Rank 9] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36148.0 | max reserved: 36148.0
|
| 22219 |
+
[Rank 0] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 35588.0 | max reserved: 35588.0
|
| 22220 |
+
[Rank 13] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36148.0 | max reserved: 36148.0
|
| 22221 |
+
[Rank 6] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36108.0 | max reserved: 36108.0
|
| 22222 |
+
[Rank 4] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36108.0 | max reserved: 36108.0
|
| 22223 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22224 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22225 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22226 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22227 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22228 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22229 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22230 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22231 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22232 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22233 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22234 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22235 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22236 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22237 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22238 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22239 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22240 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22241 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22242 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22243 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22244 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22245 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22246 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22247 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22248 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22249 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22250 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22251 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22252 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22253 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22254 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22255 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22256 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22257 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22258 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22259 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22260 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22261 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22262 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22263 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22264 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22265 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22266 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22267 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22268 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22269 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22270 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22271 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22272 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22273 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22274 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22275 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22276 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22277 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22278 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22279 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22280 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22281 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22282 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22283 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22284 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22285 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22286 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22287 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22288 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22289 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22290 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22291 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22292 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22293 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22294 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22295 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22296 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22297 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22298 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22299 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22300 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22301 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22302 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22303 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22304 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22305 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22306 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22307 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22308 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22309 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22310 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22311 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22312 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22313 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22314 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22315 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22316 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22317 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22318 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22319 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22320 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22321 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22322 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22323 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22324 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22325 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22326 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22327 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22328 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22329 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22330 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22331 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22332 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22333 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22334 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22335 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22336 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22337 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22338 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22339 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22340 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22341 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22342 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22343 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22344 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22345 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22346 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22347 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22348 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22349 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22350 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22351 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22352 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22353 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22354 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22355 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22356 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22357 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22358 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22359 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22360 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22361 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22362 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22363 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22364 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22365 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22366 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22367 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22368 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22369 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22370 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22371 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22372 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22373 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22374 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22375 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22376 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22377 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22378 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22379 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22380 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22381 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22382 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22383 |
+
Start exporting trace 1
|
| 22384 |
+
Done exporting trace 1
|
| 22385 |
+
[2025-06-21 21:08:05] iteration 2/ 10 | consumed samples: 2 | elapsed time per iteration (ms): 8086.1 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 2147483648.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
| 22386 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22387 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22388 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22389 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22390 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22391 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22392 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22393 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22394 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22395 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22396 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22397 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22398 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22399 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22400 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22401 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22402 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22403 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22404 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22405 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22406 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22407 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22408 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22409 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22410 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22411 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22412 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22413 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22414 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22415 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22416 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22417 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22418 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22419 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22420 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22421 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22422 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22423 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22424 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22425 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22426 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22427 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22428 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22429 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22430 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22431 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22432 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22433 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22434 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22435 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22436 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22437 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22438 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22439 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22440 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22441 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22442 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22443 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22444 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22445 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22446 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22447 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22448 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22449 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22450 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22451 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22452 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22453 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22454 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22455 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22456 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22457 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22458 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22459 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22460 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22461 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22462 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22463 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22464 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22465 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22466 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22467 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22468 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22469 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22470 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22471 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22472 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22473 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22474 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22475 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22476 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22477 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22478 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22479 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22480 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22481 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22482 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22483 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22484 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22485 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22486 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22487 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22488 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22489 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22490 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22491 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22492 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22493 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22494 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22495 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22496 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22497 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22498 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22499 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22500 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22501 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22502 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22503 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22504 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22505 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22506 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22507 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22508 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22509 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22510 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22511 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22512 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22513 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22514 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22515 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22516 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22517 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22518 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22519 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22520 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22521 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22522 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22523 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22524 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22525 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22526 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22527 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22528 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22529 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22530 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22531 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22532 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22533 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22534 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22535 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22536 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22537 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22538 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22539 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22540 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22541 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22542 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22543 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22544 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22545 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22546 |
+
Start exporting trace 2
|
| 22547 |
+
Done exporting trace 2
|
| 22548 |
+
[2025-06-21 21:08:13] iteration 3/ 10 | consumed samples: 3 | elapsed time per iteration (ms): 7945.0 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 1073741824.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
| 22549 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22550 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22551 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22552 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22553 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22554 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22555 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22556 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22557 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22558 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22559 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22560 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22561 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22562 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22563 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22564 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22565 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22566 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22567 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22568 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22569 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22570 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22571 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22572 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22573 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22574 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22575 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22576 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22577 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22578 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22579 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22580 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22581 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22582 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22583 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22584 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22585 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22586 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22587 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22588 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22589 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22590 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22591 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22592 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22593 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22594 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22595 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22596 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22597 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22598 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22599 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22600 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22601 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22602 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22603 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22604 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22605 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22606 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22607 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22608 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22609 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22610 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22611 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22612 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22613 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22614 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22615 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22616 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22617 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22618 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22619 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22620 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22621 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22622 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22623 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22624 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22625 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22626 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22627 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22628 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22629 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22630 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22631 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22632 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22633 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22634 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22635 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22636 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22637 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22638 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22639 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22640 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22641 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22642 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22643 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22644 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22645 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22646 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22647 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22648 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22649 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22650 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22651 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22652 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22653 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22654 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22655 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22656 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22657 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22658 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22659 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22660 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22661 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22662 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22663 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22664 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22665 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22666 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22667 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22668 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22669 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22670 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22671 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22672 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22673 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22674 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22675 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22676 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22677 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22678 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22679 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22680 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22681 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22682 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22683 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22684 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22685 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22686 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22687 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22688 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22689 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22690 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22691 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22692 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22693 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22694 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22695 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22696 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22697 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22698 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22699 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22700 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22701 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22702 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22703 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22704 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22705 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22706 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22707 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22708 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22709 |
+
Start exporting trace 3
|
| 22710 |
+
Done exporting trace 3
|
| 22711 |
+
[2025-06-21 21:08:21] iteration 4/ 10 | consumed samples: 4 | elapsed time per iteration (ms): 7924.7 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 536870912.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
| 22712 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22713 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22714 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22715 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22716 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22717 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22718 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22719 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22720 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22721 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22722 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22723 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22724 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22725 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22726 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22727 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22728 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22729 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22730 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22731 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22732 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22733 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22734 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22735 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22736 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22737 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22738 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22739 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22740 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22741 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22742 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22743 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22744 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22745 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22746 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22747 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22748 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22749 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22750 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22751 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22752 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22753 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22754 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22755 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22756 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22757 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22758 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22759 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22760 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22761 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22762 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22763 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22764 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22765 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22766 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22767 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22768 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22769 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22770 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22771 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22772 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22773 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22774 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22775 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22776 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22777 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22778 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22779 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22780 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22781 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22782 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22783 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22784 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22785 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22786 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22787 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22788 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22789 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22790 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22791 |
+
batch tensor after cp: position_idsbatch tensor: torch.Size([2, 40960])
|
| 22792 |
+
tokens torch.Size([2, 81920])
|
| 22793 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22794 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22795 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22796 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22797 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22798 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22799 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22800 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22801 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22802 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22803 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22804 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22805 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22806 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22807 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22808 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22809 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22810 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22811 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22812 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22813 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22814 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22815 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22816 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22817 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22818 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22819 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22820 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22821 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22822 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22823 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22824 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22825 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22826 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22827 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22828 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22829 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22830 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22831 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22832 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22833 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22834 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22835 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22836 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22837 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22838 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22839 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22840 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22841 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22842 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22843 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22844 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22845 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22846 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22847 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22848 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22849 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22850 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22851 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22852 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22853 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22854 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22855 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22856 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22857 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22858 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22859 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22860 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22861 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22862 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22863 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22864 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22865 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22866 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22867 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22868 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22869 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22870 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22871 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22872 |
+
Start exporting trace 4
|
| 22873 |
+
Done exporting trace 4
|
| 22874 |
+
[2025-06-21 21:08:28] iteration 5/ 10 | consumed samples: 5 | elapsed time per iteration (ms): 7646.0 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 268435456.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
| 22875 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22876 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22877 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22878 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22879 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22880 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22881 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22882 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22883 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22884 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22885 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22886 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22887 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22888 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22889 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22890 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22891 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22892 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22893 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22894 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22895 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22896 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22897 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22898 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22899 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22900 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22901 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22902 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22903 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22904 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22905 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22906 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22907 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22908 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22909 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22910 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22911 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22912 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22913 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22914 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22915 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22916 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22917 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22918 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22919 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22920 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22921 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22922 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22923 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22924 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22925 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22926 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22927 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22928 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22929 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22930 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22931 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22932 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22933 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22934 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22935 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22936 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22937 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22938 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22939 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22940 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22941 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22942 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22943 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22944 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22945 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22946 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22947 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22948 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22949 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22950 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22951 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22952 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22953 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22954 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22955 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22956 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22957 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22958 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22959 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22960 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22961 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22962 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22963 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22964 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22965 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22966 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22967 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22968 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22969 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22970 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22971 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22972 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22973 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22974 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22975 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22976 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22977 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22978 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22979 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22980 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22981 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22982 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22983 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22984 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22985 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22986 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22987 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22988 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22989 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 22990 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 22991 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 22992 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 22993 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 22994 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 22995 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 22996 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 22997 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 22998 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 22999 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23000 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23001 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23002 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23003 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23004 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23005 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23006 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23007 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23008 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23009 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23010 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23011 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23012 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23013 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23014 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23015 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23016 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23017 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23018 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23019 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23020 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23021 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23022 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23023 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23024 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23025 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23026 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23027 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23028 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23029 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23030 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23031 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23032 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23033 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23034 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23035 |
+
Start exporting trace 5
|
| 23036 |
+
Done exporting trace 5
|
| 23037 |
+
[2025-06-21 21:08:36] iteration 6/ 10 | consumed samples: 6 | elapsed time per iteration (ms): 7989.7 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 134217728.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
| 23038 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23039 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23040 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23041 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23042 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23043 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23044 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23045 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23046 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23047 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23048 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23049 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23050 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23051 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23052 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23053 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23054 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23055 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23056 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23057 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23058 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23059 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23060 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23061 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23062 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23063 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23064 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23065 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23066 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23067 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23068 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23069 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23070 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23071 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23072 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23073 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23074 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23075 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23076 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23077 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23078 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23079 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23080 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23081 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23082 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23083 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23084 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23085 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23086 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23087 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23088 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23089 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23090 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23091 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23092 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23093 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23094 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23095 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23096 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23097 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23098 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23099 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23100 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23101 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23102 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23103 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23104 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23105 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23106 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23107 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23108 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23109 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23110 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23111 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23112 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23113 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23114 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23115 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23116 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23117 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23118 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23119 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23120 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23121 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23122 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23123 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23124 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23125 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23126 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23127 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23128 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23129 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23130 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23131 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23132 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23133 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23134 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23135 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23136 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23137 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23138 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23139 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23140 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23141 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23142 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23143 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23144 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23145 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23146 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23147 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23148 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23149 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23150 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23151 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23152 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23153 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23154 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23155 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23156 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23157 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23158 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23159 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23160 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23161 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23162 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23163 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23164 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23165 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23166 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23167 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23168 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23169 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23170 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23171 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23172 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23173 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23174 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23175 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23176 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23177 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23178 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23179 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23180 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23181 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23182 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23183 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23184 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23185 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23186 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23187 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23188 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23189 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23190 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23191 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23192 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23193 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23194 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23195 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23196 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23197 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23198 |
+
Start exporting trace 6
|
| 23199 |
+
Done exporting trace 6
|
| 23200 |
+
[2025-06-21 21:08:44] iteration 7/ 10 | consumed samples: 7 | elapsed time per iteration (ms): 7801.4 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 67108864.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
| 23201 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23202 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23203 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23204 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23205 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23206 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23207 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23208 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23209 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23210 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23211 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23212 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23213 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23214 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23215 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23216 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23217 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23218 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23219 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23220 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23221 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23222 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23223 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23224 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23225 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23226 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23227 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23228 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23229 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23230 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23231 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23232 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23233 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23234 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23235 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23236 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23237 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23238 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23239 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23240 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23241 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23242 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23243 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23244 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23245 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23246 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23247 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23248 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23249 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23250 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23251 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23252 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23253 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23254 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23255 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23256 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23257 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23258 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23259 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23260 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23261 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23262 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23263 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23264 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23265 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23266 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23267 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23268 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23269 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23270 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23271 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23272 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23273 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23274 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23275 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23276 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23277 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23278 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23279 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23280 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23281 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23282 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23283 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23284 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23285 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23286 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23287 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23288 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23289 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23290 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23291 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23292 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23293 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23294 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23295 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23296 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23297 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23298 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23299 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23300 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23301 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23302 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23303 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23304 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23305 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23306 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23307 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23308 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23309 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23310 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23311 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23312 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23313 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23314 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23315 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23316 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23317 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23318 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23319 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23320 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23321 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23322 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23323 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23324 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23325 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23326 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23327 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23328 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23329 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23330 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23331 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23332 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23333 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23334 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23335 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23336 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23337 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23338 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23339 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23340 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23341 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23342 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23343 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23344 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23345 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23346 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23347 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23348 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23349 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23350 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
| 23351 |
+
batch tensor: tokens torch.Size([2, 81920])
|
| 23352 |
+
batch tensor: labels torch.Size([2, 81920])
|
| 23353 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
| 23354 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
| 23355 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
| 23356 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
| 23357 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
| 23358 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
| 23359 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
| 23360 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|