Upload folder using huggingface_hub
Browse files- attnserver.run_attnserver.slurm.sh.343188.out.log +30 -0
- attnserver.run_attnserver.slurm.sh.343195.out.log +323 -0
- attnserver.run_attnserver.slurm.sh.343196.err.log +2 -2
- attnserver.run_attnserver.slurm.sh.343196.out.log +0 -0
- attnserver.run_attnserver.slurm.sh.343197.err.log +0 -0
- attnserver.run_attnserver.slurm.sh.343197.out.log +0 -0
- attnserver.run_attnserver.slurm.sh.343198.err.log +0 -0
- attnserver.run_attnserver.slurm.sh.343198.out.log +0 -0
- attnserver.run_attnserver.slurm.sh.343199.err.log +0 -0
- attnserver.run_attnserver.slurm.sh.343199.out.log +0 -0
- attnserver.run_attnserver.slurm.sh.343200.err.log +186 -0
- attnserver.run_attnserver.slurm.sh.343200.out.log +37 -0
- attnserver.run_attnserver.slurm.sh.343202.out.log +1162 -0
attnserver.run_attnserver.slurm.sh.343188.out.log
CHANGED
@@ -122183,3 +122183,33 @@ batch tensor after cp: labels torch.Size([1, 16384])
|
|
122183 |
batch tensor after cp: loss_mask torch.Size([1, 16384])
|
122184 |
batch tensor after cp: attention_mask torch.Size([1, 1, 16384, 131072])
|
122185 |
batch tensor after cp: position_ids torch.Size([1, 16384])
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
122183 |
batch tensor after cp: loss_mask torch.Size([1, 16384])
|
122184 |
batch tensor after cp: attention_mask torch.Size([1, 1, 16384, 131072])
|
122185 |
batch tensor after cp: position_ids torch.Size([1, 16384])
|
122186 |
+
batch tensor: tokens torch.Size([1, 131072])
|
122187 |
+
batch tensor: labels torch.Size([1, 131072])
|
122188 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
122189 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
122190 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
122191 |
+
batch tensor after cp: tokens torch.Size([1, 16384])
|
122192 |
+
batch tensor after cp: labels torch.Size([1, 16384])
|
122193 |
+
batch tensor after cp: loss_mask torch.Size([1, 16384])
|
122194 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 16384, 131072])
|
122195 |
+
batch tensor after cp: position_ids torch.Size([1, 16384])
|
122196 |
+
batch tensor: tokens torch.Size([1, 131072])
|
122197 |
+
batch tensor: labels torch.Size([1, 131072])
|
122198 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
122199 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
122200 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
122201 |
+
batch tensor after cp: tokens torch.Size([1, 16384])
|
122202 |
+
batch tensor after cp: labels torch.Size([1, 16384])
|
122203 |
+
batch tensor after cp: loss_mask torch.Size([1, 16384])
|
122204 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 16384, 131072])
|
122205 |
+
batch tensor after cp: position_ids torch.Size([1, 16384])
|
122206 |
+
batch tensor: tokens torch.Size([1, 131072])
|
122207 |
+
batch tensor: labels torch.Size([1, 131072])
|
122208 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
122209 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
122210 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
122211 |
+
batch tensor after cp: tokens torch.Size([1, 16384])
|
122212 |
+
batch tensor after cp: labels torch.Size([1, 16384])
|
122213 |
+
batch tensor after cp: loss_mask torch.Size([1, 16384])
|
122214 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 16384, 131072])
|
122215 |
+
batch tensor after cp: position_ids torch.Size([1, 16384])
|
attnserver.run_attnserver.slurm.sh.343195.out.log
CHANGED
@@ -66428,3 +66428,326 @@ batch tensor after cp: labels torch.Size([1, 32768])
|
|
66428 |
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66429 |
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66430 |
batch tensor after cp: position_ids torch.Size([1, 32768])
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
66428 |
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66429 |
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66430 |
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66431 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66432 |
+
batch tensor: labels torch.Size([1, 131072])
|
66433 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66434 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66435 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66436 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66437 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66438 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66439 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66440 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66441 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66442 |
+
batch tensor: labels torch.Size([1, 131072])
|
66443 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66444 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66445 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66446 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66447 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66448 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66449 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66450 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66451 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66452 |
+
batch tensor: labels torch.Size([1, 131072])
|
66453 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66454 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66455 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66456 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66457 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66458 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66459 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66460 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66461 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66462 |
+
batch tensor: labels torch.Size([1, 131072])
|
66463 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66464 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66465 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66466 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66467 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66468 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66469 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66470 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66471 |
+
Start exporting trace 2
|
66472 |
+
Done exporting trace 2
|
66473 |
+
[2025-06-21 21:07:56] iteration 3/ 10 | consumed samples: 3 | elapsed time per iteration (ms): 174501.8 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 1073741824.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
66474 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66475 |
+
batch tensor: labels torch.Size([1, 131072])
|
66476 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66477 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66478 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66479 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66480 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66481 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66482 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66483 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66484 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66485 |
+
batch tensor: labels torch.Size([1, 131072])
|
66486 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66487 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66488 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66489 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66490 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66491 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66492 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66493 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66494 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66495 |
+
batch tensor: labels torch.Size([1, 131072])
|
66496 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66497 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66498 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66499 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66500 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66501 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66502 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66503 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66504 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66505 |
+
batch tensor: labels torch.Size([1, 131072])
|
66506 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66507 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66508 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66509 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66510 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66511 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66512 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66513 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66514 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66515 |
+
batch tensor: labels torch.Size([1, 131072])
|
66516 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66517 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66518 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66519 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66520 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66521 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66522 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66523 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66524 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66525 |
+
batch tensor: labels torch.Size([1, 131072])
|
66526 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66527 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66528 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66529 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66530 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66531 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66532 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66533 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66534 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66535 |
+
batch tensor: labels torch.Size([1, 131072])
|
66536 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66537 |
+
batch tensor:batch tensor: attention_mask tokens torch.Size([1, 1, 131072, 131072])
|
66538 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66539 |
+
torch.Size([1, 131072])
|
66540 |
+
batch tensor: labels torch.Size([1, 131072])
|
66541 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66542 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66543 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66544 |
+
batch tensor after cp: tokensbatch tensor after cp: tokens torch.Size([1, 32768])torch.Size([1, 32768])
|
66545 |
+
|
66546 |
+
batch tensor after cp:batch tensor after cp: labelslabels torch.Size([1, 32768])torch.Size([1, 32768])
|
66547 |
+
|
66548 |
+
batch tensor after cp:batch tensor after cp: loss_maskloss_mask torch.Size([1, 32768])torch.Size([1, 32768])
|
66549 |
+
|
66550 |
+
batch tensor after cp:batch tensor after cp: attention_maskattention_mask torch.Size([1, 1, 32768, 131072])
|
66551 |
+
torch.Size([1, 1, 32768, 131072])batch tensor after cp:
|
66552 |
+
position_idsbatch tensor after cp: position_idstorch.Size([1, 32768])
|
66553 |
+
torch.Size([1, 32768])
|
66554 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66555 |
+
batch tensor: labels torch.Size([1, 131072])
|
66556 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66557 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66558 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66559 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66560 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66561 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66562 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66563 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66564 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66565 |
+
batch tensor: labels torch.Size([1, 131072])
|
66566 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66567 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66568 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66569 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66570 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66571 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66572 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66573 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66574 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66575 |
+
batch tensor: labels torch.Size([1, 131072])
|
66576 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66577 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66578 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66579 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66580 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66581 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66582 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66583 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66584 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66585 |
+
batch tensor: labels torch.Size([1, 131072])
|
66586 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66587 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66588 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66589 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66590 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66591 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66592 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66593 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66594 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66595 |
+
batch tensor: labels torch.Size([1, 131072])
|
66596 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66597 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66598 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66599 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66600 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66601 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66602 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66603 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66604 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66605 |
+
batch tensor: labels torch.Size([1, 131072])
|
66606 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66607 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66608 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66609 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66610 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66611 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66612 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66613 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66614 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66615 |
+
batch tensor: labels torch.Size([1, 131072])
|
66616 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66617 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66618 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66619 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66620 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66621 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66622 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66623 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66624 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66625 |
+
batch tensor: labels torch.Size([1, 131072])
|
66626 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66627 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66628 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66629 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66630 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66631 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66632 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66633 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66634 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66635 |
+
batch tensor: labels torch.Size([1, 131072])
|
66636 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66637 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66638 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66639 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66640 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66641 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66642 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66643 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66644 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66645 |
+
batch tensor: labels torch.Size([1, 131072])
|
66646 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66647 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66648 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66649 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66650 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66651 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66652 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66653 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66654 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66655 |
+
batch tensor: labels torch.Size([1, 131072])
|
66656 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66657 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66658 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66659 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66660 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66661 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66662 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66663 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66664 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66665 |
+
batch tensor: labels torch.Size([1, 131072])
|
66666 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66667 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66668 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66669 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66670 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66671 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66672 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66673 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66674 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66675 |
+
batch tensor: labels torch.Size([1, 131072])
|
66676 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66677 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66678 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66679 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66680 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66681 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66682 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66683 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66684 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66685 |
+
batch tensor: labels torch.Size([1, 131072])
|
66686 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66687 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66688 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66689 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66690 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66691 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66692 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66693 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66694 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66695 |
+
batch tensor: labels torch.Size([1, 131072])
|
66696 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66697 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66698 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66699 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66700 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66701 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66702 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66703 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66704 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66705 |
+
batch tensor: labels torch.Size([1, 131072])
|
66706 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66707 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66708 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66709 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66710 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66711 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66712 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66713 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66714 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66715 |
+
batch tensor: labels torch.Size([1, 131072])
|
66716 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66717 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66718 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66719 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66720 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66721 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66722 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66723 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66724 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66725 |
+
batch tensor: labels torch.Size([1, 131072])
|
66726 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66727 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66728 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66729 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66730 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66731 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66732 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66733 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66734 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66735 |
+
batch tensor: labels torch.Size([1, 131072])
|
66736 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66737 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66738 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66739 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66740 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66741 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66742 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66743 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
66744 |
+
batch tensor: tokens torch.Size([1, 131072])
|
66745 |
+
batch tensor: labels torch.Size([1, 131072])
|
66746 |
+
batch tensor: loss_mask torch.Size([1, 131072])
|
66747 |
+
batch tensor: attention_mask torch.Size([1, 1, 131072, 131072])
|
66748 |
+
batch tensor: position_ids torch.Size([1, 131072])
|
66749 |
+
batch tensor after cp: tokens torch.Size([1, 32768])
|
66750 |
+
batch tensor after cp: labels torch.Size([1, 32768])
|
66751 |
+
batch tensor after cp: loss_mask torch.Size([1, 32768])
|
66752 |
+
batch tensor after cp: attention_mask torch.Size([1, 1, 32768, 131072])
|
66753 |
+
batch tensor after cp: position_ids torch.Size([1, 32768])
|
attnserver.run_attnserver.slurm.sh.343196.err.log
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c78b4349417cacd0f3c66f837e9bed0c1241276553a9acecd79541d563f8e013
|
3 |
+
size 30470641
|
attnserver.run_attnserver.slurm.sh.343196.out.log
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
attnserver.run_attnserver.slurm.sh.343197.err.log
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
attnserver.run_attnserver.slurm.sh.343197.out.log
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
attnserver.run_attnserver.slurm.sh.343198.err.log
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
attnserver.run_attnserver.slurm.sh.343198.out.log
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
attnserver.run_attnserver.slurm.sh.343199.err.log
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
attnserver.run_attnserver.slurm.sh.343199.out.log
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
attnserver.run_attnserver.slurm.sh.343200.err.log
ADDED
@@ -0,0 +1,186 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
+ source /mnt/weka/home/hao.zhang/conda/miniconda/bin/activate
|
2 |
+
++ _CONDA_ROOT=/mnt/weka/home/hao.zhang/conda/miniconda
|
3 |
+
++ . /mnt/weka/home/hao.zhang/conda/miniconda/etc/profile.d/conda.sh
|
4 |
+
+++ export CONDA_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda
|
5 |
+
+++ CONDA_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda
|
6 |
+
+++ export _CE_M=
|
7 |
+
+++ _CE_M=
|
8 |
+
+++ export _CE_CONDA=
|
9 |
+
+++ _CE_CONDA=
|
10 |
+
+++ export CONDA_PYTHON_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/python
|
11 |
+
+++ CONDA_PYTHON_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/python
|
12 |
+
+++ '[' -z x ']'
|
13 |
+
++ conda activate
|
14 |
+
++ local cmd=activate
|
15 |
+
++ case "$cmd" in
|
16 |
+
++ __conda_activate activate
|
17 |
+
++ '[' -n '' ']'
|
18 |
+
++ local ask_conda
|
19 |
+
+++ PS1=
|
20 |
+
+++ __conda_exe shell.posix activate
|
21 |
+
+++ '[' -n '' ']'
|
22 |
+
+++ /mnt/weka/home/hao.zhang/conda/miniconda/bin/conda shell.posix activate
|
23 |
+
++ ask_conda='unset _CE_M
|
24 |
+
unset _CE_CONDA
|
25 |
+
PS1='\''(base) '\''
|
26 |
+
export PATH='\''/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin'\''
|
27 |
+
export CONDA_SHLVL='\''1'\''
|
28 |
+
export CONDA_PROMPT_MODIFIER='\''(base) '\''
|
29 |
+
export CONDA_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda'\''
|
30 |
+
export CONDA_PYTHON_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/python'\'''
|
31 |
+
++ eval 'unset _CE_M
|
32 |
+
unset _CE_CONDA
|
33 |
+
PS1='\''(base) '\''
|
34 |
+
export PATH='\''/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin'\''
|
35 |
+
export CONDA_SHLVL='\''1'\''
|
36 |
+
export CONDA_PROMPT_MODIFIER='\''(base) '\''
|
37 |
+
export CONDA_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda'\''
|
38 |
+
export CONDA_PYTHON_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/python'\'''
|
39 |
+
+++ unset _CE_M
|
40 |
+
+++ unset _CE_CONDA
|
41 |
+
+++ PS1='(base) '
|
42 |
+
+++ export PATH=/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin
|
43 |
+
+++ PATH=/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin
|
44 |
+
+++ export CONDA_SHLVL=1
|
45 |
+
+++ CONDA_SHLVL=1
|
46 |
+
+++ export 'CONDA_PROMPT_MODIFIER=(base) '
|
47 |
+
+++ CONDA_PROMPT_MODIFIER='(base) '
|
48 |
+
+++ export CONDA_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda
|
49 |
+
+++ CONDA_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda
|
50 |
+
+++ export CONDA_PYTHON_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/python
|
51 |
+
+++ CONDA_PYTHON_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/python
|
52 |
+
++ __conda_hashr
|
53 |
+
++ '[' -n '' ']'
|
54 |
+
++ '[' -n '' ']'
|
55 |
+
++ hash -r
|
56 |
+
+ conda activate junda-attnserver
|
57 |
+
+ local cmd=activate
|
58 |
+
+ case "$cmd" in
|
59 |
+
+ __conda_activate activate junda-attnserver
|
60 |
+
+ '[' -n '' ']'
|
61 |
+
+ local ask_conda
|
62 |
+
++ PS1='(base) '
|
63 |
+
++ __conda_exe shell.posix activate junda-attnserver
|
64 |
+
++ '[' -n '' ']'
|
65 |
+
++ /mnt/weka/home/hao.zhang/conda/miniconda/bin/conda shell.posix activate junda-attnserver
|
66 |
+
+ ask_conda='unset _CE_M
|
67 |
+
unset _CE_CONDA
|
68 |
+
PS1='\''(junda-attnserver) '\''
|
69 |
+
export PATH='\''/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin'\''
|
70 |
+
export CONDA_PREFIX='\''/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver'\''
|
71 |
+
export CONDA_SHLVL='\''2'\''
|
72 |
+
export CONDA_DEFAULT_ENV='\''junda-attnserver'\''
|
73 |
+
export CONDA_PROMPT_MODIFIER='\''(junda-attnserver) '\''
|
74 |
+
export CONDA_PREFIX_1='\''/mnt/weka/home/hao.zhang/conda/miniconda'\''
|
75 |
+
export CONDA_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda'\''
|
76 |
+
export CONDA_PYTHON_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/python'\'''
|
77 |
+
+ eval 'unset _CE_M
|
78 |
+
unset _CE_CONDA
|
79 |
+
PS1='\''(junda-attnserver) '\''
|
80 |
+
export PATH='\''/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin'\''
|
81 |
+
export CONDA_PREFIX='\''/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver'\''
|
82 |
+
export CONDA_SHLVL='\''2'\''
|
83 |
+
export CONDA_DEFAULT_ENV='\''junda-attnserver'\''
|
84 |
+
export CONDA_PROMPT_MODIFIER='\''(junda-attnserver) '\''
|
85 |
+
export CONDA_PREFIX_1='\''/mnt/weka/home/hao.zhang/conda/miniconda'\''
|
86 |
+
export CONDA_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda'\''
|
87 |
+
export CONDA_PYTHON_EXE='\''/mnt/weka/home/hao.zhang/conda/miniconda/bin/python'\'''
|
88 |
+
++ unset _CE_M
|
89 |
+
++ unset _CE_CONDA
|
90 |
+
++ PS1='(junda-attnserver) '
|
91 |
+
++ export PATH=/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin
|
92 |
+
++ PATH=/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/.local/bin:/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin:/mnt/weka/home/hao.zhang/conda/miniconda/condabin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin
|
93 |
+
++ export CONDA_PREFIX=/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver
|
94 |
+
++ CONDA_PREFIX=/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver
|
95 |
+
++ export CONDA_SHLVL=2
|
96 |
+
++ CONDA_SHLVL=2
|
97 |
+
++ export CONDA_DEFAULT_ENV=junda-attnserver
|
98 |
+
++ CONDA_DEFAULT_ENV=junda-attnserver
|
99 |
+
++ export 'CONDA_PROMPT_MODIFIER=(junda-attnserver) '
|
100 |
+
++ CONDA_PROMPT_MODIFIER='(junda-attnserver) '
|
101 |
+
++ export CONDA_PREFIX_1=/mnt/weka/home/hao.zhang/conda/miniconda
|
102 |
+
++ CONDA_PREFIX_1=/mnt/weka/home/hao.zhang/conda/miniconda
|
103 |
+
++ export CONDA_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda
|
104 |
+
++ CONDA_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/conda
|
105 |
+
++ export CONDA_PYTHON_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/python
|
106 |
+
++ CONDA_PYTHON_EXE=/mnt/weka/home/hao.zhang/conda/miniconda/bin/python
|
107 |
+
+ __conda_hashr
|
108 |
+
+ '[' -n '' ']'
|
109 |
+
+ '[' -n '' ']'
|
110 |
+
+ hash -r
|
111 |
+
+ export CHROME_TRACE_PREFIX=/mnt/sharefs/users/hao.zhang/junda/megatron-prof-data--unstable-v5
|
112 |
+
+ CHROME_TRACE_PREFIX=/mnt/sharefs/users/hao.zhang/junda/megatron-prof-data--unstable-v5
|
113 |
+
+ mkdir -p /mnt/sharefs/users/hao.zhang/junda/megatron-prof-data--unstable-v5
|
114 |
+
+ export PROF_TP_SIZE=8
|
115 |
+
+ PROF_TP_SIZE=8
|
116 |
+
+ export PROF_CP_SIZE=4
|
117 |
+
+ PROF_CP_SIZE=4
|
118 |
+
+ export PROF_BS=32
|
119 |
+
+ PROF_BS=32
|
120 |
+
+ for ctx_length in 1024 2048 4096 8192 12288 16384 24576 32768 40960 49152 65536 81920 98304 131072
|
121 |
+
+ export PROF_CTX_LENGTH=1024
|
122 |
+
+ PROF_CTX_LENGTH=1024
|
123 |
+
+ name='/mnt/sharefs/users/hao.zhang/junda/megatron-prof-data--unstable-v5/mytrace.L1024*tp8.cp4.bs32.json'
|
124 |
+
+ '[' -f '/mnt/sharefs/users/hao.zhang/junda/megatron-prof-data--unstable-v5/mytrace.L1024*tp8.cp4.bs32.json' ']'
|
125 |
+
+ echo 'Running ctx_length=1024, TP_SIZE=8, CP_SIZE=4, BATCH_SIZE=32'
|
126 |
+
+ srun bash ./attnserver.sh
|
127 |
+
+ which python3
|
128 |
+
+ python3 -m torch.distributed.launch --nproc_per_node 8 --nnodes 4 --node_rank 3 --rdzv_id 343200 --rdzv_backend c10d --rdzv_endpoint fs-mbz-gpu-852:29500 ./pretrain_gpt_profile.py --tensor-model-parallel-size 8 --context-parallel-size 4 --num-layers 2 --hidden-size 4096 --num-attention-heads 64 --group-query-attention --num-query-groups 16 --seq-length 1024 --max-position-embeddings 1024 --micro-batch-size 1 --global-batch-size 1 --lr 0.0005 --train-iters 10 --lr-decay-iters 150000 --lr-decay-style cosine --lr-warmup-iters 2 --weight-decay .1 --adam-beta2 .999 --fp16 --log-interval 1 --save-interval 16 --eval-interval 16 --eval-iters 1 --vocab-file vocab.json --merge-file merges.txt --save gpt-checkpoint --load gpt-checkpoint --logging-level 0 --mock-data --tensorboard-dir tensorboard-logs/
|
129 |
+
+ which python3
|
130 |
+
+ python3 -m torch.distributed.launch --nproc_per_node 8 --nnodes 4 --node_rank 2 --rdzv_id 343200 --rdzv_backend c10d --rdzv_endpoint fs-mbz-gpu-852:29500 ./pretrain_gpt_profile.py --tensor-model-parallel-size 8 --context-parallel-size 4 --num-layers 2 --hidden-size 4096 --num-attention-heads 64 --group-query-attention --num-query-groups 16 --seq-length 1024 --max-position-embeddings 1024 --micro-batch-size 1 --global-batch-size 1 --lr 0.0005 --train-iters 10 --lr-decay-iters 150000 --lr-decay-style cosine --lr-warmup-iters 2 --weight-decay .1 --adam-beta2 .999 --fp16 --log-interval 1 --save-interval 16 --eval-interval 16 --eval-iters 1 --vocab-file vocab.json --merge-file merges.txt --save gpt-checkpoint --load gpt-checkpoint --logging-level 0 --mock-data --tensorboard-dir tensorboard-logs/
|
131 |
+
+ which python3
|
132 |
+
+ python3 -m torch.distributed.launch --nproc_per_node 8 --nnodes 4 --node_rank 0 --rdzv_id 343200 --rdzv_backend c10d --rdzv_endpoint fs-mbz-gpu-852:29500 ./pretrain_gpt_profile.py --tensor-model-parallel-size 8 --context-parallel-size 4 --num-layers 2 --hidden-size 4096 --num-attention-heads 64 --group-query-attention --num-query-groups 16 --seq-length 1024 --max-position-embeddings 1024 --micro-batch-size 1 --global-batch-size 1 --lr 0.0005 --train-iters 10 --lr-decay-iters 150000 --lr-decay-style cosine --lr-warmup-iters 2 --weight-decay .1 --adam-beta2 .999 --fp16 --log-interval 1 --save-interval 16 --eval-interval 16 --eval-iters 1 --vocab-file vocab.json --merge-file merges.txt --save gpt-checkpoint --load gpt-checkpoint --logging-level 0 --mock-data --tensorboard-dir tensorboard-logs/
|
133 |
+
+ which python3
|
134 |
+
+ python3 -m torch.distributed.launch --nproc_per_node 8 --nnodes 4 --node_rank 1 --rdzv_id 343200 --rdzv_backend c10d --rdzv_endpoint fs-mbz-gpu-852:29500 ./pretrain_gpt_profile.py --tensor-model-parallel-size 8 --context-parallel-size 4 --num-layers 2 --hidden-size 4096 --num-attention-heads 64 --group-query-attention --num-query-groups 16 --seq-length 1024 --max-position-embeddings 1024 --micro-batch-size 1 --global-batch-size 1 --lr 0.0005 --train-iters 10 --lr-decay-iters 150000 --lr-decay-style cosine --lr-warmup-iters 2 --weight-decay .1 --adam-beta2 .999 --fp16 --log-interval 1 --save-interval 16 --eval-interval 16 --eval-iters 1 --vocab-file vocab.json --merge-file merges.txt --save gpt-checkpoint --load gpt-checkpoint --logging-level 0 --mock-data --tensorboard-dir tensorboard-logs/
|
135 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/lib/python3.12/site-packages/torch/distributed/launch.py:207: FutureWarning: The module torch.distributed.launch is deprecated
|
136 |
+
and will be removed in future. Use torchrun.
|
137 |
+
Note that --use-env is set by default in torchrun.
|
138 |
+
If your script expects `--local-rank` argument to be set, please
|
139 |
+
change it to read from `os.environ['LOCAL_RANK']` instead. See
|
140 |
+
https://pytorch.org/docs/stable/distributed.html#launch-utility for
|
141 |
+
further instructions
|
142 |
+
|
143 |
+
main()
|
144 |
+
W0621 21:08:49.911000 121562 site-packages/torch/distributed/run.py:766]
|
145 |
+
W0621 21:08:49.911000 121562 site-packages/torch/distributed/run.py:766] *****************************************
|
146 |
+
W0621 21:08:49.911000 121562 site-packages/torch/distributed/run.py:766] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
|
147 |
+
W0621 21:08:49.911000 121562 site-packages/torch/distributed/run.py:766] *****************************************
|
148 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/lib/python3.12/site-packages/torch/distributed/launch.py:207: FutureWarning: The module torch.distributed.launch is deprecated
|
149 |
+
and will be removed in future. Use torchrun.
|
150 |
+
Note that --use-env is set by default in torchrun.
|
151 |
+
If your script expects `--local-rank` argument to be set, please
|
152 |
+
change it to read from `os.environ['LOCAL_RANK']` instead. See
|
153 |
+
https://pytorch.org/docs/stable/distributed.html#launch-utility for
|
154 |
+
further instructions
|
155 |
+
|
156 |
+
main()
|
157 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/lib/python3.12/site-packages/torch/distributed/launch.py:207: FutureWarning: The module torch.distributed.launch is deprecated
|
158 |
+
and will be removed in future. Use torchrun.
|
159 |
+
Note that --use-env is set by default in torchrun.
|
160 |
+
If your script expects `--local-rank` argument to be set, please
|
161 |
+
change it to read from `os.environ['LOCAL_RANK']` instead. See
|
162 |
+
https://pytorch.org/docs/stable/distributed.html#launch-utility for
|
163 |
+
further instructions
|
164 |
+
|
165 |
+
main()
|
166 |
+
W0621 21:08:49.997000 2041342 site-packages/torch/distributed/run.py:766]
|
167 |
+
W0621 21:08:49.997000 2041342 site-packages/torch/distributed/run.py:766] *****************************************
|
168 |
+
W0621 21:08:49.997000 2041342 site-packages/torch/distributed/run.py:766] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
|
169 |
+
W0621 21:08:49.997000 2041342 site-packages/torch/distributed/run.py:766] *****************************************
|
170 |
+
W0621 21:08:49.997000 3410371 site-packages/torch/distributed/run.py:766]
|
171 |
+
W0621 21:08:49.997000 3410371 site-packages/torch/distributed/run.py:766] *****************************************
|
172 |
+
W0621 21:08:49.997000 3410371 site-packages/torch/distributed/run.py:766] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
|
173 |
+
W0621 21:08:49.997000 3410371 site-packages/torch/distributed/run.py:766] *****************************************
|
174 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/lib/python3.12/site-packages/torch/distributed/launch.py:207: FutureWarning: The module torch.distributed.launch is deprecated
|
175 |
+
and will be removed in future. Use torchrun.
|
176 |
+
Note that --use-env is set by default in torchrun.
|
177 |
+
If your script expects `--local-rank` argument to be set, please
|
178 |
+
change it to read from `os.environ['LOCAL_RANK']` instead. See
|
179 |
+
https://pytorch.org/docs/stable/distributed.html#launch-utility for
|
180 |
+
further instructions
|
181 |
+
|
182 |
+
main()
|
183 |
+
W0621 21:08:50.019000 3341756 site-packages/torch/distributed/run.py:766]
|
184 |
+
W0621 21:08:50.019000 3341756 site-packages/torch/distributed/run.py:766] *****************************************
|
185 |
+
W0621 21:08:50.019000 3341756 site-packages/torch/distributed/run.py:766] Setting OMP_NUM_THREADS environment variable for each process to be 1 in default, to avoid your system being overloaded, please further tune the variable for optimal performance in your application as needed.
|
186 |
+
W0621 21:08:50.019000 3341756 site-packages/torch/distributed/run.py:766] *****************************************
|
attnserver.run_attnserver.slurm.sh.343200.out.log
ADDED
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
Running ctx_length=1024, TP_SIZE=8, CP_SIZE=4, BATCH_SIZE=32
|
2 |
+
Cleaning up checkpoint directory: gpt-checkpoint
|
3 |
+
--------------------------------
|
4 |
+
CTX_LENGTH: 1024
|
5 |
+
TP_SIZE: 8
|
6 |
+
CP_SIZE: 4
|
7 |
+
CHECKPOINT_PATH: gpt-checkpoint
|
8 |
+
PWD: /mnt/weka/home/hao.zhang/junda/attnserver-megatron
|
9 |
+
--------------------------------
|
10 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin/python3
|
11 |
+
Cleaning up checkpoint directory: gpt-checkpoint
|
12 |
+
--------------------------------
|
13 |
+
CTX_LENGTH: 1024
|
14 |
+
TP_SIZE: 8
|
15 |
+
CP_SIZE: 4
|
16 |
+
CHECKPOINT_PATH: gpt-checkpoint
|
17 |
+
PWD: /mnt/weka/home/hao.zhang/junda/attnserver-megatron
|
18 |
+
--------------------------------
|
19 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin/python3
|
20 |
+
Cleaning up checkpoint directory: gpt-checkpoint
|
21 |
+
--------------------------------
|
22 |
+
CTX_LENGTH: 1024
|
23 |
+
TP_SIZE: 8
|
24 |
+
CP_SIZE: 4
|
25 |
+
CHECKPOINT_PATH: gpt-checkpoint
|
26 |
+
PWD: /mnt/weka/home/hao.zhang/junda/attnserver-megatron
|
27 |
+
--------------------------------
|
28 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin/python3
|
29 |
+
Cleaning up checkpoint directory: gpt-checkpoint
|
30 |
+
--------------------------------
|
31 |
+
CTX_LENGTH: 1024
|
32 |
+
TP_SIZE: 8
|
33 |
+
CP_SIZE: 4
|
34 |
+
CHECKPOINT_PATH: gpt-checkpoint
|
35 |
+
PWD: /mnt/weka/home/hao.zhang/junda/attnserver-megatron
|
36 |
+
--------------------------------
|
37 |
+
/mnt/weka/home/hao.zhang/conda/miniconda/envs/junda-attnserver/bin/python3
|
attnserver.run_attnserver.slurm.sh.343202.out.log
CHANGED
@@ -22196,3 +22196,1165 @@ batch tensor after cp: labels torch.Size([2, 40960])
|
|
22196 |
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22197 |
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22198 |
batch tensor after cp: position_ids torch.Size([2, 40960])
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
22196 |
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22197 |
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22198 |
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22199 |
+
Start exporting trace 0
|
22200 |
+
Done exporting trace 0
|
22201 |
+
Number of parameters in transformer block in billions: 0.35
|
22202 |
+
Number of parameters in embedding layers in billions: 0.21
|
22203 |
+
Total number of parameters in billions: 0.56
|
22204 |
+
Number of parameters in most loaded shard in billions: 0.0703
|
22205 |
+
Theoretical memory footprints: weight and optimizer=1206.09 MB
|
22206 |
+
[2025-06-21 21:07:57] iteration 1/ 10 | consumed samples: 1 | elapsed time per iteration (ms): 29934.7 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 4294967296.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
22207 |
+
[Rank 5] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36108.0 | max reserved: 36108.0
|
22208 |
+
[Rank 1] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36108.0 | max reserved: 36108.0
|
22209 |
+
[Rank 7] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36108.0 | max reserved: 36108.0
|
22210 |
+
[Rank 14] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36148.0 | max reserved: 36148.0
|
22211 |
+
[Rank 12] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36148.0 | max reserved: 36148.0
|
22212 |
+
[Rank 11] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36148.0 | max reserved: 36148.0
|
22213 |
+
[Rank 10] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36148.0 | max reserved: 36148.0
|
22214 |
+
[Rank 3] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36108.0 | max reserved: 36108.0
|
22215 |
+
[Rank 15] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36148.0 | max reserved: 36148.0
|
22216 |
+
[Rank 8] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 35508.0 | max reserved: 35508.0
|
22217 |
+
[Rank 2] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36108.0 | max reserved: 36108.0
|
22218 |
+
[Rank 9] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36148.0 | max reserved: 36148.0
|
22219 |
+
[Rank 0] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 35588.0 | max reserved: 35588.0
|
22220 |
+
[Rank 13] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36148.0 | max reserved: 36148.0
|
22221 |
+
[Rank 6] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36108.0 | max reserved: 36108.0
|
22222 |
+
[Rank 4] (after 1 iterations) memory (MB) | allocated: 22070.03955078125 | max allocated: 33779.52001953125 | reserved: 36108.0 | max reserved: 36108.0
|
22223 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22224 |
+
batch tensor: labels torch.Size([2, 81920])
|
22225 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22226 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22227 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22228 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22229 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22230 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22231 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22232 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22233 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22234 |
+
batch tensor: labels torch.Size([2, 81920])
|
22235 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22236 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22237 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22238 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22239 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22240 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22241 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22242 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22243 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22244 |
+
batch tensor: labels torch.Size([2, 81920])
|
22245 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22246 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22247 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22248 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22249 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22250 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22251 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22252 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22253 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22254 |
+
batch tensor: labels torch.Size([2, 81920])
|
22255 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22256 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22257 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22258 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22259 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22260 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22261 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22262 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22263 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22264 |
+
batch tensor: labels torch.Size([2, 81920])
|
22265 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22266 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22267 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22268 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22269 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22270 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22271 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22272 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22273 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22274 |
+
batch tensor: labels torch.Size([2, 81920])
|
22275 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22276 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22277 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22278 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22279 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22280 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22281 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22282 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22283 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22284 |
+
batch tensor: labels torch.Size([2, 81920])
|
22285 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22286 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22287 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22288 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22289 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22290 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22291 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22292 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22293 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22294 |
+
batch tensor: labels torch.Size([2, 81920])
|
22295 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22296 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22297 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22298 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22299 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22300 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22301 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22302 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22303 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22304 |
+
batch tensor: labels torch.Size([2, 81920])
|
22305 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22306 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22307 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22308 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22309 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22310 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22311 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22312 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22313 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22314 |
+
batch tensor: labels torch.Size([2, 81920])
|
22315 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22316 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22317 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22318 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22319 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22320 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22321 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22322 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22323 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22324 |
+
batch tensor: labels torch.Size([2, 81920])
|
22325 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22326 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22327 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22328 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22329 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22330 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22331 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22332 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22333 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22334 |
+
batch tensor: labels torch.Size([2, 81920])
|
22335 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22336 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22337 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22338 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22339 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22340 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22341 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22342 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22343 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22344 |
+
batch tensor: labels torch.Size([2, 81920])
|
22345 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22346 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22347 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22348 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22349 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22350 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22351 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22352 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22353 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22354 |
+
batch tensor: labels torch.Size([2, 81920])
|
22355 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22356 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22357 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22358 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22359 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22360 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22361 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22362 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22363 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22364 |
+
batch tensor: labels torch.Size([2, 81920])
|
22365 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22366 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22367 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22368 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22369 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22370 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22371 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22372 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22373 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22374 |
+
batch tensor: labels torch.Size([2, 81920])
|
22375 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22376 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22377 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22378 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22379 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22380 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22381 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22382 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22383 |
+
Start exporting trace 1
|
22384 |
+
Done exporting trace 1
|
22385 |
+
[2025-06-21 21:08:05] iteration 2/ 10 | consumed samples: 2 | elapsed time per iteration (ms): 8086.1 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 2147483648.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
22386 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22387 |
+
batch tensor: labels torch.Size([2, 81920])
|
22388 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22389 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22390 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22391 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22392 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22393 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22394 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22395 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22396 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22397 |
+
batch tensor: labels torch.Size([2, 81920])
|
22398 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22399 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22400 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22401 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22402 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22403 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22404 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22405 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22406 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22407 |
+
batch tensor: labels torch.Size([2, 81920])
|
22408 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22409 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22410 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22411 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22412 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22413 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22414 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22415 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22416 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22417 |
+
batch tensor: labels torch.Size([2, 81920])
|
22418 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22419 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22420 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22421 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22422 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22423 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22424 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22425 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22426 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22427 |
+
batch tensor: labels torch.Size([2, 81920])
|
22428 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22429 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22430 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22431 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22432 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22433 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22434 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22435 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22436 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22437 |
+
batch tensor: labels torch.Size([2, 81920])
|
22438 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22439 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22440 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22441 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22442 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22443 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22444 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22445 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22446 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22447 |
+
batch tensor: labels torch.Size([2, 81920])
|
22448 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22449 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22450 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22451 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22452 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22453 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22454 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22455 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22456 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22457 |
+
batch tensor: labels torch.Size([2, 81920])
|
22458 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22459 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22460 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22461 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22462 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22463 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22464 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22465 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22466 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22467 |
+
batch tensor: labels torch.Size([2, 81920])
|
22468 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22469 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22470 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22471 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22472 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22473 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22474 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22475 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22476 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22477 |
+
batch tensor: labels torch.Size([2, 81920])
|
22478 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22479 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22480 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22481 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22482 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22483 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22484 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22485 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22486 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22487 |
+
batch tensor: labels torch.Size([2, 81920])
|
22488 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22489 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22490 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22491 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22492 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22493 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22494 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22495 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22496 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22497 |
+
batch tensor: labels torch.Size([2, 81920])
|
22498 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22499 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22500 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22501 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22502 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22503 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22504 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22505 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22506 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22507 |
+
batch tensor: labels torch.Size([2, 81920])
|
22508 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22509 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22510 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22511 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22512 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22513 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22514 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22515 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22516 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22517 |
+
batch tensor: labels torch.Size([2, 81920])
|
22518 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22519 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22520 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22521 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22522 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22523 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22524 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22525 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22526 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22527 |
+
batch tensor: labels torch.Size([2, 81920])
|
22528 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22529 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22530 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22531 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22532 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22533 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22534 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22535 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22536 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22537 |
+
batch tensor: labels torch.Size([2, 81920])
|
22538 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22539 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22540 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22541 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22542 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22543 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22544 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22545 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22546 |
+
Start exporting trace 2
|
22547 |
+
Done exporting trace 2
|
22548 |
+
[2025-06-21 21:08:13] iteration 3/ 10 | consumed samples: 3 | elapsed time per iteration (ms): 7945.0 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 1073741824.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
22549 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22550 |
+
batch tensor: labels torch.Size([2, 81920])
|
22551 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22552 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22553 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22554 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22555 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22556 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22557 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22558 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22559 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22560 |
+
batch tensor: labels torch.Size([2, 81920])
|
22561 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22562 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22563 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22564 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22565 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22566 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22567 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22568 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22569 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22570 |
+
batch tensor: labels torch.Size([2, 81920])
|
22571 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22572 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22573 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22574 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22575 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22576 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22577 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22578 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22579 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22580 |
+
batch tensor: labels torch.Size([2, 81920])
|
22581 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22582 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22583 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22584 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22585 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22586 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22587 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22588 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22589 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22590 |
+
batch tensor: labels torch.Size([2, 81920])
|
22591 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22592 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22593 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22594 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22595 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22596 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22597 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22598 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22599 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22600 |
+
batch tensor: labels torch.Size([2, 81920])
|
22601 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22602 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22603 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22604 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22605 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22606 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22607 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22608 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22609 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22610 |
+
batch tensor: labels torch.Size([2, 81920])
|
22611 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22612 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22613 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22614 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22615 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22616 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22617 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22618 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22619 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22620 |
+
batch tensor: labels torch.Size([2, 81920])
|
22621 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22622 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22623 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22624 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22625 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22626 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22627 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22628 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22629 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22630 |
+
batch tensor: labels torch.Size([2, 81920])
|
22631 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22632 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22633 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22634 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22635 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22636 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22637 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22638 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22639 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22640 |
+
batch tensor: labels torch.Size([2, 81920])
|
22641 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22642 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22643 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22644 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22645 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22646 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22647 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22648 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22649 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22650 |
+
batch tensor: labels torch.Size([2, 81920])
|
22651 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22652 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22653 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22654 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22655 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22656 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22657 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22658 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22659 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22660 |
+
batch tensor: labels torch.Size([2, 81920])
|
22661 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22662 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22663 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22664 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22665 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22666 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22667 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22668 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22669 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22670 |
+
batch tensor: labels torch.Size([2, 81920])
|
22671 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22672 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22673 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22674 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22675 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22676 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22677 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22678 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22679 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22680 |
+
batch tensor: labels torch.Size([2, 81920])
|
22681 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22682 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22683 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22684 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22685 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22686 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22687 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22688 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22689 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22690 |
+
batch tensor: labels torch.Size([2, 81920])
|
22691 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22692 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22693 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22694 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22695 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22696 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22697 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22698 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22699 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22700 |
+
batch tensor: labels torch.Size([2, 81920])
|
22701 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22702 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22703 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22704 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22705 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22706 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22707 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22708 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22709 |
+
Start exporting trace 3
|
22710 |
+
Done exporting trace 3
|
22711 |
+
[2025-06-21 21:08:21] iteration 4/ 10 | consumed samples: 4 | elapsed time per iteration (ms): 7924.7 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 536870912.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
22712 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22713 |
+
batch tensor: labels torch.Size([2, 81920])
|
22714 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22715 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22716 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22717 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22718 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22719 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22720 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22721 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22722 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22723 |
+
batch tensor: labels torch.Size([2, 81920])
|
22724 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22725 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22726 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22727 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22728 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22729 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22730 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22731 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22732 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22733 |
+
batch tensor: labels torch.Size([2, 81920])
|
22734 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22735 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22736 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22737 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22738 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22739 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22740 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22741 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22742 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22743 |
+
batch tensor: labels torch.Size([2, 81920])
|
22744 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22745 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22746 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22747 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22748 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22749 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22750 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22751 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22752 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22753 |
+
batch tensor: labels torch.Size([2, 81920])
|
22754 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22755 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22756 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22757 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22758 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22759 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22760 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22761 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22762 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22763 |
+
batch tensor: labels torch.Size([2, 81920])
|
22764 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22765 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22766 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22767 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22768 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22769 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22770 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22771 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22772 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22773 |
+
batch tensor: labels torch.Size([2, 81920])
|
22774 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22775 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22776 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22777 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22778 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22779 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22780 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22781 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22782 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22783 |
+
batch tensor: labels torch.Size([2, 81920])
|
22784 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22785 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22786 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22787 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22788 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22789 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22790 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22791 |
+
batch tensor after cp: position_idsbatch tensor: torch.Size([2, 40960])
|
22792 |
+
tokens torch.Size([2, 81920])
|
22793 |
+
batch tensor: labels torch.Size([2, 81920])
|
22794 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22795 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22796 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22797 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22798 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22799 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22800 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22801 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22802 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22803 |
+
batch tensor: labels torch.Size([2, 81920])
|
22804 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22805 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22806 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22807 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22808 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22809 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22810 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22811 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22812 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22813 |
+
batch tensor: labels torch.Size([2, 81920])
|
22814 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22815 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22816 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22817 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22818 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22819 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22820 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22821 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22822 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22823 |
+
batch tensor: labels torch.Size([2, 81920])
|
22824 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22825 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22826 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22827 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22828 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22829 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22830 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22831 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22832 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22833 |
+
batch tensor: labels torch.Size([2, 81920])
|
22834 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22835 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22836 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22837 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22838 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22839 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22840 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22841 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22842 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22843 |
+
batch tensor: labels torch.Size([2, 81920])
|
22844 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22845 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22846 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22847 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22848 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22849 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22850 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22851 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22852 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22853 |
+
batch tensor: labels torch.Size([2, 81920])
|
22854 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22855 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22856 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22857 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22858 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22859 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22860 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22861 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22862 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22863 |
+
batch tensor: labels torch.Size([2, 81920])
|
22864 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22865 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22866 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22867 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22868 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22869 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22870 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22871 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22872 |
+
Start exporting trace 4
|
22873 |
+
Done exporting trace 4
|
22874 |
+
[2025-06-21 21:08:28] iteration 5/ 10 | consumed samples: 5 | elapsed time per iteration (ms): 7646.0 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 268435456.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
22875 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22876 |
+
batch tensor: labels torch.Size([2, 81920])
|
22877 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22878 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22879 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22880 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22881 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22882 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22883 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22884 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22885 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22886 |
+
batch tensor: labels torch.Size([2, 81920])
|
22887 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22888 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22889 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22890 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22891 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22892 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22893 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22894 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22895 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22896 |
+
batch tensor: labels torch.Size([2, 81920])
|
22897 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22898 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22899 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22900 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22901 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22902 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22903 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22904 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22905 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22906 |
+
batch tensor: labels torch.Size([2, 81920])
|
22907 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22908 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22909 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22910 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22911 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22912 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22913 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22914 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22915 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22916 |
+
batch tensor: labels torch.Size([2, 81920])
|
22917 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22918 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22919 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22920 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22921 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22922 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22923 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22924 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22925 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22926 |
+
batch tensor: labels torch.Size([2, 81920])
|
22927 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22928 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22929 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22930 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22931 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22932 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22933 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22934 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22935 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22936 |
+
batch tensor: labels torch.Size([2, 81920])
|
22937 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22938 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22939 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22940 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22941 |
+
batch tensor: labels torch.Size([2, 81920])
|
22942 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22943 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22944 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22945 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22946 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22947 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22948 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22949 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22950 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22951 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22952 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22953 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22954 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22955 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22956 |
+
batch tensor: labels torch.Size([2, 81920])
|
22957 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22958 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22959 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22960 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22961 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22962 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22963 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22964 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22965 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22966 |
+
batch tensor: labels torch.Size([2, 81920])
|
22967 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22968 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22969 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22970 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22971 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22972 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22973 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22974 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22975 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22976 |
+
batch tensor: labels torch.Size([2, 81920])
|
22977 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22978 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22979 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22980 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22981 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22982 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22983 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22984 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22985 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22986 |
+
batch tensor: labels torch.Size([2, 81920])
|
22987 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22988 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22989 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
22990 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
22991 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
22992 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
22993 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
22994 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
22995 |
+
batch tensor: tokens torch.Size([2, 81920])
|
22996 |
+
batch tensor: labels torch.Size([2, 81920])
|
22997 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
22998 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
22999 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23000 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23001 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23002 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23003 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23004 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23005 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23006 |
+
batch tensor: labels torch.Size([2, 81920])
|
23007 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23008 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23009 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23010 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23011 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23012 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23013 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23014 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23015 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23016 |
+
batch tensor: labels torch.Size([2, 81920])
|
23017 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23018 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23019 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23020 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23021 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23022 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23023 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23024 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23025 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23026 |
+
batch tensor: labels torch.Size([2, 81920])
|
23027 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23028 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23029 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23030 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23031 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23032 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23033 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23034 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23035 |
+
Start exporting trace 5
|
23036 |
+
Done exporting trace 5
|
23037 |
+
[2025-06-21 21:08:36] iteration 6/ 10 | consumed samples: 6 | elapsed time per iteration (ms): 7989.7 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 134217728.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
23038 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23039 |
+
batch tensor: labels torch.Size([2, 81920])
|
23040 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23041 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23042 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23043 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23044 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23045 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23046 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23047 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23048 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23049 |
+
batch tensor: labels torch.Size([2, 81920])
|
23050 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23051 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23052 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23053 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23054 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23055 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23056 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23057 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23058 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23059 |
+
batch tensor: labels torch.Size([2, 81920])
|
23060 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23061 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23062 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23063 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23064 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23065 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23066 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23067 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23068 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23069 |
+
batch tensor: labels torch.Size([2, 81920])
|
23070 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23071 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23072 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23073 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23074 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23075 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23076 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23077 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23078 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23079 |
+
batch tensor: labels torch.Size([2, 81920])
|
23080 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23081 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23082 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23083 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23084 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23085 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23086 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23087 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23088 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23089 |
+
batch tensor: labels torch.Size([2, 81920])
|
23090 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23091 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23092 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23093 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23094 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23095 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23096 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23097 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23098 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23099 |
+
batch tensor: labels torch.Size([2, 81920])
|
23100 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23101 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23102 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23103 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23104 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23105 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23106 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23107 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23108 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23109 |
+
batch tensor: labels torch.Size([2, 81920])
|
23110 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23111 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23112 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23113 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23114 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23115 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23116 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23117 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23118 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23119 |
+
batch tensor: labels torch.Size([2, 81920])
|
23120 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23121 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23122 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23123 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23124 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23125 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23126 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23127 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23128 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23129 |
+
batch tensor: labels torch.Size([2, 81920])
|
23130 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23131 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23132 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23133 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23134 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23135 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23136 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23137 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23138 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23139 |
+
batch tensor: labels torch.Size([2, 81920])
|
23140 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23141 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23142 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23143 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23144 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23145 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23146 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23147 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23148 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23149 |
+
batch tensor: labels torch.Size([2, 81920])
|
23150 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23151 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23152 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23153 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23154 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23155 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23156 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23157 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23158 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23159 |
+
batch tensor: labels torch.Size([2, 81920])
|
23160 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23161 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23162 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23163 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23164 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23165 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23166 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23167 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23168 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23169 |
+
batch tensor: labels torch.Size([2, 81920])
|
23170 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23171 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23172 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23173 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23174 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23175 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23176 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23177 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23178 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23179 |
+
batch tensor: labels torch.Size([2, 81920])
|
23180 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23181 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23182 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23183 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23184 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23185 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23186 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23187 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23188 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23189 |
+
batch tensor: labels torch.Size([2, 81920])
|
23190 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23191 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23192 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23193 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23194 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23195 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23196 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23197 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23198 |
+
Start exporting trace 6
|
23199 |
+
Done exporting trace 6
|
23200 |
+
[2025-06-21 21:08:44] iteration 7/ 10 | consumed samples: 7 | elapsed time per iteration (ms): 7801.4 | learning rate: 0.000000E+00 | global batch size: 1 | loss scale: 67108864.0 | number of skipped iterations: 1 | number of nan iterations: 0 |
|
23201 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23202 |
+
batch tensor: labels torch.Size([2, 81920])
|
23203 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23204 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23205 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23206 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23207 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23208 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23209 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23210 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23211 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23212 |
+
batch tensor: labels torch.Size([2, 81920])
|
23213 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23214 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23215 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23216 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23217 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23218 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23219 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23220 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23221 |
+
batch tensor: labels torch.Size([2, 81920])
|
23222 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23223 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23224 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23225 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23226 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23227 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23228 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23229 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23230 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23231 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23232 |
+
batch tensor: labels torch.Size([2, 81920])
|
23233 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23234 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23235 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23236 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23237 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23238 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23239 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23240 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23241 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23242 |
+
batch tensor: labels torch.Size([2, 81920])
|
23243 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23244 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23245 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23246 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23247 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23248 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23249 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23250 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23251 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23252 |
+
batch tensor: labels torch.Size([2, 81920])
|
23253 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23254 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23255 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23256 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23257 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23258 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23259 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23260 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23261 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23262 |
+
batch tensor: labels torch.Size([2, 81920])
|
23263 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23264 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23265 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23266 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23267 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23268 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23269 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23270 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23271 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23272 |
+
batch tensor: labels torch.Size([2, 81920])
|
23273 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23274 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23275 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23276 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23277 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23278 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23279 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23280 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23281 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23282 |
+
batch tensor: labels torch.Size([2, 81920])
|
23283 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23284 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23285 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23286 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23287 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23288 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23289 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23290 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23291 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23292 |
+
batch tensor: labels torch.Size([2, 81920])
|
23293 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23294 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23295 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23296 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23297 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23298 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23299 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23300 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23301 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23302 |
+
batch tensor: labels torch.Size([2, 81920])
|
23303 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23304 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23305 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23306 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23307 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23308 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23309 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23310 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23311 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23312 |
+
batch tensor: labels torch.Size([2, 81920])
|
23313 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23314 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23315 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23316 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23317 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23318 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23319 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23320 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23321 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23322 |
+
batch tensor: labels torch.Size([2, 81920])
|
23323 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23324 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23325 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23326 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23327 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23328 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23329 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23330 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23331 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23332 |
+
batch tensor: labels torch.Size([2, 81920])
|
23333 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23334 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23335 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23336 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23337 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23338 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23339 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23340 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23341 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23342 |
+
batch tensor: labels torch.Size([2, 81920])
|
23343 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23344 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23345 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23346 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23347 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23348 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23349 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23350 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|
23351 |
+
batch tensor: tokens torch.Size([2, 81920])
|
23352 |
+
batch tensor: labels torch.Size([2, 81920])
|
23353 |
+
batch tensor: loss_mask torch.Size([2, 81920])
|
23354 |
+
batch tensor: attention_mask torch.Size([2, 1, 81920, 81920])
|
23355 |
+
batch tensor: position_ids torch.Size([2, 81920])
|
23356 |
+
batch tensor after cp: tokens torch.Size([2, 40960])
|
23357 |
+
batch tensor after cp: labels torch.Size([2, 40960])
|
23358 |
+
batch tensor after cp: loss_mask torch.Size([2, 40960])
|
23359 |
+
batch tensor after cp: attention_mask torch.Size([2, 1, 40960, 81920])
|
23360 |
+
batch tensor after cp: position_ids torch.Size([2, 40960])
|