SystemAdmin123 commited on
Commit
1888915
·
verified ·
1 Parent(s): 34bc671

Training in progress, step 2500, checkpoint

Browse files
last-checkpoint/model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:39b65f1f5bdc8e68c678fe87b09e42b54b9eba1640a82fd33222297da0ee47b3
3
  size 2066752
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4156994b0d2538aa2d37af1880314c638cc4d24a008e661dbcf7a289e3478dca
3
  size 2066752
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef4e35f5dcad96cb1089fe9a38ce911a58e849647e472ddf5a739d5e6986aa33
3
  size 2162798
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fba8dbfefb423350e220e5a13f8bccbed1d56b9ffd0aa6cc568c8e5687f191cf
3
  size 2162798
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:19ab3d6cfcb43de67f16e412d0cb4f86309db602f8242d16f2b203a0212d6cbb
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:97474a95cf2d0e6166f036d8937e33ebebb2adb23cf1177f88edc10dc549c905
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c88b3aeb8ec2bf995149291b90b69667d3f268ff2f13afbeab1a220b8cc27590
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b204b20cdc92a140e2e21e015bdaa04af008c00e0bde30e59edf0f23817a338
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.5918910920390648,
5
  "eval_steps": 200,
6
- "global_step": 2000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1495,6 +1495,372 @@
1495
  "eval_samples_per_second": 94.292,
1496
  "eval_steps_per_second": 23.604,
1497
  "step": 2000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1498
  }
1499
  ],
1500
  "logging_steps": 10,
@@ -1509,12 +1875,12 @@
1509
  "should_evaluate": false,
1510
  "should_log": false,
1511
  "should_save": true,
1512
- "should_training_stop": false
1513
  },
1514
  "attributes": {}
1515
  }
1516
  },
1517
- "total_flos": 51272680734720.0,
1518
  "train_batch_size": 4,
1519
  "trial_name": null,
1520
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.7398638650488311,
5
  "eval_steps": 200,
6
+ "global_step": 2500,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1495
  "eval_samples_per_second": 94.292,
1496
  "eval_steps_per_second": 23.604,
1497
  "step": 2000
1498
+ },
1499
+ {
1500
+ "epoch": 0.5948505474992601,
1501
+ "grad_norm": 0.55078125,
1502
+ "learning_rate": 2.0280410844810428e-05,
1503
+ "loss": 8.6746,
1504
+ "step": 2010
1505
+ },
1506
+ {
1507
+ "epoch": 0.5978100029594554,
1508
+ "grad_norm": 0.490234375,
1509
+ "learning_rate": 1.9488821249060297e-05,
1510
+ "loss": 8.7101,
1511
+ "step": 2020
1512
+ },
1513
+ {
1514
+ "epoch": 0.6007694584196508,
1515
+ "grad_norm": 0.546875,
1516
+ "learning_rate": 1.871131877836879e-05,
1517
+ "loss": 8.6891,
1518
+ "step": 2030
1519
+ },
1520
+ {
1521
+ "epoch": 0.6037289138798461,
1522
+ "grad_norm": 0.60546875,
1523
+ "learning_rate": 1.7948039473155554e-05,
1524
+ "loss": 8.6504,
1525
+ "step": 2040
1526
+ },
1527
+ {
1528
+ "epoch": 0.6066883693400414,
1529
+ "grad_norm": 1.0390625,
1530
+ "learning_rate": 1.7199116885197995e-05,
1531
+ "loss": 8.7523,
1532
+ "step": 2050
1533
+ },
1534
+ {
1535
+ "epoch": 0.6096478248002367,
1536
+ "grad_norm": 0.4921875,
1537
+ "learning_rate": 1.646468205426377e-05,
1538
+ "loss": 8.6832,
1539
+ "step": 2060
1540
+ },
1541
+ {
1542
+ "epoch": 0.6126072802604321,
1543
+ "grad_norm": 0.466796875,
1544
+ "learning_rate": 1.5744863485182537e-05,
1545
+ "loss": 8.7104,
1546
+ "step": 2070
1547
+ },
1548
+ {
1549
+ "epoch": 0.6155667357206274,
1550
+ "grad_norm": 0.55859375,
1551
+ "learning_rate": 1.5039787125361326e-05,
1552
+ "loss": 8.6838,
1553
+ "step": 2080
1554
+ },
1555
+ {
1556
+ "epoch": 0.6185261911808227,
1557
+ "grad_norm": 0.7421875,
1558
+ "learning_rate": 1.4349576342747462e-05,
1559
+ "loss": 8.7315,
1560
+ "step": 2090
1561
+ },
1562
+ {
1563
+ "epoch": 0.621485646641018,
1564
+ "grad_norm": 1.2109375,
1565
+ "learning_rate": 1.3674351904242611e-05,
1566
+ "loss": 8.622,
1567
+ "step": 2100
1568
+ },
1569
+ {
1570
+ "epoch": 0.6244451021012134,
1571
+ "grad_norm": 0.4765625,
1572
+ "learning_rate": 1.3014231954572287e-05,
1573
+ "loss": 8.6708,
1574
+ "step": 2110
1575
+ },
1576
+ {
1577
+ "epoch": 0.6274045575614087,
1578
+ "grad_norm": 0.6171875,
1579
+ "learning_rate": 1.2369331995613665e-05,
1580
+ "loss": 8.6699,
1581
+ "step": 2120
1582
+ },
1583
+ {
1584
+ "epoch": 0.630364013021604,
1585
+ "grad_norm": 0.56640625,
1586
+ "learning_rate": 1.173976486618631e-05,
1587
+ "loss": 8.6815,
1588
+ "step": 2130
1589
+ },
1590
+ {
1591
+ "epoch": 0.6333234684817993,
1592
+ "grad_norm": 0.69921875,
1593
+ "learning_rate": 1.1125640722308628e-05,
1594
+ "loss": 8.7564,
1595
+ "step": 2140
1596
+ },
1597
+ {
1598
+ "epoch": 0.6362829239419947,
1599
+ "grad_norm": 1.1328125,
1600
+ "learning_rate": 1.0527067017923654e-05,
1601
+ "loss": 8.6414,
1602
+ "step": 2150
1603
+ },
1604
+ {
1605
+ "epoch": 0.63924237940219,
1606
+ "grad_norm": 0.4609375,
1607
+ "learning_rate": 9.944148486097793e-06,
1608
+ "loss": 8.7073,
1609
+ "step": 2160
1610
+ },
1611
+ {
1612
+ "epoch": 0.6422018348623854,
1613
+ "grad_norm": 0.45703125,
1614
+ "learning_rate": 9.376987120695545e-06,
1615
+ "loss": 8.6823,
1616
+ "step": 2170
1617
+ },
1618
+ {
1619
+ "epoch": 0.6451612903225806,
1620
+ "grad_norm": 0.56640625,
1621
+ "learning_rate": 8.825682158533554e-06,
1622
+ "loss": 8.6332,
1623
+ "step": 2180
1624
+ },
1625
+ {
1626
+ "epoch": 0.648120745782776,
1627
+ "grad_norm": 0.6953125,
1628
+ "learning_rate": 8.290330062017016e-06,
1629
+ "loss": 8.6951,
1630
+ "step": 2190
1631
+ },
1632
+ {
1633
+ "epoch": 0.6510802012429713,
1634
+ "grad_norm": 0.95703125,
1635
+ "learning_rate": 7.771024502261526e-06,
1636
+ "loss": 8.7376,
1637
+ "step": 2200
1638
+ },
1639
+ {
1640
+ "epoch": 0.6510802012429713,
1641
+ "eval_loss": 8.695413589477539,
1642
+ "eval_runtime": 18.9893,
1643
+ "eval_samples_per_second": 79.097,
1644
+ "eval_steps_per_second": 19.801,
1645
+ "step": 2200
1646
+ },
1647
+ {
1648
+ "epoch": 0.6540396567031667,
1649
+ "grad_norm": 0.455078125,
1650
+ "learning_rate": 7.267856342703461e-06,
1651
+ "loss": 8.714,
1652
+ "step": 2210
1653
+ },
1654
+ {
1655
+ "epoch": 0.6569991121633619,
1656
+ "grad_norm": 0.52734375,
1657
+ "learning_rate": 6.780913623201346e-06,
1658
+ "loss": 8.6495,
1659
+ "step": 2220
1660
+ },
1661
+ {
1662
+ "epoch": 0.6599585676235573,
1663
+ "grad_norm": 0.61328125,
1664
+ "learning_rate": 6.310281544631546e-06,
1665
+ "loss": 8.7043,
1666
+ "step": 2230
1667
+ },
1668
+ {
1669
+ "epoch": 0.6629180230837526,
1670
+ "grad_norm": 0.62890625,
1671
+ "learning_rate": 5.856042453980526e-06,
1672
+ "loss": 8.6877,
1673
+ "step": 2240
1674
+ },
1675
+ {
1676
+ "epoch": 0.665877478543948,
1677
+ "grad_norm": 1.25,
1678
+ "learning_rate": 5.418275829936537e-06,
1679
+ "loss": 8.6366,
1680
+ "step": 2250
1681
+ },
1682
+ {
1683
+ "epoch": 0.6688369340041432,
1684
+ "grad_norm": 0.470703125,
1685
+ "learning_rate": 4.997058268983135e-06,
1686
+ "loss": 8.7548,
1687
+ "step": 2260
1688
+ },
1689
+ {
1690
+ "epoch": 0.6717963894643386,
1691
+ "grad_norm": 0.53515625,
1692
+ "learning_rate": 4.592463471997022e-06,
1693
+ "loss": 8.6756,
1694
+ "step": 2270
1695
+ },
1696
+ {
1697
+ "epoch": 0.6747558449245339,
1698
+ "grad_norm": 0.482421875,
1699
+ "learning_rate": 4.204562231352516e-06,
1700
+ "loss": 8.6466,
1701
+ "step": 2280
1702
+ },
1703
+ {
1704
+ "epoch": 0.6777153003847292,
1705
+ "grad_norm": 0.58203125,
1706
+ "learning_rate": 3.83342241853496e-06,
1707
+ "loss": 8.6794,
1708
+ "step": 2290
1709
+ },
1710
+ {
1711
+ "epoch": 0.6806747558449245,
1712
+ "grad_norm": 1.171875,
1713
+ "learning_rate": 3.4791089722651436e-06,
1714
+ "loss": 8.6301,
1715
+ "step": 2300
1716
+ },
1717
+ {
1718
+ "epoch": 0.6836342113051198,
1719
+ "grad_norm": 0.546875,
1720
+ "learning_rate": 3.1416838871368924e-06,
1721
+ "loss": 8.6403,
1722
+ "step": 2310
1723
+ },
1724
+ {
1725
+ "epoch": 0.6865936667653152,
1726
+ "grad_norm": 0.5,
1727
+ "learning_rate": 2.821206202769899e-06,
1728
+ "loss": 8.7124,
1729
+ "step": 2320
1730
+ },
1731
+ {
1732
+ "epoch": 0.6895531222255105,
1733
+ "grad_norm": 0.765625,
1734
+ "learning_rate": 2.5177319934794e-06,
1735
+ "loss": 8.6445,
1736
+ "step": 2330
1737
+ },
1738
+ {
1739
+ "epoch": 0.6925125776857058,
1740
+ "grad_norm": 0.6328125,
1741
+ "learning_rate": 2.2313143584648423e-06,
1742
+ "loss": 8.6972,
1743
+ "step": 2340
1744
+ },
1745
+ {
1746
+ "epoch": 0.6954720331459011,
1747
+ "grad_norm": 1.0625,
1748
+ "learning_rate": 1.9620034125190644e-06,
1749
+ "loss": 8.9133,
1750
+ "step": 2350
1751
+ },
1752
+ {
1753
+ "epoch": 0.6984314886060965,
1754
+ "grad_norm": 0.44921875,
1755
+ "learning_rate": 1.7098462772596302e-06,
1756
+ "loss": 8.6366,
1757
+ "step": 2360
1758
+ },
1759
+ {
1760
+ "epoch": 0.7013909440662918,
1761
+ "grad_norm": 0.51953125,
1762
+ "learning_rate": 1.4748870728839347e-06,
1763
+ "loss": 8.6377,
1764
+ "step": 2370
1765
+ },
1766
+ {
1767
+ "epoch": 0.7043503995264871,
1768
+ "grad_norm": 0.470703125,
1769
+ "learning_rate": 1.2571669104494256e-06,
1770
+ "loss": 8.6656,
1771
+ "step": 2380
1772
+ },
1773
+ {
1774
+ "epoch": 0.7073098549866824,
1775
+ "grad_norm": 0.71484375,
1776
+ "learning_rate": 1.0567238846803996e-06,
1777
+ "loss": 8.6525,
1778
+ "step": 2390
1779
+ },
1780
+ {
1781
+ "epoch": 0.7102693104468778,
1782
+ "grad_norm": 0.7421875,
1783
+ "learning_rate": 8.735930673024806e-07,
1784
+ "loss": 8.573,
1785
+ "step": 2400
1786
+ },
1787
+ {
1788
+ "epoch": 0.7102693104468778,
1789
+ "eval_loss": 8.698899269104004,
1790
+ "eval_runtime": 10.0141,
1791
+ "eval_samples_per_second": 149.989,
1792
+ "eval_steps_per_second": 37.547,
1793
+ "step": 2400
1794
+ },
1795
+ {
1796
+ "epoch": 0.7132287659070731,
1797
+ "grad_norm": 0.4921875,
1798
+ "learning_rate": 7.078065009060941e-07,
1799
+ "loss": 8.6492,
1800
+ "step": 2410
1801
+ },
1802
+ {
1803
+ "epoch": 0.7161882213672685,
1804
+ "grad_norm": 0.4765625,
1805
+ "learning_rate": 5.593931933399854e-07,
1806
+ "loss": 8.6976,
1807
+ "step": 2420
1808
+ },
1809
+ {
1810
+ "epoch": 0.7191476768274637,
1811
+ "grad_norm": 0.515625,
1812
+ "learning_rate": 4.2837911263562404e-07,
1813
+ "loss": 8.6823,
1814
+ "step": 2430
1815
+ },
1816
+ {
1817
+ "epoch": 0.7221071322876591,
1818
+ "grad_norm": 0.6875,
1819
+ "learning_rate": 3.1478718246357173e-07,
1820
+ "loss": 8.7054,
1821
+ "step": 2440
1822
+ },
1823
+ {
1824
+ "epoch": 0.7250665877478544,
1825
+ "grad_norm": 1.28125,
1826
+ "learning_rate": 2.1863727812254653e-07,
1827
+ "loss": 8.81,
1828
+ "step": 2450
1829
+ },
1830
+ {
1831
+ "epoch": 0.7280260432080498,
1832
+ "grad_norm": 0.427734375,
1833
+ "learning_rate": 1.3994622306173765e-07,
1834
+ "loss": 8.6948,
1835
+ "step": 2460
1836
+ },
1837
+ {
1838
+ "epoch": 0.730985498668245,
1839
+ "grad_norm": 0.48046875,
1840
+ "learning_rate": 7.872778593728258e-08,
1841
+ "loss": 8.6751,
1842
+ "step": 2470
1843
+ },
1844
+ {
1845
+ "epoch": 0.7339449541284404,
1846
+ "grad_norm": 0.59375,
1847
+ "learning_rate": 3.499267820307184e-08,
1848
+ "loss": 8.7253,
1849
+ "step": 2480
1850
+ },
1851
+ {
1852
+ "epoch": 0.7369044095886357,
1853
+ "grad_norm": 0.60546875,
1854
+ "learning_rate": 8.748552236603757e-09,
1855
+ "loss": 8.7342,
1856
+ "step": 2490
1857
+ },
1858
+ {
1859
+ "epoch": 0.7398638650488311,
1860
+ "grad_norm": 0.91796875,
1861
+ "learning_rate": 0.0,
1862
+ "loss": 8.645,
1863
+ "step": 2500
1864
  }
1865
  ],
1866
  "logging_steps": 10,
 
1875
  "should_evaluate": false,
1876
  "should_log": false,
1877
  "should_save": true,
1878
+ "should_training_stop": true
1879
  },
1880
  "attributes": {}
1881
  }
1882
  },
1883
+ "total_flos": 64161175044096.0,
1884
  "train_batch_size": 4,
1885
  "trial_name": null,
1886
  "trial_params": null