ardaspear commited on
Commit
07960d1
·
verified ·
1 Parent(s): f825057

Training in progress, step 34, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df3189200d2a19afd6bf65e1b7abee07cc56a66a3605f3e697d63af3c3d6a814
3
  size 159967880
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef352b3755cd2c3514a70e0f1913121f3c90f77b381ac85eeb778430e7a0492f
3
  size 159967880
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c6c4b751dac4de342f4aba0e1bf7d18453a9fe7ff26a6340f73105e47974449c
3
  size 81730196
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:255220845625b8a532eac68bcefb4bfe60d87d73a6ff5f76ef1a4f52d1f270d8
3
  size 81730196
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8dc6c5ef268f1183211150fcad03d95bec24795d013490d0fe87a02db0afe595
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:624617eb67c5ae66a26f74399b042214fe4beaf1a0d11ffcede0e4228de76c62
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:61d8ec93266e4bc8a64331da6b277e98dfae3d3d4bab70d66d65fad27bc26c8c
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0af27ed5b5e5c3013f1da7a97e494138e751399ceff1f8e7486b6e269bc7092f
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.00436316971446904,
5
  "eval_steps": 17,
6
- "global_step": 17,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -58,6 +58,56 @@
58
  "eval_samples_per_second": 13.291,
59
  "eval_steps_per_second": 1.663,
60
  "step": 17
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
61
  }
62
  ],
63
  "logging_steps": 3,
@@ -77,7 +127,7 @@
77
  "attributes": {}
78
  }
79
  },
80
- "total_flos": 2.4391533604110336e+16,
81
  "train_batch_size": 8,
82
  "trial_name": null,
83
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.00872633942893808,
5
  "eval_steps": 17,
6
+ "global_step": 34,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
58
  "eval_samples_per_second": 13.291,
59
  "eval_steps_per_second": 1.663,
60
  "step": 17
61
+ },
62
+ {
63
+ "epoch": 0.0046198267564966315,
64
+ "grad_norm": 0.3675502836704254,
65
+ "learning_rate": 9.956320346634876e-05,
66
+ "loss": 2.4945,
67
+ "step": 18
68
+ },
69
+ {
70
+ "epoch": 0.005389797882579404,
71
+ "grad_norm": 0.4536501169204712,
72
+ "learning_rate": 9.917525374361912e-05,
73
+ "loss": 2.5246,
74
+ "step": 21
75
+ },
76
+ {
77
+ "epoch": 0.006159769008662175,
78
+ "grad_norm": 0.39700689911842346,
79
+ "learning_rate": 9.86663298624003e-05,
80
+ "loss": 2.4664,
81
+ "step": 24
82
+ },
83
+ {
84
+ "epoch": 0.006929740134744947,
85
+ "grad_norm": 0.3603559732437134,
86
+ "learning_rate": 9.803768380684242e-05,
87
+ "loss": 2.4274,
88
+ "step": 27
89
+ },
90
+ {
91
+ "epoch": 0.007699711260827719,
92
+ "grad_norm": 0.368173211812973,
93
+ "learning_rate": 9.729086208503174e-05,
94
+ "loss": 2.4661,
95
+ "step": 30
96
+ },
97
+ {
98
+ "epoch": 0.00846968238691049,
99
+ "grad_norm": 0.35239627957344055,
100
+ "learning_rate": 9.642770192448536e-05,
101
+ "loss": 2.3349,
102
+ "step": 33
103
+ },
104
+ {
105
+ "epoch": 0.00872633942893808,
106
+ "eval_loss": 2.3943283557891846,
107
+ "eval_runtime": 493.7443,
108
+ "eval_samples_per_second": 13.292,
109
+ "eval_steps_per_second": 1.663,
110
+ "step": 34
111
  }
112
  ],
113
  "logging_steps": 3,
 
127
  "attributes": {}
128
  }
129
  },
130
+ "total_flos": 4.977191316514406e+16,
131
  "train_batch_size": 8,
132
  "trial_name": null,
133
  "trial_params": null