End of training
Browse files
README.md
ADDED
@@ -0,0 +1,256 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: transformers
|
3 |
+
tags:
|
4 |
+
- generated_from_trainer
|
5 |
+
metrics:
|
6 |
+
- accuracy
|
7 |
+
model-index:
|
8 |
+
- name: pretrained-hist-l2_tenKQ_finetune-itemseg_v12-tssp-m0
|
9 |
+
results: []
|
10 |
+
---
|
11 |
+
|
12 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
13 |
+
should probably proofread and complete it, then remove this comment. -->
|
14 |
+
|
15 |
+
# pretrained-hist-l2_tenKQ_finetune-itemseg_v12-tssp-m0
|
16 |
+
|
17 |
+
This model was trained from scratch on an unknown dataset.
|
18 |
+
It achieves the following results on the evaluation set:
|
19 |
+
- Loss: 1.3248
|
20 |
+
- Accuracy: 0.9294
|
21 |
+
- Macro F1: 0.8204
|
22 |
+
|
23 |
+
## Model description
|
24 |
+
|
25 |
+
More information needed
|
26 |
+
|
27 |
+
## Intended uses & limitations
|
28 |
+
|
29 |
+
More information needed
|
30 |
+
|
31 |
+
## Training and evaluation data
|
32 |
+
|
33 |
+
More information needed
|
34 |
+
|
35 |
+
## Training procedure
|
36 |
+
|
37 |
+
### Training hyperparameters
|
38 |
+
|
39 |
+
The following hyperparameters were used during training:
|
40 |
+
- learning_rate: 0.0001
|
41 |
+
- train_batch_size: 16
|
42 |
+
- eval_batch_size: 16
|
43 |
+
- seed: 42
|
44 |
+
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
45 |
+
- lr_scheduler_type: linear
|
46 |
+
- lr_scheduler_warmup_steps: 3365
|
47 |
+
- training_steps: 67312
|
48 |
+
|
49 |
+
### Training results
|
50 |
+
|
51 |
+
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Macro F1 |
|
52 |
+
|:-------------:|:--------:|:-----:|:---------------:|:--------:|:--------:|
|
53 |
+
| 18.3081 | 2.0010 | 201 | 92.0443 | 0.1413 | 0.0501 |
|
54 |
+
| 6.328 | 5.0009 | 402 | 133.2266 | 0.5761 | 0.1625 |
|
55 |
+
| 4.838 | 8.0009 | 603 | 145.9778 | 0.6359 | 0.1990 |
|
56 |
+
| 3.6763 | 11.0008 | 804 | 119.4216 | 0.6674 | 0.2212 |
|
57 |
+
| 2.7128 | 14.0008 | 1005 | 51.8430 | 0.7070 | 0.2573 |
|
58 |
+
| 2.3342 | 17.0007 | 1206 | 27.4778 | 0.7250 | 0.2894 |
|
59 |
+
| 1.8815 | 20.0007 | 1407 | 19.0125 | 0.7462 | 0.3373 |
|
60 |
+
| 1.65 | 23.0007 | 1608 | 13.0926 | 0.7608 | 0.3662 |
|
61 |
+
| 1.4902 | 26.0006 | 1809 | 8.4115 | 0.7567 | 0.4072 |
|
62 |
+
| 1.2953 | 29.0006 | 2010 | 7.8777 | 0.7823 | 0.4407 |
|
63 |
+
| 1.1867 | 32.0005 | 2211 | 6.7955 | 0.7812 | 0.4557 |
|
64 |
+
| 1.0704 | 35.0005 | 2412 | 5.3807 | 0.7841 | 0.4912 |
|
65 |
+
| 0.9792 | 38.0004 | 2613 | 4.6642 | 0.8160 | 0.5335 |
|
66 |
+
| 0.8649 | 41.0004 | 2814 | 4.1745 | 0.8178 | 0.5658 |
|
67 |
+
| 0.7768 | 44.0003 | 3015 | 3.8817 | 0.8240 | 0.5736 |
|
68 |
+
| 0.7169 | 47.0003 | 3216 | 4.2183 | 0.8330 | 0.5971 |
|
69 |
+
| 0.6459 | 50.0003 | 3417 | 3.4359 | 0.8295 | 0.6069 |
|
70 |
+
| 0.5664 | 53.0002 | 3618 | 3.3957 | 0.8538 | 0.6294 |
|
71 |
+
| 0.516 | 56.0002 | 3819 | 3.8816 | 0.8540 | 0.6424 |
|
72 |
+
| 0.4745 | 59.0001 | 4020 | 3.9219 | 0.8692 | 0.6628 |
|
73 |
+
| 0.4204 | 62.0001 | 4221 | 3.8606 | 0.8709 | 0.6657 |
|
74 |
+
| 0.3926 | 65.0000 | 4422 | 4.2133 | 0.8666 | 0.6703 |
|
75 |
+
| 0.3569 | 67.0010 | 4623 | 4.5565 | 0.8644 | 0.6744 |
|
76 |
+
| 0.3342 | 70.0010 | 4824 | 4.9640 | 0.8809 | 0.6919 |
|
77 |
+
| 0.303 | 73.0009 | 5025 | 5.1881 | 0.8813 | 0.7000 |
|
78 |
+
| 0.2858 | 76.0009 | 5226 | 5.0722 | 0.8878 | 0.7098 |
|
79 |
+
| 0.2741 | 79.0008 | 5427 | 6.1618 | 0.8920 | 0.7198 |
|
80 |
+
| 0.2592 | 82.0008 | 5628 | 6.5562 | 0.8897 | 0.7178 |
|
81 |
+
| 0.2432 | 85.0007 | 5829 | 7.2766 | 0.8939 | 0.7291 |
|
82 |
+
| 0.2265 | 88.0007 | 6030 | 8.0923 | 0.8949 | 0.7300 |
|
83 |
+
| 0.2201 | 91.0006 | 6231 | 7.3557 | 0.8984 | 0.7339 |
|
84 |
+
| 0.2093 | 94.0006 | 6432 | 7.3704 | 0.8954 | 0.7321 |
|
85 |
+
| 0.1983 | 97.0005 | 6633 | 8.7920 | 0.8962 | 0.7394 |
|
86 |
+
| 0.195 | 100.0005 | 6834 | 8.1665 | 0.8975 | 0.7421 |
|
87 |
+
| 0.1828 | 103.0005 | 7035 | 7.8452 | 0.9033 | 0.7492 |
|
88 |
+
| 0.18 | 106.0004 | 7236 | 7.6154 | 0.9012 | 0.7537 |
|
89 |
+
| 0.1723 | 109.0004 | 7437 | 6.4922 | 0.9060 | 0.7585 |
|
90 |
+
| 0.1694 | 112.0003 | 7638 | 7.5438 | 0.9040 | 0.7616 |
|
91 |
+
| 0.1672 | 115.0003 | 7839 | 6.9874 | 0.9053 | 0.7650 |
|
92 |
+
| 0.1552 | 118.0002 | 8040 | 7.1907 | 0.9063 | 0.7634 |
|
93 |
+
| 0.1588 | 121.0002 | 8241 | 7.0263 | 0.9086 | 0.7636 |
|
94 |
+
| 0.1509 | 124.0001 | 8442 | 6.8458 | 0.9070 | 0.7644 |
|
95 |
+
| 0.1514 | 127.0001 | 8643 | 5.4099 | 0.9089 | 0.7701 |
|
96 |
+
| 0.1457 | 130.0001 | 8844 | 5.8632 | 0.9107 | 0.7727 |
|
97 |
+
| 0.1425 | 133.0000 | 9045 | 5.6742 | 0.9111 | 0.7753 |
|
98 |
+
| 0.1415 | 135.0010 | 9246 | 5.7430 | 0.9117 | 0.7761 |
|
99 |
+
| 0.1436 | 138.0009 | 9447 | 5.0272 | 0.9103 | 0.7734 |
|
100 |
+
| 0.1434 | 141.0009 | 9648 | 4.4936 | 0.9096 | 0.7746 |
|
101 |
+
| 0.1377 | 144.0008 | 9849 | 5.2322 | 0.9078 | 0.7769 |
|
102 |
+
| 0.1383 | 147.0008 | 10050 | 4.7013 | 0.9149 | 0.7835 |
|
103 |
+
| 0.133 | 150.0008 | 10251 | 4.6257 | 0.9142 | 0.7836 |
|
104 |
+
| 0.1349 | 153.0007 | 10452 | 3.9570 | 0.9153 | 0.7830 |
|
105 |
+
| 0.1273 | 156.0007 | 10653 | 3.6838 | 0.9142 | 0.7806 |
|
106 |
+
| 0.1255 | 159.0006 | 10854 | 3.6121 | 0.9144 | 0.7836 |
|
107 |
+
| 0.1247 | 162.0006 | 11055 | 3.2919 | 0.9160 | 0.7879 |
|
108 |
+
| 0.129 | 165.0005 | 11256 | 3.1283 | 0.9132 | 0.7823 |
|
109 |
+
| 0.1215 | 168.0005 | 11457 | 3.2925 | 0.9152 | 0.7854 |
|
110 |
+
| 0.1216 | 171.0004 | 11658 | 3.1251 | 0.9168 | 0.7869 |
|
111 |
+
| 0.1205 | 174.0004 | 11859 | 3.1548 | 0.9182 | 0.7893 |
|
112 |
+
| 0.1231 | 177.0004 | 12060 | 2.7044 | 0.9122 | 0.7835 |
|
113 |
+
| 0.1308 | 180.0003 | 12261 | 2.7894 | 0.9158 | 0.7891 |
|
114 |
+
| 0.1254 | 183.0003 | 12462 | 2.6146 | 0.9158 | 0.7864 |
|
115 |
+
| 0.1194 | 186.0002 | 12663 | 2.9000 | 0.9179 | 0.7899 |
|
116 |
+
| 0.1171 | 189.0002 | 12864 | 2.7811 | 0.9157 | 0.7899 |
|
117 |
+
| 0.1159 | 192.0001 | 13065 | 2.6063 | 0.9180 | 0.7926 |
|
118 |
+
| 0.1164 | 195.0001 | 13266 | 2.7002 | 0.9180 | 0.7915 |
|
119 |
+
| 0.1148 | 198.0000 | 13467 | 2.5541 | 0.9192 | 0.7946 |
|
120 |
+
| 0.1172 | 200.0010 | 13668 | 2.3960 | 0.9173 | 0.7894 |
|
121 |
+
| 0.1252 | 203.0010 | 13869 | 2.2429 | 0.9164 | 0.7887 |
|
122 |
+
| 0.1122 | 206.0009 | 14070 | 2.1204 | 0.9187 | 0.7935 |
|
123 |
+
| 0.115 | 209.0009 | 14271 | 2.3396 | 0.9190 | 0.7961 |
|
124 |
+
| 0.108 | 212.0008 | 14472 | 2.0145 | 0.9198 | 0.7957 |
|
125 |
+
| 0.1081 | 215.0008 | 14673 | 2.1137 | 0.9202 | 0.7969 |
|
126 |
+
| 0.1093 | 218.0007 | 14874 | 2.0863 | 0.9205 | 0.7972 |
|
127 |
+
| 0.1073 | 221.0007 | 15075 | 1.8909 | 0.9196 | 0.7946 |
|
128 |
+
| 0.11 | 224.0007 | 15276 | 1.8784 | 0.9212 | 0.7977 |
|
129 |
+
| 0.1087 | 227.0006 | 15477 | 1.8960 | 0.9224 | 0.8016 |
|
130 |
+
| 0.1065 | 230.0006 | 15678 | 1.8995 | 0.9186 | 0.7946 |
|
131 |
+
| 0.1065 | 233.0005 | 15879 | 1.8899 | 0.9200 | 0.7968 |
|
132 |
+
| 0.1108 | 236.0005 | 16080 | 1.8815 | 0.9205 | 0.7982 |
|
133 |
+
| 0.1064 | 239.0004 | 16281 | 1.8124 | 0.9220 | 0.7989 |
|
134 |
+
| 0.1038 | 242.0004 | 16482 | 1.8478 | 0.9235 | 0.8016 |
|
135 |
+
| 0.1047 | 245.0003 | 16683 | 1.8597 | 0.9218 | 0.8015 |
|
136 |
+
| 0.1042 | 248.0003 | 16884 | 1.8254 | 0.9221 | 0.8011 |
|
137 |
+
| 0.1064 | 251.0003 | 17085 | 1.7683 | 0.9222 | 0.8005 |
|
138 |
+
| 0.1007 | 254.0002 | 17286 | 1.7810 | 0.9216 | 0.8044 |
|
139 |
+
| 0.1009 | 257.0002 | 17487 | 1.8142 | 0.9225 | 0.8025 |
|
140 |
+
| 0.1011 | 260.0001 | 17688 | 1.7981 | 0.9203 | 0.7992 |
|
141 |
+
| 0.0963 | 263.0001 | 17889 | 1.6290 | 0.9222 | 0.8056 |
|
142 |
+
| 0.1089 | 266.0000 | 18090 | 1.6993 | 0.9214 | 0.8007 |
|
143 |
+
| 0.1119 | 268.0010 | 18291 | 1.6256 | 0.9214 | 0.8022 |
|
144 |
+
| 0.1008 | 271.0010 | 18492 | 1.7093 | 0.9232 | 0.8026 |
|
145 |
+
| 0.0963 | 274.0009 | 18693 | 1.6050 | 0.9234 | 0.8048 |
|
146 |
+
| 0.0997 | 277.0009 | 18894 | 1.5403 | 0.9237 | 0.8080 |
|
147 |
+
| 0.0978 | 280.0008 | 19095 | 1.6604 | 0.9229 | 0.8034 |
|
148 |
+
| 0.0964 | 283.0008 | 19296 | 1.5679 | 0.9232 | 0.8071 |
|
149 |
+
| 0.0963 | 286.0007 | 19497 | 1.5459 | 0.9230 | 0.8048 |
|
150 |
+
| 0.0973 | 289.0007 | 19698 | 1.4886 | 0.9196 | 0.8025 |
|
151 |
+
| 0.0996 | 292.0006 | 19899 | 1.4917 | 0.9208 | 0.8024 |
|
152 |
+
| 0.0963 | 295.0006 | 20100 | 1.3997 | 0.9225 | 0.8051 |
|
153 |
+
| 0.095 | 298.0005 | 20301 | 1.4833 | 0.9228 | 0.8050 |
|
154 |
+
| 0.0955 | 301.0005 | 20502 | 1.5128 | 0.9217 | 0.8033 |
|
155 |
+
| 0.0969 | 304.0005 | 20703 | 1.4938 | 0.9208 | 0.8025 |
|
156 |
+
| 0.1025 | 307.0004 | 20904 | 1.5228 | 0.9236 | 0.8031 |
|
157 |
+
| 0.096 | 310.0004 | 21105 | 1.5700 | 0.9241 | 0.8084 |
|
158 |
+
| 0.0981 | 313.0003 | 21306 | 1.4501 | 0.9202 | 0.8003 |
|
159 |
+
| 0.0968 | 316.0003 | 21507 | 1.5587 | 0.9238 | 0.8046 |
|
160 |
+
| 0.0958 | 319.0002 | 21708 | 1.4369 | 0.9254 | 0.8081 |
|
161 |
+
| 0.091 | 322.0002 | 21909 | 1.4935 | 0.9251 | 0.8074 |
|
162 |
+
| 0.0945 | 325.0001 | 22110 | 1.4459 | 0.9269 | 0.8120 |
|
163 |
+
| 0.0923 | 328.0001 | 22311 | 1.3652 | 0.9259 | 0.8109 |
|
164 |
+
| 0.0916 | 331.0001 | 22512 | 1.5116 | 0.9260 | 0.8099 |
|
165 |
+
| 0.091 | 334.0000 | 22713 | 1.4167 | 0.9245 | 0.8084 |
|
166 |
+
| 0.0924 | 336.0010 | 22914 | 1.4038 | 0.9258 | 0.8095 |
|
167 |
+
| 0.0917 | 339.0009 | 23115 | 1.4282 | 0.9258 | 0.8103 |
|
168 |
+
| 0.0892 | 342.0009 | 23316 | 1.4295 | 0.9267 | 0.8108 |
|
169 |
+
| 0.0898 | 345.0008 | 23517 | 1.4366 | 0.9257 | 0.8137 |
|
170 |
+
| 0.0916 | 348.0008 | 23718 | 1.4560 | 0.9234 | 0.8075 |
|
171 |
+
| 0.088 | 351.0008 | 23919 | 1.3876 | 0.9258 | 0.8118 |
|
172 |
+
| 0.0896 | 354.0007 | 24120 | 1.5172 | 0.9227 | 0.8065 |
|
173 |
+
| 0.088 | 357.0007 | 24321 | 1.5016 | 0.9261 | 0.8097 |
|
174 |
+
| 0.0874 | 360.0006 | 24522 | 1.4182 | 0.9254 | 0.8089 |
|
175 |
+
| 0.0918 | 363.0006 | 24723 | 1.4840 | 0.9241 | 0.8045 |
|
176 |
+
| 0.09 | 366.0005 | 24924 | 1.5154 | 0.9233 | 0.8081 |
|
177 |
+
| 0.0888 | 369.0005 | 25125 | 1.3604 | 0.9277 | 0.8143 |
|
178 |
+
| 0.0866 | 372.0004 | 25326 | 1.4135 | 0.9240 | 0.8081 |
|
179 |
+
| 0.089 | 375.0004 | 25527 | 1.4785 | 0.9260 | 0.8117 |
|
180 |
+
| 0.0928 | 378.0004 | 25728 | 1.3908 | 0.9256 | 0.8094 |
|
181 |
+
| 0.0927 | 381.0003 | 25929 | 1.4106 | 0.9246 | 0.8058 |
|
182 |
+
| 0.0877 | 384.0003 | 26130 | 1.3791 | 0.9263 | 0.8111 |
|
183 |
+
| 0.0859 | 387.0002 | 26331 | 1.4158 | 0.9256 | 0.8112 |
|
184 |
+
| 0.0878 | 390.0002 | 26532 | 1.3359 | 0.9256 | 0.8111 |
|
185 |
+
| 0.085 | 393.0001 | 26733 | 1.3835 | 0.9249 | 0.8096 |
|
186 |
+
| 0.0863 | 396.0001 | 26934 | 1.3753 | 0.9255 | 0.8109 |
|
187 |
+
| 0.0854 | 399.0000 | 27135 | 1.3575 | 0.9266 | 0.8108 |
|
188 |
+
| 0.0856 | 401.0010 | 27336 | 1.3802 | 0.9270 | 0.8122 |
|
189 |
+
| 0.0856 | 404.0010 | 27537 | 1.3584 | 0.9278 | 0.8138 |
|
190 |
+
| 0.0868 | 407.0009 | 27738 | 1.4511 | 0.9240 | 0.8079 |
|
191 |
+
| 0.0866 | 410.0009 | 27939 | 1.3532 | 0.9270 | 0.8125 |
|
192 |
+
| 0.0843 | 413.0008 | 28140 | 1.3818 | 0.9264 | 0.8134 |
|
193 |
+
| 0.086 | 416.0008 | 28341 | 1.3797 | 0.9262 | 0.8131 |
|
194 |
+
| 0.0849 | 419.0007 | 28542 | 1.3713 | 0.9267 | 0.8149 |
|
195 |
+
| 0.0831 | 422.0007 | 28743 | 1.4248 | 0.9260 | 0.8123 |
|
196 |
+
| 0.0843 | 425.0007 | 28944 | 1.3671 | 0.9272 | 0.8133 |
|
197 |
+
| 0.0838 | 428.0006 | 29145 | 1.5038 | 0.9260 | 0.8124 |
|
198 |
+
| 0.0839 | 431.0006 | 29346 | 1.4157 | 0.9272 | 0.8136 |
|
199 |
+
| 0.0827 | 434.0005 | 29547 | 1.3950 | 0.9269 | 0.8137 |
|
200 |
+
| 0.0832 | 437.0005 | 29748 | 1.3595 | 0.9267 | 0.8128 |
|
201 |
+
| 0.0867 | 440.0004 | 29949 | 1.2980 | 0.9243 | 0.8134 |
|
202 |
+
| 0.0846 | 443.0004 | 30150 | 1.4668 | 0.9279 | 0.8149 |
|
203 |
+
| 0.0853 | 446.0003 | 30351 | 1.3490 | 0.9259 | 0.8133 |
|
204 |
+
| 0.0849 | 449.0003 | 30552 | 1.4092 | 0.9275 | 0.8148 |
|
205 |
+
| 0.0856 | 452.0003 | 30753 | 1.3654 | 0.9246 | 0.8077 |
|
206 |
+
| 0.0838 | 455.0002 | 30954 | 1.4238 | 0.9269 | 0.8115 |
|
207 |
+
| 0.0809 | 458.0002 | 31155 | 1.3319 | 0.9273 | 0.8117 |
|
208 |
+
| 0.0826 | 461.0001 | 31356 | 1.4123 | 0.9276 | 0.8129 |
|
209 |
+
| 0.0815 | 464.0001 | 31557 | 1.3352 | 0.9275 | 0.8136 |
|
210 |
+
| 0.0804 | 467.0000 | 31758 | 1.3803 | 0.9280 | 0.8159 |
|
211 |
+
| 0.0797 | 469.0010 | 31959 | 1.3948 | 0.9272 | 0.8173 |
|
212 |
+
| 0.0806 | 472.0010 | 32160 | 1.3945 | 0.9261 | 0.8104 |
|
213 |
+
| 0.0804 | 475.0009 | 32361 | 1.3707 | 0.9284 | 0.8156 |
|
214 |
+
| 0.0808 | 478.0009 | 32562 | 1.4406 | 0.9268 | 0.8152 |
|
215 |
+
| 0.0792 | 481.0008 | 32763 | 1.4056 | 0.9267 | 0.8144 |
|
216 |
+
| 0.0788 | 484.0008 | 32964 | 1.4086 | 0.9269 | 0.8145 |
|
217 |
+
| 0.0805 | 487.0007 | 33165 | 1.3563 | 0.9276 | 0.8168 |
|
218 |
+
| 0.0804 | 490.0007 | 33366 | 1.3429 | 0.9284 | 0.8149 |
|
219 |
+
| 0.0798 | 493.0006 | 33567 | 1.4408 | 0.9270 | 0.8143 |
|
220 |
+
| 0.0791 | 496.0006 | 33768 | 1.4209 | 0.9274 | 0.8152 |
|
221 |
+
| 0.0823 | 499.0005 | 33969 | 1.4354 | 0.9271 | 0.8144 |
|
222 |
+
| 0.0811 | 502.0005 | 34170 | 1.3956 | 0.9265 | 0.8146 |
|
223 |
+
| 0.0978 | 505.0005 | 34371 | 1.4652 | 0.9256 | 0.8102 |
|
224 |
+
| 0.0894 | 508.0004 | 34572 | 1.3084 | 0.9237 | 0.8135 |
|
225 |
+
| 0.0817 | 511.0004 | 34773 | 1.3392 | 0.9272 | 0.8169 |
|
226 |
+
| 0.0793 | 514.0003 | 34974 | 1.3396 | 0.9294 | 0.8198 |
|
227 |
+
| 0.0781 | 517.0003 | 35175 | 1.2850 | 0.9289 | 0.8204 |
|
228 |
+
| 0.0787 | 520.0002 | 35376 | 1.3228 | 0.9294 | 0.8204 |
|
229 |
+
| 0.0778 | 523.0002 | 35577 | 1.3750 | 0.9289 | 0.8172 |
|
230 |
+
| 0.0772 | 526.0001 | 35778 | 1.4147 | 0.9274 | 0.8143 |
|
231 |
+
| 0.0777 | 529.0001 | 35979 | 1.3052 | 0.9278 | 0.8170 |
|
232 |
+
| 0.0783 | 532.0001 | 36180 | 1.2464 | 0.9286 | 0.8198 |
|
233 |
+
| 0.077 | 535.0000 | 36381 | 1.3571 | 0.9283 | 0.8184 |
|
234 |
+
| 0.0778 | 537.0010 | 36582 | 1.3503 | 0.9290 | 0.8184 |
|
235 |
+
| 0.0787 | 540.0009 | 36783 | 1.3391 | 0.9286 | 0.8184 |
|
236 |
+
| 0.077 | 543.0009 | 36984 | 1.4027 | 0.9292 | 0.8170 |
|
237 |
+
| 0.0779 | 546.0008 | 37185 | 1.3167 | 0.9281 | 0.8165 |
|
238 |
+
| 0.0781 | 549.0008 | 37386 | 1.3828 | 0.9284 | 0.8147 |
|
239 |
+
| 0.0781 | 552.0008 | 37587 | 1.3284 | 0.9287 | 0.8196 |
|
240 |
+
| 0.0765 | 555.0007 | 37788 | 1.4137 | 0.9286 | 0.8186 |
|
241 |
+
| 0.0778 | 558.0007 | 37989 | 1.4190 | 0.9286 | 0.8174 |
|
242 |
+
| 0.0747 | 561.0006 | 38190 | 1.3886 | 0.9289 | 0.8159 |
|
243 |
+
| 0.0768 | 564.0006 | 38391 | 1.3433 | 0.9291 | 0.8178 |
|
244 |
+
| 0.0765 | 567.0005 | 38592 | 1.4049 | 0.9289 | 0.8169 |
|
245 |
+
| 0.0781 | 570.0005 | 38793 | 1.4670 | 0.9276 | 0.8141 |
|
246 |
+
| 0.0776 | 573.0004 | 38994 | 1.5075 | 0.9280 | 0.8166 |
|
247 |
+
| 0.0753 | 576.0004 | 39195 | 1.3683 | 0.9280 | 0.8164 |
|
248 |
+
| 0.0756 | 579.0004 | 39396 | 1.3317 | 0.9281 | 0.8186 |
|
249 |
+
|
250 |
+
|
251 |
+
### Framework versions
|
252 |
+
|
253 |
+
- Transformers 4.46.0
|
254 |
+
- Pytorch 2.3.1+cu121
|
255 |
+
- Datasets 2.20.0
|
256 |
+
- Tokenizers 0.20.1
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 126048336
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e0ff83760b5c6dcc42dd14851ff136a26014c60b60907f48ff84623701a6fdab
|
3 |
size 126048336
|
runs/num_train_epochs=1000.0-tssp_mode=0-drp=0.0-trp=0.0-v6/events.out.tfevents.1738986472.dipa.3930354.1
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b23f3540bc74b8cf5e4a60367b666e35d2880f9e0ac2365c57d7c211f73e96a2
|
3 |
+
size 470
|