izaitova commited on
Commit
bbc34bc
·
verified ·
1 Parent(s): 7878f3a

End of training

Browse files
Files changed (2) hide show
  1. README.md +28 -15
  2. pytorch_model.bin +1 -1
README.md CHANGED
@@ -1,4 +1,5 @@
1
  ---
 
2
  base_model: ai-forever/ruBert-large
3
  tags:
4
  - generated_from_trainer
@@ -24,16 +25,16 @@ model-index:
24
  metrics:
25
  - name: Precision
26
  type: precision
27
- value: 0.7868756986512067
28
  - name: Recall
29
  type: recall
30
- value: 0.7638984758119328
31
  - name: F1
32
  type: f1
33
- value: 0.772014566268761
34
  - name: Accuracy
35
  type: accuracy
36
- value: 0.8826334008229377
37
  ---
38
 
39
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -43,11 +44,11 @@ should probably proofread and complete it, then remove this comment. -->
43
 
44
  This model is a fine-tuned version of [ai-forever/ruBert-large](https://huggingface.co/ai-forever/ruBert-large) on the universal_dependencies dataset.
45
  It achieves the following results on the evaluation set:
46
- - Loss: 0.8204
47
- - Precision: 0.7869
48
- - Recall: 0.7639
49
- - F1: 0.7720
50
- - Accuracy: 0.8826
51
 
52
  ## Model description
53
 
@@ -67,20 +68,32 @@ More information needed
67
 
68
  The following hyperparameters were used during training:
69
  - learning_rate: 5e-05
70
- - train_batch_size: 32
71
  - eval_batch_size: 8
72
  - seed: 42
73
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
74
  - lr_scheduler_type: linear
75
- - num_epochs: 20
76
 
77
  ### Training results
78
 
 
 
 
 
 
 
 
 
 
 
 
 
79
 
80
 
81
  ### Framework versions
82
 
83
- - Transformers 4.39.3
84
- - Pytorch 1.11.0a0+17540c5
85
- - Datasets 2.21.0
86
- - Tokenizers 0.15.2
 
1
  ---
2
+ library_name: transformers
3
  base_model: ai-forever/ruBert-large
4
  tags:
5
  - generated_from_trainer
 
25
  metrics:
26
  - name: Precision
27
  type: precision
28
+ value: 0.8307441967265208
29
  - name: Recall
30
  type: recall
31
+ value: 0.7502322735093846
32
  - name: F1
33
  type: f1
34
+ value: 0.783084706036028
35
  - name: Accuracy
36
  type: accuracy
37
+ value: 0.868562326706389
38
  ---
39
 
40
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
44
 
45
  This model is a fine-tuned version of [ai-forever/ruBert-large](https://huggingface.co/ai-forever/ruBert-large) on the universal_dependencies dataset.
46
  It achieves the following results on the evaluation set:
47
+ - Loss: 0.4344
48
+ - Precision: 0.8307
49
+ - Recall: 0.7502
50
+ - F1: 0.7831
51
+ - Accuracy: 0.8686
52
 
53
  ## Model description
54
 
 
68
 
69
  The following hyperparameters were used during training:
70
  - learning_rate: 5e-05
71
+ - train_batch_size: 16
72
  - eval_batch_size: 8
73
  - seed: 42
74
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
75
  - lr_scheduler_type: linear
76
+ - num_epochs: 10
77
 
78
  ### Training results
79
 
80
+ | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
81
+ |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
82
+ | No log | 1.0 | 338 | 0.4759 | 0.7967 | 0.7249 | 0.7532 | 0.8557 |
83
+ | No log | 2.0 | 676 | 0.4344 | 0.8307 | 0.7502 | 0.7831 | 0.8686 |
84
+ | No log | 3.0 | 1014 | 0.6906 | 0.7842 | 0.7480 | 0.7563 | 0.8674 |
85
+ | No log | 4.0 | 1352 | 0.4757 | 0.8185 | 0.7578 | 0.7777 | 0.8816 |
86
+ | No log | 5.0 | 1690 | 0.6291 | 0.7791 | 0.7721 | 0.7670 | 0.8792 |
87
+ | No log | 6.0 | 2028 | 0.6466 | 0.7967 | 0.7677 | 0.7721 | 0.8863 |
88
+ | No log | 7.0 | 2366 | 0.7072 | 0.7751 | 0.7700 | 0.7704 | 0.8809 |
89
+ | No log | 8.0 | 2704 | 0.7623 | 0.7957 | 0.7678 | 0.7749 | 0.8838 |
90
+ | No log | 9.0 | 3042 | 0.7458 | 0.7922 | 0.7716 | 0.7773 | 0.8873 |
91
+ | No log | 10.0 | 3380 | 0.7560 | 0.7916 | 0.7709 | 0.7767 | 0.8869 |
92
 
93
 
94
  ### Framework versions
95
 
96
+ - Transformers 4.44.2
97
+ - Pytorch 2.4.1+cu121
98
+ - Datasets 3.0.1
99
+ - Tokenizers 0.19.1
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d5e2c4e64f5be90393df2d9e256a2c421d3eb1c48ef09ab5e4ed157e0122b7e4
3
  size 1703644326
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0721adfb413fb80835d1ae25e61eea5d72246d0c82655c20a9ea7dac8475a14e
3
  size 1703644326