xuancoblab2023 commited on
Commit
e33a76f
·
verified ·
1 Parent(s): ee0c7c1

Training in progress, epoch 1

Browse files
Files changed (38) hide show
  1. logs/events.out.tfevents.1709748697.d65e8524b0f6.5185.21 +2 -2
  2. logs/events.out.tfevents.1709749139.d65e8524b0f6.5185.22 +3 -0
  3. model.safetensors +1 -1
  4. run-21/checkpoint-156/config.json +34 -0
  5. run-21/checkpoint-156/model.safetensors +3 -0
  6. run-21/checkpoint-156/optimizer.pt +3 -0
  7. run-21/checkpoint-156/rng_state.pth +3 -0
  8. run-21/checkpoint-156/scheduler.pt +3 -0
  9. run-21/checkpoint-156/special_tokens_map.json +7 -0
  10. run-21/checkpoint-156/tokenizer.json +0 -0
  11. run-21/checkpoint-156/tokenizer_config.json +57 -0
  12. run-21/checkpoint-156/trainer_state.json +84 -0
  13. run-21/checkpoint-156/training_args.bin +3 -0
  14. run-21/checkpoint-156/vocab.txt +0 -0
  15. run-21/checkpoint-208/config.json +34 -0
  16. run-21/checkpoint-208/model.safetensors +3 -0
  17. run-21/checkpoint-208/optimizer.pt +3 -0
  18. run-21/checkpoint-208/rng_state.pth +3 -0
  19. run-21/checkpoint-208/scheduler.pt +3 -0
  20. run-21/checkpoint-208/special_tokens_map.json +7 -0
  21. run-21/checkpoint-208/tokenizer.json +0 -0
  22. run-21/checkpoint-208/tokenizer_config.json +57 -0
  23. run-21/checkpoint-208/trainer_state.json +103 -0
  24. run-21/checkpoint-208/training_args.bin +3 -0
  25. run-21/checkpoint-208/vocab.txt +0 -0
  26. run-22/checkpoint-73/config.json +34 -0
  27. run-22/checkpoint-73/model.safetensors +3 -0
  28. run-22/checkpoint-73/optimizer.pt +3 -0
  29. run-22/checkpoint-73/rng_state.pth +3 -0
  30. run-22/checkpoint-73/scheduler.pt +3 -0
  31. run-22/checkpoint-73/special_tokens_map.json +7 -0
  32. run-22/checkpoint-73/tokenizer.json +0 -0
  33. run-22/checkpoint-73/tokenizer_config.json +57 -0
  34. run-22/checkpoint-73/trainer_state.json +46 -0
  35. run-22/checkpoint-73/training_args.bin +3 -0
  36. run-22/checkpoint-73/vocab.txt +0 -0
  37. tokenizer.json +1 -1
  38. training_args.bin +1 -1
logs/events.out.tfevents.1709748697.d65e8524b0f6.5185.21 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef590dead44772f2ee831d4bc41d013aa2689aa54b55765ef13016a1f6685cd9
3
- size 5985
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b326b8b922d402e11c9d8ccc78036e6cbe235eddea45aad933ef6fdbf25e6df
3
+ size 7705
logs/events.out.tfevents.1709749139.d65e8524b0f6.5185.22 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:243e0ee847490474fd07e9ec81e31416de0c6d1a89d6c910eb380c00b962a442
3
+ size 5313
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a3f76b8e9d1f67f9b433ce0b747eaefb4717964e4ff51b59ef03cffd276f3dbe
3
  size 17549312
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44c120b54fd669f33f415ab00eadf71e828c642fcb280379e4e68323698fc046
3
  size 17549312
run-21/checkpoint-156/config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/bert_uncased_L-2_H-128_A-2",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 128,
11
+ "id2label": {
12
+ "0": "negative",
13
+ "1": "positive"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 512,
17
+ "label2id": {
18
+ "negative": "0",
19
+ "positive": "1"
20
+ },
21
+ "layer_norm_eps": 1e-12,
22
+ "max_position_embeddings": 512,
23
+ "model_type": "bert",
24
+ "num_attention_heads": 2,
25
+ "num_hidden_layers": 2,
26
+ "pad_token_id": 0,
27
+ "position_embedding_type": "absolute",
28
+ "problem_type": "single_label_classification",
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.38.2",
31
+ "type_vocab_size": 2,
32
+ "use_cache": true,
33
+ "vocab_size": 30522
34
+ }
run-21/checkpoint-156/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8fb3e7a27c82807500b3ec68bc657387013cf10a996199e945b1f3b01c2795b
3
+ size 17549312
run-21/checkpoint-156/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1958680b9707519142c99a3618d9001f6c418cacd0b9605957ef05febb3cadf5
3
+ size 35122746
run-21/checkpoint-156/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f3f2bf3b854c592a6f1e0e90f14af49f8d742650373f022831e4b62d27ad7a2
3
+ size 14054
run-21/checkpoint-156/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccc1b9a873ffc927473096e4bcf6be58267e96371fc18313b8496d12cd4d5224
3
+ size 1064
run-21/checkpoint-156/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
run-21/checkpoint-156/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-21/checkpoint-156/tokenizer_config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
+ "never_split": null,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "strip_accents": null,
54
+ "tokenize_chinese_chars": true,
55
+ "tokenizer_class": "BertTokenizer",
56
+ "unk_token": "[UNK]"
57
+ }
run-21/checkpoint-156/trainer_state.json ADDED
@@ -0,0 +1,84 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.7504892367906066,
3
+ "best_model_checkpoint": "tiny-bert-sst2-distilled/run-21/checkpoint-104",
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 156,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "grad_norm": 2.939810276031494,
14
+ "learning_rate": 0.00013471680923458288,
15
+ "loss": 0.5033,
16
+ "step": 52
17
+ },
18
+ {
19
+ "epoch": 1.0,
20
+ "eval_accuracy": 0.7152641878669276,
21
+ "eval_f1": 0.712166172106825,
22
+ "eval_loss": 0.4479813575744629,
23
+ "eval_precision": 0.72,
24
+ "eval_recall": 0.7045009784735812,
25
+ "eval_runtime": 28.0693,
26
+ "eval_samples_per_second": 36.41,
27
+ "eval_steps_per_second": 1.14,
28
+ "step": 52
29
+ },
30
+ {
31
+ "epoch": 2.0,
32
+ "grad_norm": 8.902515411376953,
33
+ "learning_rate": 8.981120615638859e-05,
34
+ "loss": 0.4475,
35
+ "step": 104
36
+ },
37
+ {
38
+ "epoch": 2.0,
39
+ "eval_accuracy": 0.7504892367906066,
40
+ "eval_f1": 0.7578347578347578,
41
+ "eval_loss": 0.42175495624542236,
42
+ "eval_precision": 0.7361623616236163,
43
+ "eval_recall": 0.7808219178082192,
44
+ "eval_runtime": 27.9965,
45
+ "eval_samples_per_second": 36.505,
46
+ "eval_steps_per_second": 1.143,
47
+ "step": 104
48
+ },
49
+ {
50
+ "epoch": 3.0,
51
+ "grad_norm": 4.328202724456787,
52
+ "learning_rate": 4.4905603078194294e-05,
53
+ "loss": 0.4233,
54
+ "step": 156
55
+ },
56
+ {
57
+ "epoch": 3.0,
58
+ "eval_accuracy": 0.7201565557729941,
59
+ "eval_f1": 0.6857142857142856,
60
+ "eval_loss": 0.415389746427536,
61
+ "eval_precision": 0.7819548872180451,
62
+ "eval_recall": 0.6105675146771037,
63
+ "eval_runtime": 28.1971,
64
+ "eval_samples_per_second": 36.245,
65
+ "eval_steps_per_second": 1.135,
66
+ "step": 156
67
+ }
68
+ ],
69
+ "logging_steps": 500,
70
+ "max_steps": 208,
71
+ "num_input_tokens_seen": 0,
72
+ "num_train_epochs": 4,
73
+ "save_steps": 500,
74
+ "total_flos": 707085591840.0,
75
+ "train_batch_size": 60,
76
+ "trial_name": null,
77
+ "trial_params": {
78
+ "alpha": 0.7020704368642945,
79
+ "learning_rate": 0.00017962241231277718,
80
+ "num_train_epochs": 4,
81
+ "per_device_train_batch_size": 60,
82
+ "temperature": 19
83
+ }
84
+ }
run-21/checkpoint-156/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fe829315cd1ce1b368ca5aa1539a6e655572843d060f6af981de772385e9701
3
+ size 4920
run-21/checkpoint-156/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
run-21/checkpoint-208/config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/bert_uncased_L-2_H-128_A-2",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 128,
11
+ "id2label": {
12
+ "0": "negative",
13
+ "1": "positive"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 512,
17
+ "label2id": {
18
+ "negative": "0",
19
+ "positive": "1"
20
+ },
21
+ "layer_norm_eps": 1e-12,
22
+ "max_position_embeddings": 512,
23
+ "model_type": "bert",
24
+ "num_attention_heads": 2,
25
+ "num_hidden_layers": 2,
26
+ "pad_token_id": 0,
27
+ "position_embedding_type": "absolute",
28
+ "problem_type": "single_label_classification",
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.38.2",
31
+ "type_vocab_size": 2,
32
+ "use_cache": true,
33
+ "vocab_size": 30522
34
+ }
run-21/checkpoint-208/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa8c12b34db4caa4aa05168b916ae65ef07eec0279de7abd53e115744a573909
3
+ size 17549312
run-21/checkpoint-208/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:707810b82af2863c354b43e11120fe875bdde16840fa76be0d6f6e96e00df9f6
3
+ size 35122746
run-21/checkpoint-208/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b6ed2075102dd75f9c18fe89ccd178319808314220b3c31c67cb30308a85909
3
+ size 14054
run-21/checkpoint-208/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ff997d3a0b6bdcde1a920243f3b68a38d81cbfb189c7690b7052ae0d3a22aa2
3
+ size 1064
run-21/checkpoint-208/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
run-21/checkpoint-208/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-21/checkpoint-208/tokenizer_config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
+ "never_split": null,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "strip_accents": null,
54
+ "tokenize_chinese_chars": true,
55
+ "tokenizer_class": "BertTokenizer",
56
+ "unk_token": "[UNK]"
57
+ }
run-21/checkpoint-208/trainer_state.json ADDED
@@ -0,0 +1,103 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.7906066536203522,
3
+ "best_model_checkpoint": "tiny-bert-sst2-distilled/run-21/checkpoint-208",
4
+ "epoch": 4.0,
5
+ "eval_steps": 500,
6
+ "global_step": 208,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "grad_norm": 2.939810276031494,
14
+ "learning_rate": 0.00013471680923458288,
15
+ "loss": 0.5033,
16
+ "step": 52
17
+ },
18
+ {
19
+ "epoch": 1.0,
20
+ "eval_accuracy": 0.7152641878669276,
21
+ "eval_f1": 0.712166172106825,
22
+ "eval_loss": 0.4479813575744629,
23
+ "eval_precision": 0.72,
24
+ "eval_recall": 0.7045009784735812,
25
+ "eval_runtime": 28.0693,
26
+ "eval_samples_per_second": 36.41,
27
+ "eval_steps_per_second": 1.14,
28
+ "step": 52
29
+ },
30
+ {
31
+ "epoch": 2.0,
32
+ "grad_norm": 8.902515411376953,
33
+ "learning_rate": 8.981120615638859e-05,
34
+ "loss": 0.4475,
35
+ "step": 104
36
+ },
37
+ {
38
+ "epoch": 2.0,
39
+ "eval_accuracy": 0.7504892367906066,
40
+ "eval_f1": 0.7578347578347578,
41
+ "eval_loss": 0.42175495624542236,
42
+ "eval_precision": 0.7361623616236163,
43
+ "eval_recall": 0.7808219178082192,
44
+ "eval_runtime": 27.9965,
45
+ "eval_samples_per_second": 36.505,
46
+ "eval_steps_per_second": 1.143,
47
+ "step": 104
48
+ },
49
+ {
50
+ "epoch": 3.0,
51
+ "grad_norm": 4.328202724456787,
52
+ "learning_rate": 4.4905603078194294e-05,
53
+ "loss": 0.4233,
54
+ "step": 156
55
+ },
56
+ {
57
+ "epoch": 3.0,
58
+ "eval_accuracy": 0.7201565557729941,
59
+ "eval_f1": 0.6857142857142856,
60
+ "eval_loss": 0.415389746427536,
61
+ "eval_precision": 0.7819548872180451,
62
+ "eval_recall": 0.6105675146771037,
63
+ "eval_runtime": 28.1971,
64
+ "eval_samples_per_second": 36.245,
65
+ "eval_steps_per_second": 1.135,
66
+ "step": 156
67
+ },
68
+ {
69
+ "epoch": 4.0,
70
+ "grad_norm": 6.079658508300781,
71
+ "learning_rate": 0.0,
72
+ "loss": 0.4193,
73
+ "step": 208
74
+ },
75
+ {
76
+ "epoch": 4.0,
77
+ "eval_accuracy": 0.7906066536203522,
78
+ "eval_f1": 0.8007448789571694,
79
+ "eval_loss": 0.4070701003074646,
80
+ "eval_precision": 0.7637655417406749,
81
+ "eval_recall": 0.8414872798434442,
82
+ "eval_runtime": 27.9567,
83
+ "eval_samples_per_second": 36.557,
84
+ "eval_steps_per_second": 1.145,
85
+ "step": 208
86
+ }
87
+ ],
88
+ "logging_steps": 500,
89
+ "max_steps": 208,
90
+ "num_input_tokens_seen": 0,
91
+ "num_train_epochs": 4,
92
+ "save_steps": 500,
93
+ "total_flos": 942780789120.0,
94
+ "train_batch_size": 60,
95
+ "trial_name": null,
96
+ "trial_params": {
97
+ "alpha": 0.7020704368642945,
98
+ "learning_rate": 0.00017962241231277718,
99
+ "num_train_epochs": 4,
100
+ "per_device_train_batch_size": 60,
101
+ "temperature": 19
102
+ }
103
+ }
run-21/checkpoint-208/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fe829315cd1ce1b368ca5aa1539a6e655572843d060f6af981de772385e9701
3
+ size 4920
run-21/checkpoint-208/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
run-22/checkpoint-73/config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/bert_uncased_L-2_H-128_A-2",
3
+ "architectures": [
4
+ "BertForSequenceClassification"
5
+ ],
6
+ "attention_probs_dropout_prob": 0.1,
7
+ "classifier_dropout": null,
8
+ "hidden_act": "gelu",
9
+ "hidden_dropout_prob": 0.1,
10
+ "hidden_size": 128,
11
+ "id2label": {
12
+ "0": "negative",
13
+ "1": "positive"
14
+ },
15
+ "initializer_range": 0.02,
16
+ "intermediate_size": 512,
17
+ "label2id": {
18
+ "negative": "0",
19
+ "positive": "1"
20
+ },
21
+ "layer_norm_eps": 1e-12,
22
+ "max_position_embeddings": 512,
23
+ "model_type": "bert",
24
+ "num_attention_heads": 2,
25
+ "num_hidden_layers": 2,
26
+ "pad_token_id": 0,
27
+ "position_embedding_type": "absolute",
28
+ "problem_type": "single_label_classification",
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.38.2",
31
+ "type_vocab_size": 2,
32
+ "use_cache": true,
33
+ "vocab_size": 30522
34
+ }
run-22/checkpoint-73/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44c120b54fd669f33f415ab00eadf71e828c642fcb280379e4e68323698fc046
3
+ size 17549312
run-22/checkpoint-73/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:611ef02219b5048bf7d76ef651f30f7843251339392e4c484c0e4a579cf0d056
3
+ size 35122746
run-22/checkpoint-73/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66c42846858a5e0d6a622704e299ce8fbbbb35d1bc8a006bfa388d93deb44fa1
3
+ size 14054
run-22/checkpoint-73/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e2553c8ee13b391e83f4e8036597f099a9b8867f2315b9a912ebba86609070a
3
+ size 1064
run-22/checkpoint-73/special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
run-22/checkpoint-73/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
run-22/checkpoint-73/tokenizer_config.json ADDED
@@ -0,0 +1,57 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "clean_up_tokenization_spaces": true,
45
+ "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": true,
48
+ "mask_token": "[MASK]",
49
+ "model_max_length": 512,
50
+ "never_split": null,
51
+ "pad_token": "[PAD]",
52
+ "sep_token": "[SEP]",
53
+ "strip_accents": null,
54
+ "tokenize_chinese_chars": true,
55
+ "tokenizer_class": "BertTokenizer",
56
+ "unk_token": "[UNK]"
57
+ }
run-22/checkpoint-73/trainer_state.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.538160469667319,
3
+ "best_model_checkpoint": "tiny-bert-sst2-distilled/run-22/checkpoint-73",
4
+ "epoch": 1.0,
5
+ "eval_steps": 500,
6
+ "global_step": 73,
7
+ "is_hyper_param_search": true,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.0,
13
+ "grad_norm": 0.7895054817199707,
14
+ "learning_rate": 4.53929460427656e-05,
15
+ "loss": 0.592,
16
+ "step": 73
17
+ },
18
+ {
19
+ "epoch": 1.0,
20
+ "eval_accuracy": 0.538160469667319,
21
+ "eval_f1": 0.1945392491467577,
22
+ "eval_loss": 0.5740381479263306,
23
+ "eval_precision": 0.76,
24
+ "eval_recall": 0.11154598825831702,
25
+ "eval_runtime": 28.0661,
26
+ "eval_samples_per_second": 36.414,
27
+ "eval_steps_per_second": 1.14,
28
+ "step": 73
29
+ }
30
+ ],
31
+ "logging_steps": 500,
32
+ "max_steps": 219,
33
+ "num_input_tokens_seen": 0,
34
+ "num_train_epochs": 3,
35
+ "save_steps": 500,
36
+ "total_flos": 235695197280.0,
37
+ "train_batch_size": 42,
38
+ "trial_name": null,
39
+ "trial_params": {
40
+ "alpha": 0.8218865571627406,
41
+ "learning_rate": 6.80894190641484e-05,
42
+ "num_train_epochs": 3,
43
+ "per_device_train_batch_size": 42,
44
+ "temperature": 18
45
+ }
46
+ }
run-22/checkpoint-73/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af01c8cc931e90be8609f6f102e02c380f096eaceeec5ba79858d8a3dfed157c
3
+ size 4920
run-22/checkpoint-73/vocab.txt ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.json CHANGED
@@ -2,7 +2,7 @@
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
- "max_length": 33,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
 
2
  "version": "1.0",
3
  "truncation": {
4
  "direction": "Right",
5
+ "max_length": 31,
6
  "strategy": "LongestFirst",
7
  "stride": 0
8
  },
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:32f042525736aa17960ee8344101391e82f28237778df28fde71ca14364b664f
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af01c8cc931e90be8609f6f102e02c380f096eaceeec5ba79858d8a3dfed157c
3
  size 4920