bonurtek commited on
Commit
7b00647
·
verified ·
1 Parent(s): 82239ef

End of training

Browse files
README.md CHANGED
@@ -11,8 +11,6 @@ metrics:
11
  model-index:
12
  - name: flan-t5-small-hallucination-text-classification
13
  results: []
14
- datasets:
15
- - cemuluoglakci/hallucination_acceptance_agent_instruction_dataset
16
  ---
17
 
18
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -20,13 +18,13 @@ should probably proofread and complete it, then remove this comment. -->
20
 
21
  # flan-t5-small-hallucination-text-classification
22
 
23
- This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on the Hallucination Acceptance Agent Instruction dataset.
24
  It achieves the following results on the evaluation set:
25
- - Loss: 0.6364
26
- - Precision: 0.7374
27
- - Recall: 0.7420
28
- - F1: 0.7370
29
- - Accuracy: 0.7420
30
 
31
  ## Model description
32
 
@@ -46,26 +44,18 @@ More information needed
46
 
47
  The following hyperparameters were used during training:
48
  - learning_rate: 0.0003
49
- - train_batch_size: 8
50
- - eval_batch_size: 8
51
  - seed: 42
52
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
53
  - lr_scheduler_type: linear
54
- - num_epochs: 2
55
 
56
  ### Training results
57
 
58
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
59
  |:-------------:|:------:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
60
- | 0.9155 | 0.2008 | 100 | 0.8142 | 0.6260 | 0.6365 | 0.6213 | 0.6365 |
61
- | 0.777 | 0.4016 | 200 | 0.7090 | 0.7062 | 0.7098 | 0.7003 | 0.7098 |
62
- | 0.7625 | 0.6024 | 300 | 0.6595 | 0.7365 | 0.7299 | 0.7212 | 0.7299 |
63
- | 0.7136 | 0.8032 | 400 | 0.7185 | 0.7084 | 0.7108 | 0.7090 | 0.7108 |
64
- | 0.6807 | 1.0040 | 500 | 0.7207 | 0.7216 | 0.7008 | 0.7066 | 0.7008 |
65
- | 0.6522 | 1.2048 | 600 | 0.6221 | 0.7340 | 0.7329 | 0.7321 | 0.7329 |
66
- | 0.6355 | 1.4056 | 700 | 0.6216 | 0.7358 | 0.7400 | 0.7368 | 0.7400 |
67
- | 0.6074 | 1.6064 | 800 | 0.6261 | 0.7463 | 0.7490 | 0.7462 | 0.7490 |
68
- | 0.5999 | 1.8072 | 900 | 0.6364 | 0.7374 | 0.7420 | 0.7370 | 0.7420 |
69
 
70
 
71
  ### Framework versions
@@ -73,4 +63,4 @@ The following hyperparameters were used during training:
73
  - Transformers 4.41.2
74
  - Pytorch 2.3.0+cu121
75
  - Datasets 2.19.1
76
- - Tokenizers 0.19.1
 
11
  model-index:
12
  - name: flan-t5-small-hallucination-text-classification
13
  results: []
 
 
14
  ---
15
 
16
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
18
 
19
  # flan-t5-small-hallucination-text-classification
20
 
21
+ This model is a fine-tuned version of [google/flan-t5-small](https://huggingface.co/google/flan-t5-small) on the None dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.6695
24
+ - Precision: 0.7097
25
+ - Recall: 0.7129
26
+ - F1: 0.7034
27
+ - Accuracy: 0.7129
28
 
29
  ## Model description
30
 
 
44
 
45
  The following hyperparameters were used during training:
46
  - learning_rate: 0.0003
47
+ - train_batch_size: 16
48
+ - eval_batch_size: 16
49
  - seed: 42
50
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
  - lr_scheduler_type: linear
52
+ - num_epochs: 1
53
 
54
  ### Training results
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
57
  |:-------------:|:------:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
58
+ | 0.827 | 0.8032 | 200 | 0.6695 | 0.7097 | 0.7129 | 0.7034 | 0.7129 |
 
 
 
 
 
 
 
 
59
 
60
 
61
  ### Framework versions
 
63
  - Transformers 4.41.2
64
  - Pytorch 2.3.0+cu121
65
  - Datasets 2.19.1
66
+ - Tokenizers 0.19.1
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:383849146455fb075eafec8612a22fa07fe77ea990b0c1bbb40e827458ac3e31
3
  size 243128252
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:100aecc5061f1504661aa48caad4fec0fd1a39b1bdc9dbc5df6c2014fb65f0af
3
  size 243128252
runs/Jun01_14-34-21_b5a25765b253/events.out.tfevents.1717253446.b5a25765b253.9583.3 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:068997926963bb27c4d5222cce97aaeee3b64fcab3901ffe919bee4d3c081999
3
+ size 560
runs/Jun01_15-06-59_b5a25765b253/events.out.tfevents.1717254435.b5a25765b253.9583.4 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c369ee9b35bc87a4028bceaea8f02ae792e8c9a68fbcee826ace8b5a30b868d6
3
+ size 6988
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0dff502ea4b20a183d35fafd8d9f341411e78a45e1c198d7153d8bc3ac9fe671
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d9865670cf48304a5ec199938d12dd6ed72e7839b8ad45ae173458aa90ab8a0
3
  size 5240