weathon commited on
Commit
765702f
·
verified ·
1 Parent(s): f51e8de

End of training

Browse files
Files changed (3) hide show
  1. README.md +11 -9
  2. model.safetensors +1 -1
  3. training_args.bin +1 -1
README.md CHANGED
@@ -16,13 +16,13 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [Salesforce/blip-image-captioning-base](https://huggingface.co/Salesforce/blip-image-captioning-base) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - eval_loss: 0.0174
20
- - eval_accuracy: 0.8140
21
- - eval_runtime: 2.6673
22
- - eval_samples_per_second: 64.486
23
- - eval_steps_per_second: 2.249
24
- - epoch: 2.2621
25
- - step: 10100
26
 
27
  ## Model description
28
 
@@ -41,16 +41,18 @@ More information needed
41
  ### Training hyperparameters
42
 
43
  The following hyperparameters were used during training:
44
- - learning_rate: 3e-05
45
  - train_batch_size: 32
46
  - eval_batch_size: 32
47
  - seed: 42
48
  - gradient_accumulation_steps: 2
49
  - total_train_batch_size: 64
50
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
51
- - lr_scheduler_type: linear
 
52
  - num_epochs: 5
53
  - mixed_precision_training: Native AMP
 
54
 
55
  ### Framework versions
56
 
 
16
 
17
  This model is a fine-tuned version of [Salesforce/blip-image-captioning-base](https://huggingface.co/Salesforce/blip-image-captioning-base) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - eval_loss: 1.3598
20
+ - eval_accuracy: 1.0
21
+ - eval_runtime: 8.776
22
+ - eval_samples_per_second: 83.865
23
+ - eval_steps_per_second: 2.621
24
+ - epoch: 0.2240
25
+ - step: 1000
26
 
27
  ## Model description
28
 
 
41
  ### Training hyperparameters
42
 
43
  The following hyperparameters were used during training:
44
+ - learning_rate: 5e-05
45
  - train_batch_size: 32
46
  - eval_batch_size: 32
47
  - seed: 42
48
  - gradient_accumulation_steps: 2
49
  - total_train_batch_size: 64
50
  - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
51
+ - lr_scheduler_type: cosine
52
+ - lr_scheduler_warmup_ratio: 0.05
53
  - num_epochs: 5
54
  - mixed_precision_training: Native AMP
55
+ - label_smoothing_factor: 0.1
56
 
57
  ### Framework versions
58
 
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b012239faa9b38faf68818954dc660c665fb509d0c7e7e7b9815b279effbfe7c
3
  size 989717056
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9dfe76ff4de65e0cb96593c91b4402e18346ef2647bdadd33f8e7b40aa972b3b
3
  size 989717056
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:963641b19ed5d01dc0a98fc594075bc6e8077ba61258d93e24154573a89ef0be
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e6bf3aeb60da940eeccd2692a261595a243beddc5da15b78a7d6226c67c3f28
3
  size 5304