gimarchetti commited on
Commit
3c8125d
·
verified ·
1 Parent(s): 0f98fcd

End of training

Browse files
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: transformers
3
  license: gemma
4
- base_model: google/paligemma-3b-pt-224
5
  tags:
6
  - generated_from_trainer
7
  model-index:
@@ -14,9 +14,7 @@ should probably proofread and complete it, then remove this comment. -->
14
 
15
  # paligemma-rocov2
16
 
17
- This model is a fine-tuned version of [google/paligemma-3b-pt-224](https://huggingface.co/google/paligemma-3b-pt-224) on an unknown dataset.
18
- It achieves the following results on the evaluation set:
19
- - Loss: 2.4124
20
 
21
  ## Model description
22
 
@@ -46,40 +44,6 @@ The following hyperparameters were used during training:
46
 
47
  ### Training results
48
 
49
- | Training Loss | Epoch | Step | Validation Loss |
50
- |:-------------:|:------:|:-----:|:---------------:|
51
- | 2.5859 | 0.1223 | 1000 | 2.5663 |
52
- | 2.5199 | 0.2446 | 2000 | 2.5207 |
53
- | 2.4867 | 0.3668 | 3000 | 2.4848 |
54
- | 2.4727 | 0.4891 | 4000 | 2.4690 |
55
- | 2.4836 | 0.6114 | 5000 | 2.4517 |
56
- | 2.4412 | 0.7337 | 6000 | 2.4379 |
57
- | 2.4192 | 0.8560 | 7000 | 2.4280 |
58
- | 2.427 | 0.9782 | 8000 | 2.4206 |
59
- | 2.2767 | 1.1005 | 9000 | 2.4229 |
60
- | 2.3071 | 1.2228 | 10000 | 2.4188 |
61
- | 2.2753 | 1.3451 | 11000 | 2.4153 |
62
- | 2.2722 | 1.4674 | 12000 | 2.4150 |
63
- | 2.2533 | 1.5896 | 13000 | 2.4124 |
64
- | 2.2494 | 1.7119 | 14000 | 2.4099 |
65
- | 2.2546 | 1.8342 | 15000 | 2.4086 |
66
- | 2.2548 | 1.9565 | 16000 | 2.4068 |
67
- | 2.2179 | 2.0787 | 17000 | 2.4116 |
68
- | 2.2062 | 2.2010 | 18000 | 2.4125 |
69
- | 2.2588 | 2.3233 | 19000 | 2.4117 |
70
- | 2.2389 | 2.4456 | 20000 | 2.4122 |
71
- | 2.2231 | 2.5679 | 21000 | 2.4113 |
72
- | 2.232 | 2.6901 | 22000 | 2.4112 |
73
- | 2.2101 | 2.8124 | 23000 | 2.4109 |
74
- | 2.2038 | 2.9347 | 24000 | 2.4110 |
75
- | 2.2442 | 3.0570 | 25000 | 2.4116 |
76
- | 2.2474 | 3.1793 | 26000 | 2.4118 |
77
- | 2.2272 | 3.3015 | 27000 | 2.4123 |
78
- | 2.1801 | 3.4238 | 28000 | 2.4125 |
79
- | 2.1884 | 3.5461 | 29000 | 2.4125 |
80
- | 2.2271 | 3.6684 | 30000 | 2.4124 |
81
- | 2.182 | 3.7907 | 31000 | 2.4125 |
82
- | 2.1832 | 3.9129 | 32000 | 2.4124 |
83
 
84
 
85
  ### Framework versions
 
1
  ---
2
  library_name: transformers
3
  license: gemma
4
+ base_model: google/paligemma-3b-pt-448
5
  tags:
6
  - generated_from_trainer
7
  model-index:
 
14
 
15
  # paligemma-rocov2
16
 
17
+ This model is a fine-tuned version of [google/paligemma-3b-pt-448](https://huggingface.co/google/paligemma-3b-pt-448) on an unknown dataset.
 
 
18
 
19
  ## Model description
20
 
 
44
 
45
  ### Training results
46
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
 
48
 
49
  ### Framework versions
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "google/paligemma-3b-pt-224",
3
  "architectures": [
4
  "PaliGemmaForConditionalGeneration"
5
  ],
@@ -17,7 +17,7 @@
17
  "model_type": "gemma",
18
  "num_attention_heads": 8,
19
  "num_hidden_layers": 18,
20
- "num_image_tokens": 256,
21
  "num_key_value_heads": 1,
22
  "torch_dtype": "float32",
23
  "vocab_size": 257216
@@ -26,11 +26,12 @@
26
  "transformers_version": "4.44.2",
27
  "vision_config": {
28
  "hidden_size": 1152,
 
29
  "intermediate_size": 4304,
30
  "model_type": "siglip_vision_model",
31
  "num_attention_heads": 16,
32
  "num_hidden_layers": 27,
33
- "num_image_tokens": 256,
34
  "patch_size": 14,
35
  "projection_dim": 2048,
36
  "projector_hidden_act": "gelu_fast",
 
1
  {
2
+ "_name_or_path": "google/paligemma-3b-pt-448",
3
  "architectures": [
4
  "PaliGemmaForConditionalGeneration"
5
  ],
 
17
  "model_type": "gemma",
18
  "num_attention_heads": 8,
19
  "num_hidden_layers": 18,
20
+ "num_image_tokens": 1024,
21
  "num_key_value_heads": 1,
22
  "torch_dtype": "float32",
23
  "vocab_size": 257216
 
26
  "transformers_version": "4.44.2",
27
  "vision_config": {
28
  "hidden_size": 1152,
29
+ "image_size": 448,
30
  "intermediate_size": 4304,
31
  "model_type": "siglip_vision_model",
32
  "num_attention_heads": 16,
33
  "num_hidden_layers": 27,
34
+ "num_image_tokens": 1024,
35
  "patch_size": 14,
36
  "projection_dim": 2048,
37
  "projector_hidden_act": "gelu_fast",
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6c44c379ac1b3e96f9ab51a19c8be0808e6bdb46f01e1f5d3675188e62a591d
3
- size 4985044392
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f386d29458bd3506cfc56a0160995c78a4ff64f9b54af19cd882d62a4eb25ae
3
+ size 4986813864
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:381e9b36537211866504dd5519102ad4b215522f1b803d22780b3c27506ca49f
3
  size 861970608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:932622f017540e0293d32ffa49bdb2617f5911415e9874aee228242cf1bf9c85
3
  size 861970608
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 5846932960
4
  },
5
  "weight_map": {
6
  "language_model.model.embed_tokens.weight": "model-00001-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 5848702432
4
  },
5
  "weight_map": {
6
  "language_model.model.embed_tokens.weight": "model-00001-of-00002.safetensors",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da7a48c361d9879cee70373f3501b32a8ab2639e92fe721f2146a61b2d590603
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1feedb5d06bfd80099941e903459d7ca332bf5eb36d369b55b76106ac12dcc63
3
  size 5176