danielhanchen commited on
Commit
a0eadae
·
verified ·
1 Parent(s): 79a79c0

Upload folder using huggingface_hub

Browse files
config.json CHANGED
@@ -23,6 +23,7 @@
23
  32
24
  ],
25
  "sscp_conv_eps": 0.001,
 
26
  "sscp_conv_kernel_size": [
27
  [
28
  3,
@@ -80,12 +81,12 @@
80
  "modality_router",
81
  "embedding_projection",
82
  "input_proj_linear",
83
- "laurel",
84
- "per_layer_input_gate",
85
  "per_layer_projection",
86
  "audio_tower",
87
  "embed_vision",
88
- "embed_audio"
 
 
89
  ],
90
  "llm_int8_threshold": 6.0,
91
  "load_in_4bit": true,
@@ -138,7 +139,38 @@
138
  "hidden_size": 2048,
139
  "hidden_size_per_layer_input": 256,
140
  "initializer_range": 0.02,
141
- "intermediate_size": 8192,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
142
  "laurel_rank": 64,
143
  "layer_types": [
144
  "sliding_attention",
@@ -190,7 +222,7 @@
190
  "vocab_size_per_layer_input": 262144
191
  },
192
  "torch_dtype": "bfloat16",
193
- "transformers_version": "4.53.0.dev0",
194
  "unsloth_fixed": true,
195
  "vision_config": {
196
  "architecture": "mobilenetv5_300m_enc",
@@ -201,6 +233,7 @@
201
  "LABEL_0",
202
  "LABEL_1"
203
  ],
 
204
  "model_type": "gemma3n_vision",
205
  "num_classes": 2,
206
  "rms_norm_eps": 1e-06,
 
23
  32
24
  ],
25
  "sscp_conv_eps": 0.001,
26
+ "sscp_conv_group_norm_eps": 0.001,
27
  "sscp_conv_kernel_size": [
28
  [
29
  3,
 
81
  "modality_router",
82
  "embedding_projection",
83
  "input_proj_linear",
 
 
84
  "per_layer_projection",
85
  "audio_tower",
86
  "embed_vision",
87
+ "embed_audio",
88
+ "altup_projections",
89
+ "altup_unembed_projections"
90
  ],
91
  "llm_int8_threshold": 6.0,
92
  "load_in_4bit": true,
 
139
  "hidden_size": 2048,
140
  "hidden_size_per_layer_input": 256,
141
  "initializer_range": 0.02,
142
+ "intermediate_size": [
143
+ 8192,
144
+ 8192,
145
+ 8192,
146
+ 8192,
147
+ 8192,
148
+ 8192,
149
+ 8192,
150
+ 8192,
151
+ 8192,
152
+ 8192,
153
+ 8192,
154
+ 8192,
155
+ 8192,
156
+ 8192,
157
+ 8192,
158
+ 8192,
159
+ 8192,
160
+ 8192,
161
+ 8192,
162
+ 8192,
163
+ 8192,
164
+ 8192,
165
+ 8192,
166
+ 8192,
167
+ 8192,
168
+ 8192,
169
+ 8192,
170
+ 8192,
171
+ 8192,
172
+ 8192
173
+ ],
174
  "laurel_rank": 64,
175
  "layer_types": [
176
  "sliding_attention",
 
222
  "vocab_size_per_layer_input": 262144
223
  },
224
  "torch_dtype": "bfloat16",
225
+ "transformers_version": "4.53.0",
226
  "unsloth_fixed": true,
227
  "vision_config": {
228
  "architecture": "mobilenetv5_300m_enc",
 
233
  "LABEL_0",
234
  "LABEL_1"
235
  ],
236
+ "model_args": null,
237
  "model_type": "gemma3n_vision",
238
  "num_classes": 2,
239
  "rms_norm_eps": 1e-06,
generation_config.json CHANGED
@@ -6,5 +6,5 @@
6
  "pad_token_id": 0,
7
  "top_k": 64,
8
  "top_p": 0.95,
9
- "transformers_version": "4.53.0.dev0"
10
  }
 
6
  "pad_token_id": 0,
7
  "top_k": 64,
8
  "top_p": 0.95,
9
+ "transformers_version": "4.53.0"
10
  }
model-00001-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6744c3e6b7e06dca3de55f177bfa56e01821f39ec31e73c96c52a5d57139618d
3
- size 2684992582
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01a9c74c1d69b300df4b0816755a72b10aac6cb6239a2dd56f9fb3d8e3f28bbe
3
+ size 2650152952
model-00002-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e3b18d1e771b899eeae930afe5226e92771647b8783f3e0e55174f32c20ae569
3
- size 4999684416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b6b452078a424fcf218ab19cef69fd8120458812a9f3f6a436a577b416714465
3
+ size 4994547573
model-00003-of-00003.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:95ee0dc1d6416c141a75d8ab5807d38284e2a322102d823be189ad47962cd5c3
3
- size 426305256
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e2f57bf11393c77e467d8b7e14fe4c3e7f016ad1da47058ad41c91b5676d839
3
+ size 468779960
model.safetensors.index.json CHANGED
The diff for this file is too large to render. See raw diff
 
preprocessor_config.json CHANGED
@@ -3,6 +3,7 @@
3
  "data_format": "channels_first",
4
  "default_to_square": false,
5
  "device": null,
 
6
  "dither": 0.0,
7
  "do_center_crop": null,
8
  "do_convert_rgb": null,
 
3
  "data_format": "channels_first",
4
  "default_to_square": false,
5
  "device": null,
6
+ "disable_grouping": null,
7
  "dither": 0.0,
8
  "do_center_crop": null,
9
  "do_convert_rgb": null,