Upload folder using huggingface_hub
Browse files- config.json +1 -1
config.json
CHANGED
@@ -9,6 +9,6 @@
|
|
9 |
"embedding_size": 512,
|
10 |
"model_type": "LUAR",
|
11 |
"torch_dtype": "float32",
|
12 |
-
"transformers_version": "4.
|
13 |
"use_memory_efficient_attention": false
|
14 |
}
|
|
|
9 |
"embedding_size": 512,
|
10 |
"model_type": "LUAR",
|
11 |
"torch_dtype": "float32",
|
12 |
+
"transformers_version": "4.45.2",
|
13 |
"use_memory_efficient_attention": false
|
14 |
}
|