BinXiao commited on
Commit
20bffb1
·
1 Parent(s): b759baa

Upload model files

Browse files
config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"architectures": ["LlamaForCausalLM"], "bos_token_id": 1, "eos_token_id": 2, "hidden_act": "silu", "hidden_size": 4096, "initializer_range": 0.02, "intermediate_size": 11008, "max_position_embeddings": 2048, "model_type": "llama", "num_attention_heads": 32, "num_hidden_layers": 1, "pad_token_id": 0, "rms_norm_eps": 1e-06, "tie_word_embeddings": false, "torch_dtype": "float16", "transformers_version": "4.28.1", "use_cache": true, "vocab_size": 32000, "clover": {"num_heads": 5, "num_layers": 2, "heads_coefficient": 1.0, "decay_coefficient": 0.7}}
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81393802b7b0fe564152f16b24eeead2d63f5b5b2488dc511b38cf370d219124
3
+ size 3271823464
model_1.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:70216eb1762c6db6557dcc8bf799909c6f6a0b30c77d3bd8d53ea4aac768e494
3
+ size 524288120
optimizer.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8505eafe84e954385074b46e356453ea5c855069792071766dd10d9beab03261
3
+ size 4446518941
random_states_0.pkl ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f0a7023ff189485d80d21db2fff24c44eded85cc0d3d56c74cdbbb80a7f57992
3
+ size 21795
scheduler.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:518f0d2fcd44f64308b6444158a049b05e378e5370ce46e5fd4af79c6c35524b
3
+ size 563