danielhanchen commited on
Commit
addf43a
·
verified ·
1 Parent(s): bd490f7

Add files using upload-large-folder tool

Browse files
Files changed (1) hide show
  1. config.json +1 -25
config.json CHANGED
@@ -6,11 +6,7 @@
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "bos_token_id": 128000,
9
- "eos_token_id": [
10
- 128001,
11
- 128008,
12
- 128009
13
- ],
14
  "head_dim": 128,
15
  "hidden_act": "silu",
16
  "hidden_size": 8192,
@@ -24,26 +20,6 @@
24
  "num_key_value_heads": 8,
25
  "pad_token_id": 128004,
26
  "pretraining_tp": 1,
27
- "quantization_config": {
28
- "_load_in_4bit": true,
29
- "_load_in_8bit": false,
30
- "bnb_4bit_compute_dtype": "bfloat16",
31
- "bnb_4bit_quant_storage": "uint8",
32
- "bnb_4bit_quant_type": "nf4",
33
- "bnb_4bit_use_double_quant": true,
34
- "llm_int8_enable_fp32_cpu_offload": false,
35
- "llm_int8_has_fp16_weight": false,
36
- "llm_int8_skip_modules": [
37
- "lm_head",
38
- "multi_modal_projector",
39
- "merger",
40
- "modality_projection"
41
- ],
42
- "llm_int8_threshold": 6.0,
43
- "load_in_4bit": true,
44
- "load_in_8bit": false,
45
- "quant_method": "bitsandbytes"
46
- },
47
  "rms_norm_eps": 1e-05,
48
  "rope_scaling": {
49
  "factor": 8.0,
 
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
  "bos_token_id": 128000,
9
+ "eos_token_id": 128001,
 
 
 
 
10
  "head_dim": 128,
11
  "hidden_act": "silu",
12
  "hidden_size": 8192,
 
20
  "num_key_value_heads": 8,
21
  "pad_token_id": 128004,
22
  "pretraining_tp": 1,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  "rms_norm_eps": 1e-05,
24
  "rope_scaling": {
25
  "factor": 8.0,