lfhase commited on
Commit
3852911
·
verified ·
1 Parent(s): ebc253b

Upload folder using huggingface_hub

Browse files
Files changed (33) hide show
  1. vicuna-selfies/graph-text-molgen/forward_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/README.md +9 -0
  2. vicuna-selfies/graph-text-molgen/forward_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/adapter_config.json +26 -0
  3. vicuna-selfies/graph-text-molgen/forward_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/adapter_model.bin +3 -0
  4. vicuna-selfies/graph-text-molgen/forward_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/config.json +36 -0
  5. vicuna-selfies/graph-text-molgen/forward_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/non_lora_trainables.bin +3 -0
  6. vicuna-selfies/graph-text-molgen/forward_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/trainer_state.json +0 -0
  7. vicuna-selfies/graph-text-molgen/property_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep/README.md +9 -0
  8. vicuna-selfies/graph-text-molgen/property_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep/adapter_config.json +26 -0
  9. vicuna-selfies/graph-text-molgen/property_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep/adapter_model.bin +3 -0
  10. vicuna-selfies/graph-text-molgen/property_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep/config.json +36 -0
  11. vicuna-selfies/graph-text-molgen/property_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep/non_lora_trainables.bin +3 -0
  12. vicuna-selfies/graph-text-molgen/property_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep/trainer_state.json +0 -0
  13. vicuna-selfies/graph-text-molgen/reagent_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/README.md +9 -0
  14. vicuna-selfies/graph-text-molgen/reagent_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/adapter_config.json +26 -0
  15. vicuna-selfies/graph-text-molgen/reagent_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/adapter_model.bin +3 -0
  16. vicuna-selfies/graph-text-molgen/reagent_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/config.json +36 -0
  17. vicuna-selfies/graph-text-molgen/reagent_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/non_lora_trainables.bin +3 -0
  18. vicuna-selfies/graph-text-molgen/reagent_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/trainer_state.json +0 -0
  19. vicuna-selfies/graph-text-molgen/retrosynthesis-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/README.md +11 -0
  20. vicuna-selfies/graph-text-molgen/retrosynthesis-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/adapter_config.json +26 -0
  21. vicuna-selfies/graph-text-molgen/retrosynthesis-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/adapter_model.bin +3 -0
  22. vicuna-selfies/graph-text-molgen/retrosynthesis-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/config.json +36 -0
  23. vicuna-selfies/graph-text-molgen/retrosynthesis-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/non_lora_trainables.bin +3 -0
  24. vicuna-selfies/graph-text-molgen/retrosynthesis-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/trainer_state.json +0 -0
  25. vicuna-selfies/llava-hvqvae2-vicuna-v1-3-7b-pretrain/config.json +36 -0
  26. vicuna-selfies/llava-hvqvae2-vicuna-v1-3-7b-pretrain/mm_projector.bin +3 -0
  27. vicuna-selfies/llava-hvqvae2-vicuna-v1-3-7b-pretrain/trainer_state.json +0 -0
  28. vicuna/MoleculeNet-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-large/README.md +9 -0
  29. vicuna/MoleculeNet-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-large/adapter_config.json +26 -0
  30. vicuna/MoleculeNet-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-large/adapter_model.bin +3 -0
  31. vicuna/MoleculeNet-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-large/config.json +39 -0
  32. vicuna/MoleculeNet-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-large/non_lora_trainables.bin +3 -0
  33. vicuna/MoleculeNet-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-large/trainer_state.json +0 -0
vicuna-selfies/graph-text-molgen/forward_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.5.0
vicuna-selfies/graph-text-molgen/forward_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "vicuna-v1-3-7b",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 16,
11
+ "lora_dropout": 0.05,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 64,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "down_proj",
18
+ "o_proj",
19
+ "up_proj",
20
+ "k_proj",
21
+ "q_proj",
22
+ "v_proj",
23
+ "gate_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM"
26
+ }
vicuna-selfies/graph-text-molgen/forward_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df3eab2c5f130a406bd6b6a70ff6f7d3af7060199b9b081057595807e207ab94
3
+ size 319970957
vicuna-selfies/graph-text-molgen/forward_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "vicuna-v1-3-7b",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "freeze_mm_mlp_adapter": false,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 4096,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 11008,
13
+ "max_position_embeddings": 2048,
14
+ "mm_graph_tower": "hvqvae2",
15
+ "mm_hidden_size": 308,
16
+ "mm_projector_type": "hlinear",
17
+ "mm_use_im_patch_token": false,
18
+ "mm_use_im_start_end": false,
19
+ "model_type": "llava_graph",
20
+ "num_attention_heads": 32,
21
+ "num_hidden_layers": 32,
22
+ "num_key_value_heads": 32,
23
+ "pad_token_id": 0,
24
+ "pretraining_tp": 1,
25
+ "rms_norm_eps": 1e-06,
26
+ "rope_scaling": null,
27
+ "rope_theta": 10000.0,
28
+ "tie_word_embeddings": false,
29
+ "torch_dtype": "float16",
30
+ "transformers_version": "4.33.2",
31
+ "tune_mm_mlp_adapter": false,
32
+ "use_cache": true,
33
+ "use_lap_pe": true,
34
+ "use_mm_proj": true,
35
+ "vocab_size": 32000
36
+ }
vicuna-selfies/graph-text-molgen/forward_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/non_lora_trainables.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0c5c0f66e5988f59e86ec4f21bf39322f67bb3847df4da7ed49f2b5e227a66e0
3
+ size 11335231
vicuna-selfies/graph-text-molgen/forward_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
vicuna-selfies/graph-text-molgen/property_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.5.0
vicuna-selfies/graph-text-molgen/property_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep/adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "vicuna-v1-3-7b",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 16,
11
+ "lora_dropout": 0.05,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 64,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "down_proj",
18
+ "k_proj",
19
+ "gate_proj",
20
+ "q_proj",
21
+ "o_proj",
22
+ "v_proj",
23
+ "up_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM"
26
+ }
vicuna-selfies/graph-text-molgen/property_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fd133c19a6a2a4b6b5b403f19c54a54146895e78912b17bea72704409c68fafa
3
+ size 319970957
vicuna-selfies/graph-text-molgen/property_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "vicuna-v1-3-7b",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "freeze_mm_mlp_adapter": false,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 4096,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 11008,
13
+ "max_position_embeddings": 2048,
14
+ "mm_graph_tower": "hvqvae2",
15
+ "mm_hidden_size": 308,
16
+ "mm_projector_type": "hlinear",
17
+ "mm_use_im_patch_token": false,
18
+ "mm_use_im_start_end": false,
19
+ "model_type": "llava_graph",
20
+ "num_attention_heads": 32,
21
+ "num_hidden_layers": 32,
22
+ "num_key_value_heads": 32,
23
+ "pad_token_id": 0,
24
+ "pretraining_tp": 1,
25
+ "rms_norm_eps": 1e-06,
26
+ "rope_scaling": null,
27
+ "rope_theta": 10000.0,
28
+ "tie_word_embeddings": false,
29
+ "torch_dtype": "float16",
30
+ "transformers_version": "4.33.2",
31
+ "tune_mm_mlp_adapter": false,
32
+ "use_cache": true,
33
+ "use_lap_pe": true,
34
+ "use_mm_proj": true,
35
+ "vocab_size": 32000
36
+ }
vicuna-selfies/graph-text-molgen/property_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep/non_lora_trainables.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd9fee870b58c4519fcfef92bc257da4d3fbce206c3fd051151c930347e4fb36
3
+ size 11335231
vicuna-selfies/graph-text-molgen/property_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
vicuna-selfies/graph-text-molgen/reagent_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.5.0
vicuna-selfies/graph-text-molgen/reagent_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "vicuna-v1-3-7b",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 16,
11
+ "lora_dropout": 0.05,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 64,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "k_proj",
18
+ "o_proj",
19
+ "q_proj",
20
+ "down_proj",
21
+ "v_proj",
22
+ "gate_proj",
23
+ "up_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM"
26
+ }
vicuna-selfies/graph-text-molgen/reagent_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6dfdb3407a8583bee602ab6f9006ae849e94c47cdde8560dbaa8efd18a05b716
3
+ size 319970957
vicuna-selfies/graph-text-molgen/reagent_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "vicuna-v1-3-7b",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "freeze_mm_mlp_adapter": false,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 4096,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 11008,
13
+ "max_position_embeddings": 2048,
14
+ "mm_graph_tower": "hvqvae2",
15
+ "mm_hidden_size": 308,
16
+ "mm_projector_type": "hlinear",
17
+ "mm_use_im_patch_token": false,
18
+ "mm_use_im_start_end": false,
19
+ "model_type": "llava_graph",
20
+ "num_attention_heads": 32,
21
+ "num_hidden_layers": 32,
22
+ "num_key_value_heads": 32,
23
+ "pad_token_id": 0,
24
+ "pretraining_tp": 1,
25
+ "rms_norm_eps": 1e-06,
26
+ "rope_scaling": null,
27
+ "rope_theta": 10000.0,
28
+ "tie_word_embeddings": false,
29
+ "torch_dtype": "float16",
30
+ "transformers_version": "4.33.2",
31
+ "tune_mm_mlp_adapter": false,
32
+ "use_cache": true,
33
+ "use_lap_pe": true,
34
+ "use_mm_proj": true,
35
+ "vocab_size": 32000
36
+ }
vicuna-selfies/graph-text-molgen/reagent_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/non_lora_trainables.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b6107c13a4144d5e3d3b5d6829d32e150a30000801cbe900f70439ba4124db9
3
+ size 11335231
vicuna-selfies/graph-text-molgen/reagent_pred-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
vicuna-selfies/graph-text-molgen/retrosynthesis-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/README.md ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+ - PEFT 0.5.0
9
+ - PEFT 0.5.0
10
+
11
+ - PEFT 0.5.0
vicuna-selfies/graph-text-molgen/retrosynthesis-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "vicuna-v1-3-7b",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 16,
11
+ "lora_dropout": 0.05,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 64,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "k_proj",
18
+ "up_proj",
19
+ "o_proj",
20
+ "gate_proj",
21
+ "v_proj",
22
+ "q_proj",
23
+ "down_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM"
26
+ }
vicuna-selfies/graph-text-molgen/retrosynthesis-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f8383a6ab82cffb87089bc220430c976e4e48233126240829df9097cc036b62
3
+ size 319970957
vicuna-selfies/graph-text-molgen/retrosynthesis-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "vicuna-v1-3-7b",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "freeze_mm_mlp_adapter": false,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 4096,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 11008,
13
+ "max_position_embeddings": 2048,
14
+ "mm_graph_tower": "hvqvae2",
15
+ "mm_hidden_size": 308,
16
+ "mm_projector_type": "hlinear",
17
+ "mm_use_im_patch_token": false,
18
+ "mm_use_im_start_end": false,
19
+ "model_type": "llava_graph",
20
+ "num_attention_heads": 32,
21
+ "num_hidden_layers": 32,
22
+ "num_key_value_heads": 32,
23
+ "pad_token_id": 0,
24
+ "pretraining_tp": 1,
25
+ "rms_norm_eps": 1e-06,
26
+ "rope_scaling": null,
27
+ "rope_theta": 10000.0,
28
+ "tie_word_embeddings": false,
29
+ "torch_dtype": "float16",
30
+ "transformers_version": "4.33.2",
31
+ "tune_mm_mlp_adapter": false,
32
+ "use_cache": true,
33
+ "use_lap_pe": true,
34
+ "use_mm_proj": true,
35
+ "vocab_size": 32000
36
+ }
vicuna-selfies/graph-text-molgen/retrosynthesis-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/non_lora_trainables.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:112ad676a20bd1f3c6359b7d20a0f8b26e0cf2e82651a9fd23b5b325cf9df8bb
3
+ size 11335231
vicuna-selfies/graph-text-molgen/retrosynthesis-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-5ep16bz/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
vicuna-selfies/llava-hvqvae2-vicuna-v1-3-7b-pretrain/config.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "vicuna-v1-3-7b",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "freeze_mm_mlp_adapter": false,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 4096,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 11008,
13
+ "max_position_embeddings": 2048,
14
+ "mm_graph_tower": "hvqvae2",
15
+ "mm_hidden_size": 308,
16
+ "mm_projector_type": "hlinear",
17
+ "mm_use_im_patch_token": false,
18
+ "mm_use_im_start_end": false,
19
+ "model_type": "llava_graph",
20
+ "num_attention_heads": 32,
21
+ "num_hidden_layers": 32,
22
+ "num_key_value_heads": 32,
23
+ "pad_token_id": 0,
24
+ "pretraining_tp": 1,
25
+ "rms_norm_eps": 1e-06,
26
+ "rope_scaling": null,
27
+ "rope_theta": 10000.0,
28
+ "tie_word_embeddings": false,
29
+ "torch_dtype": "float16",
30
+ "transformers_version": "4.33.2",
31
+ "tune_mm_mlp_adapter": true,
32
+ "use_cache": true,
33
+ "use_lap_pe": true,
34
+ "use_mm_proj": true,
35
+ "vocab_size": 32000
36
+ }
vicuna-selfies/llava-hvqvae2-vicuna-v1-3-7b-pretrain/mm_projector.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:012ebe5b3092982b8285d8dd747f16ba82045b800c4e83085dfbb33e04306ef2
3
+ size 7596223
vicuna-selfies/llava-hvqvae2-vicuna-v1-3-7b-pretrain/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
vicuna/MoleculeNet-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-large/README.md ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ ---
4
+ ## Training procedure
5
+
6
+ ### Framework versions
7
+
8
+
9
+ - PEFT 0.5.0
vicuna/MoleculeNet-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-large/adapter_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "auto_mapping": null,
3
+ "base_model_name_or_path": "./vicuna-v1-3-7b",
4
+ "bias": "none",
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "layers_pattern": null,
9
+ "layers_to_transform": null,
10
+ "lora_alpha": 256,
11
+ "lora_dropout": 0.05,
12
+ "modules_to_save": null,
13
+ "peft_type": "LORA",
14
+ "r": 128,
15
+ "revision": null,
16
+ "target_modules": [
17
+ "up_proj",
18
+ "gate_proj",
19
+ "q_proj",
20
+ "v_proj",
21
+ "k_proj",
22
+ "o_proj",
23
+ "down_proj"
24
+ ],
25
+ "task_type": "CAUSAL_LM"
26
+ }
vicuna/MoleculeNet-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-large/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84efdc09bf8be031148e59d5fbdb4ad49463495be923e0333731d6f26a3cd0bf
3
+ size 639786637
vicuna/MoleculeNet-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-large/config.json ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "./vicuna-v1-3-7b",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 2,
10
+ "freeze_mm_mlp_adapter": false,
11
+ "hidden_act": "silu",
12
+ "hidden_size": 4096,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 11008,
15
+ "max_position_embeddings": 2048,
16
+ "mlp_bias": false,
17
+ "mm_graph_tower": "hvqvae2",
18
+ "mm_hidden_size": 308,
19
+ "mm_projector_type": "hlinear",
20
+ "mm_use_im_patch_token": false,
21
+ "mm_use_im_start_end": false,
22
+ "model_type": "llava_graph",
23
+ "num_attention_heads": 32,
24
+ "num_hidden_layers": 32,
25
+ "num_key_value_heads": 32,
26
+ "pad_token_id": 0,
27
+ "pretraining_tp": 1,
28
+ "rms_norm_eps": 1e-06,
29
+ "rope_scaling": null,
30
+ "rope_theta": 10000.0,
31
+ "tie_word_embeddings": false,
32
+ "torch_dtype": "float16",
33
+ "transformers_version": "4.41.0",
34
+ "tune_mm_mlp_adapter": false,
35
+ "use_cache": true,
36
+ "use_lap_pe": true,
37
+ "use_mm_proj": true,
38
+ "vocab_size": 32000
39
+ }
vicuna/MoleculeNet-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-large/non_lora_trainables.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1b1709a04ed78d9e9be71a9a58720eae5356875436ec2b7fd8db2ee13d59235
3
+ size 11335231
vicuna/MoleculeNet-llava-hvqvae2-vicuna-v1-3-7b-finetune_lora-large/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff