Kartheekb7 commited on
Commit
8bfeb9e
·
verified ·
1 Parent(s): 4a83194

Kartheekb7/peft_llava_llama_2

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- library_name: transformers
3
  license: llama3.2
4
  base_model: meta-llama/Llama-3.2-1B-Instruct
5
  tags:
@@ -33,13 +33,13 @@ More information needed
33
  ### Training hyperparameters
34
 
35
  The following hyperparameters were used during training:
36
- - learning_rate: 0.0001
37
- - train_batch_size: 4
38
- - eval_batch_size: 8
39
  - seed: 42
40
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
41
  - lr_scheduler_type: cosine
42
- - training_steps: 5
43
 
44
  ### Training results
45
 
@@ -47,7 +47,8 @@ The following hyperparameters were used during training:
47
 
48
  ### Framework versions
49
 
 
50
  - Transformers 4.45.1
51
- - Pytorch 2.4.0+cpu
52
  - Datasets 3.0.1
53
- - Tokenizers 0.20.0
 
1
  ---
2
+ library_name: peft
3
  license: llama3.2
4
  base_model: meta-llama/Llama-3.2-1B-Instruct
5
  tags:
 
33
  ### Training hyperparameters
34
 
35
  The following hyperparameters were used during training:
36
+ - learning_rate: 0.0002
37
+ - train_batch_size: 8
38
+ - eval_batch_size: 16
39
  - seed: 42
40
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
41
  - lr_scheduler_type: cosine
42
+ - training_steps: 10000
43
 
44
  ### Training results
45
 
 
47
 
48
  ### Framework versions
49
 
50
+ - PEFT 0.13.2
51
  - Transformers 4.45.1
52
+ - Pytorch 2.4.0
53
  - Datasets 3.0.1
54
+ - Tokenizers 0.20.0
adapter_config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": null,
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 16,
14
+ "lora_dropout": 0.05,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 8,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "v_proj",
24
+ "q_proj",
25
+ "k_proj",
26
+ "o_proj"
27
+ ],
28
+ "task_type": "CAUSAL_LM",
29
+ "use_dora": false,
30
+ "use_rslora": false
31
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17860c0cf5f03ab720fe637f05b45e712181b2c2af3a4773b2755899a819ad94
3
+ size 6833928
runs/Oct29_17-41-50_4be4fe3b6de8/events.out.tfevents.1730223711.4be4fe3b6de8.115.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34deba61516e7f38c72a01ee5ec596fd098d69cb4054c2a82d24d261d35e2fd6
3
+ size 27141
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:26e42dec63cc3c3adacca7fe0920c03f791419ad1dd0cc28a538866ffd176ba7
3
  size 5176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4acc6fa46b6208a4d69e9e7ae1a49babfeabd2c3e3d34bc4992689b017eb3a40
3
  size 5176