codelion commited on
Commit
814fae3
·
verified ·
1 Parent(s): 705d46a

Upload ICM-DPO enhanced Gemma PEFT adapter with comprehensive LoRA and model card

Browse files
README.md CHANGED
@@ -50,14 +50,14 @@ This model demonstrates comprehensive capability enhancement using ICM-generated
50
  ### Dataset
51
  - **Source**: [codelion/gemma-3-270m-icm-dpo](https://huggingface.co/datasets/codelion/gemma-3-270m-icm-dpo)
52
  - **Method**: ICM (Internal Coherence Maximization) for label-free preference generation
53
- - **Training Samples**: 44286
54
  - **Evaluation Samples**: 50
55
 
56
  ### Training Configuration
57
  - **Epochs**: 1
58
- - **Batch Size**: 4 (per device)
59
  - **Gradient Accumulation**: 8 steps
60
- - **Effective Batch Size**: 32
61
  - **Learning Rate**: 2e-07
62
  - **Optimizer**: paged_adamw_8bit
63
  - **Memory Optimization**: BF16, Gradient Checkpointing
 
50
  ### Dataset
51
  - **Source**: [codelion/gemma-3-270m-icm-dpo](https://huggingface.co/datasets/codelion/gemma-3-270m-icm-dpo)
52
  - **Method**: ICM (Internal Coherence Maximization) for label-free preference generation
53
+ - **Training Samples**: 1812
54
  - **Evaluation Samples**: 50
55
 
56
  ### Training Configuration
57
  - **Epochs**: 1
58
+ - **Batch Size**: 1 (per device)
59
  - **Gradient Accumulation**: 8 steps
60
+ - **Effective Batch Size**: 8
61
  - **Learning Rate**: 2e-07
62
  - **Optimizer**: paged_adamw_8bit
63
  - **Memory Optimization**: BF16, Gradient Checkpointing
adapter_config.json CHANGED
@@ -28,13 +28,13 @@
28
  "rank_pattern": {},
29
  "revision": null,
30
  "target_modules": [
31
- "o_proj",
32
- "q_proj",
33
- "v_proj",
34
  "up_proj",
35
  "k_proj",
 
36
  "down_proj",
37
- "gate_proj"
 
38
  ],
39
  "target_parameters": null,
40
  "task_type": "CAUSAL_LM",
 
28
  "rank_pattern": {},
29
  "revision": null,
30
  "target_modules": [
31
+ "gate_proj",
 
 
32
  "up_proj",
33
  "k_proj",
34
+ "v_proj",
35
  "down_proj",
36
+ "o_proj",
37
+ "q_proj"
38
  ],
39
  "target_parameters": null,
40
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b937d87b2e5243afade0703add1246907bf08177cd3a0e47f32aefa54da5e594
3
  size 701497992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8d9fbccbcd31bc8df08fed608b79046818801f5a92ce6537aee6cb7d2b10c085
3
  size 701497992
added_tokens.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "<image_soft_token>": 262144
3
+ }
results.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
  "training_metrics": {
3
- "train_runtime": 4688.6863,
4
- "train_samples_per_second": 9.445,
5
- "train_steps_per_second": 0.295,
6
  "total_flos": 0.0,
7
- "train_loss": 0.6801743100833342,
8
  "epoch": 1.0
9
  },
10
  "config": {
@@ -31,7 +31,7 @@
31
  "beta": 0.02,
32
  "max_length": 1024,
33
  "max_prompt_length": 512,
34
- "batch_size": 4,
35
  "gradient_accumulation_steps": 8,
36
  "learning_rate": 2e-07,
37
  "num_train_epochs": 1,
 
1
  {
2
  "training_metrics": {
3
+ "train_runtime": 1035.1183,
4
+ "train_samples_per_second": 1.751,
5
+ "train_steps_per_second": 0.219,
6
  "total_flos": 0.0,
7
+ "train_loss": 0.6929902463232368,
8
  "epoch": 1.0
9
  },
10
  "config": {
 
31
  "beta": 0.02,
32
  "max_length": 1024,
33
  "max_prompt_length": 512,
34
+ "batch_size": 1,
35
  "gradient_accumulation_steps": 8,
36
  "learning_rate": 2e-07,
37
  "num_train_epochs": 1,
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1299c11d7cf632ef3b4e11937501358ada021bbdf7c47638d13c0ee982f2e79c
3
+ size 4689074
training_config.json CHANGED
@@ -22,7 +22,7 @@
22
  "beta": 0.02,
23
  "max_length": 1024,
24
  "max_prompt_length": 512,
25
- "batch_size": 4,
26
  "gradient_accumulation_steps": 8,
27
  "learning_rate": 2e-07,
28
  "num_train_epochs": 1,
 
22
  "beta": 0.02,
23
  "max_length": 1024,
24
  "max_prompt_length": 512,
25
+ "batch_size": 1,
26
  "gradient_accumulation_steps": 8,
27
  "learning_rate": 2e-07,
28
  "num_train_epochs": 1,