Upload ICM-DPO enhanced Gemma PEFT adapter with comprehensive LoRA and model card
Browse files- README.md +3 -3
- adapter_config.json +4 -4
- adapter_model.safetensors +1 -1
- added_tokens.json +3 -0
- results.json +5 -5
- tokenizer.model +3 -0
- training_config.json +1 -1
README.md
CHANGED
@@ -50,14 +50,14 @@ This model demonstrates comprehensive capability enhancement using ICM-generated
|
|
50 |
### Dataset
|
51 |
- **Source**: [codelion/gemma-3-270m-icm-dpo](https://huggingface.co/datasets/codelion/gemma-3-270m-icm-dpo)
|
52 |
- **Method**: ICM (Internal Coherence Maximization) for label-free preference generation
|
53 |
-
- **Training Samples**:
|
54 |
- **Evaluation Samples**: 50
|
55 |
|
56 |
### Training Configuration
|
57 |
- **Epochs**: 1
|
58 |
-
- **Batch Size**:
|
59 |
- **Gradient Accumulation**: 8 steps
|
60 |
-
- **Effective Batch Size**:
|
61 |
- **Learning Rate**: 2e-07
|
62 |
- **Optimizer**: paged_adamw_8bit
|
63 |
- **Memory Optimization**: BF16, Gradient Checkpointing
|
|
|
50 |
### Dataset
|
51 |
- **Source**: [codelion/gemma-3-270m-icm-dpo](https://huggingface.co/datasets/codelion/gemma-3-270m-icm-dpo)
|
52 |
- **Method**: ICM (Internal Coherence Maximization) for label-free preference generation
|
53 |
+
- **Training Samples**: 1812
|
54 |
- **Evaluation Samples**: 50
|
55 |
|
56 |
### Training Configuration
|
57 |
- **Epochs**: 1
|
58 |
+
- **Batch Size**: 1 (per device)
|
59 |
- **Gradient Accumulation**: 8 steps
|
60 |
+
- **Effective Batch Size**: 8
|
61 |
- **Learning Rate**: 2e-07
|
62 |
- **Optimizer**: paged_adamw_8bit
|
63 |
- **Memory Optimization**: BF16, Gradient Checkpointing
|
adapter_config.json
CHANGED
@@ -28,13 +28,13 @@
|
|
28 |
"rank_pattern": {},
|
29 |
"revision": null,
|
30 |
"target_modules": [
|
31 |
-
"
|
32 |
-
"q_proj",
|
33 |
-
"v_proj",
|
34 |
"up_proj",
|
35 |
"k_proj",
|
|
|
36 |
"down_proj",
|
37 |
-
"
|
|
|
38 |
],
|
39 |
"target_parameters": null,
|
40 |
"task_type": "CAUSAL_LM",
|
|
|
28 |
"rank_pattern": {},
|
29 |
"revision": null,
|
30 |
"target_modules": [
|
31 |
+
"gate_proj",
|
|
|
|
|
32 |
"up_proj",
|
33 |
"k_proj",
|
34 |
+
"v_proj",
|
35 |
"down_proj",
|
36 |
+
"o_proj",
|
37 |
+
"q_proj"
|
38 |
],
|
39 |
"target_parameters": null,
|
40 |
"task_type": "CAUSAL_LM",
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 701497992
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8d9fbccbcd31bc8df08fed608b79046818801f5a92ce6537aee6cb7d2b10c085
|
3 |
size 701497992
|
added_tokens.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"<image_soft_token>": 262144
|
3 |
+
}
|
results.json
CHANGED
@@ -1,10 +1,10 @@
|
|
1 |
{
|
2 |
"training_metrics": {
|
3 |
-
"train_runtime":
|
4 |
-
"train_samples_per_second":
|
5 |
-
"train_steps_per_second": 0.
|
6 |
"total_flos": 0.0,
|
7 |
-
"train_loss": 0.
|
8 |
"epoch": 1.0
|
9 |
},
|
10 |
"config": {
|
@@ -31,7 +31,7 @@
|
|
31 |
"beta": 0.02,
|
32 |
"max_length": 1024,
|
33 |
"max_prompt_length": 512,
|
34 |
-
"batch_size":
|
35 |
"gradient_accumulation_steps": 8,
|
36 |
"learning_rate": 2e-07,
|
37 |
"num_train_epochs": 1,
|
|
|
1 |
{
|
2 |
"training_metrics": {
|
3 |
+
"train_runtime": 1035.1183,
|
4 |
+
"train_samples_per_second": 1.751,
|
5 |
+
"train_steps_per_second": 0.219,
|
6 |
"total_flos": 0.0,
|
7 |
+
"train_loss": 0.6929902463232368,
|
8 |
"epoch": 1.0
|
9 |
},
|
10 |
"config": {
|
|
|
31 |
"beta": 0.02,
|
32 |
"max_length": 1024,
|
33 |
"max_prompt_length": 512,
|
34 |
+
"batch_size": 1,
|
35 |
"gradient_accumulation_steps": 8,
|
36 |
"learning_rate": 2e-07,
|
37 |
"num_train_epochs": 1,
|
tokenizer.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1299c11d7cf632ef3b4e11937501358ada021bbdf7c47638d13c0ee982f2e79c
|
3 |
+
size 4689074
|
training_config.json
CHANGED
@@ -22,7 +22,7 @@
|
|
22 |
"beta": 0.02,
|
23 |
"max_length": 1024,
|
24 |
"max_prompt_length": 512,
|
25 |
-
"batch_size":
|
26 |
"gradient_accumulation_steps": 8,
|
27 |
"learning_rate": 2e-07,
|
28 |
"num_train_epochs": 1,
|
|
|
22 |
"beta": 0.02,
|
23 |
"max_length": 1024,
|
24 |
"max_prompt_length": 512,
|
25 |
+
"batch_size": 1,
|
26 |
"gradient_accumulation_steps": 8,
|
27 |
"learning_rate": 2e-07,
|
28 |
"num_train_epochs": 1,
|