narekvslife commited on
Commit
f5265ca
·
verified ·
1 Parent(s): cda0936

dpo_4rbrxhkg

Browse files
README.md CHANGED
@@ -17,15 +17,15 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model was trained from scratch on the None dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.6931
21
- - Rewards/chosen: 0.0
22
- - Rewards/rejected: 0.0
23
- - Rewards/accuracies: 0.0
24
- - Rewards/margins: 0.0
25
- - Logps/rejected: -32.6347
26
- - Logps/chosen: -35.1065
27
- - Logits/rejected: -1.1843
28
- - Logits/chosen: -1.2320
29
 
30
  ## Model description
31
 
@@ -44,14 +44,14 @@ More information needed
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
47
- - learning_rate: 5e-05
48
  - train_batch_size: 1
49
  - eval_batch_size: 1
50
  - seed: 0
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: cosine
53
  - lr_scheduler_warmup_steps: 100
54
- - training_steps: 2
55
 
56
  ### Training results
57
 
 
17
 
18
  This model was trained from scratch on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.7021
21
+ - Rewards/chosen: 0.5480
22
+ - Rewards/rejected: 0.4854
23
+ - Rewards/accuracies: 0.1242
24
+ - Rewards/margins: 0.0626
25
+ - Logps/rejected: -26.1813
26
+ - Logps/chosen: -28.6110
27
+ - Logits/rejected: -1.0475
28
+ - Logits/chosen: -1.0554
29
 
30
  ## Model description
31
 
 
44
  ### Training hyperparameters
45
 
46
  The following hyperparameters were used during training:
47
+ - learning_rate: 0.0001
48
  - train_batch_size: 1
49
  - eval_batch_size: 1
50
  - seed: 0
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: cosine
53
  - lr_scheduler_warmup_steps: 100
54
+ - training_steps: 2500
55
 
56
  ### Training results
57
 
adapter_config.json CHANGED
@@ -16,17 +16,17 @@
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
- "r": 16,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "out_proj",
24
- "q_proj",
25
- "fc_in",
26
  "k_proj",
27
- "v_proj",
28
  "wte",
29
- "fc_out"
 
 
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
+ "r": 64,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
23
  "k_proj",
 
24
  "wte",
25
+ "q_proj",
26
+ "fc_out",
27
+ "out_proj",
28
+ "v_proj",
29
+ "fc_in"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5ab0fda06fcdd9a8e16d9a236718f37697611b438e210d287c0a0fcbfea98ee7
3
- size 18900240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37c713439897aefab1d63b1f1938ec6a3ff714264132976bebe1d9780fd5dfe3
3
+ size 75523504
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d13602293bb6f8bb01001bbab0f60f58b3113aee181e0ee084d00eb02d9a39aa
3
  size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b116eedb32c45e3728095b6935fbdfd4c2aa57a3764d638aa33e5a1d3237a04
3
  size 5560