narekvslife commited on
Commit
b8cad6c
·
verified ·
1 Parent(s): 80fd6b0

dpo_79pt692h

Browse files
README.md CHANGED
@@ -17,15 +17,15 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  This model was trained from scratch on the None dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.6736
21
- - Rewards/chosen: 0.5769
22
- - Rewards/rejected: 0.5039
23
- - Rewards/accuracies: 0.6000
24
- - Rewards/margins: 0.0730
25
- - Logps/rejected: -171.1358
26
- - Logps/chosen: -181.4562
27
- - Logits/rejected: -0.0987
28
- - Logits/chosen: -0.0683
29
 
30
  ## Model description
31
 
@@ -48,12 +48,10 @@ The following hyperparameters were used during training:
48
  - train_batch_size: 1
49
  - eval_batch_size: 1
50
  - seed: 0
51
- - gradient_accumulation_steps: 2
52
- - total_train_batch_size: 2
53
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
54
  - lr_scheduler_type: cosine
55
  - lr_scheduler_warmup_steps: 100
56
- - training_steps: 250
57
 
58
  ### Training results
59
 
 
17
 
18
  This model was trained from scratch on the None dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.6931
21
+ - Rewards/chosen: 0.0
22
+ - Rewards/rejected: 0.0
23
+ - Rewards/accuracies: 0.0
24
+ - Rewards/margins: 0.0
25
+ - Logps/rejected: -32.6347
26
+ - Logps/chosen: -35.1065
27
+ - Logits/rejected: -1.1843
28
+ - Logits/chosen: -1.2320
29
 
30
  ## Model description
31
 
 
48
  - train_batch_size: 1
49
  - eval_batch_size: 1
50
  - seed: 0
 
 
51
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
52
  - lr_scheduler_type: cosine
53
  - lr_scheduler_warmup_steps: 100
54
+ - training_steps: 2
55
 
56
  ### Training results
57
 
adapter_config.json CHANGED
@@ -16,17 +16,17 @@
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
- "r": 256,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "q_proj",
24
- "k_proj",
25
- "v_proj",
26
  "wte",
27
- "out_proj",
28
- "fc_in",
29
- "fc_out"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
16
  "megatron_core": "megatron.core",
17
  "modules_to_save": null,
18
  "peft_type": "LORA",
19
+ "r": 16,
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
+ "out_proj",
24
  "q_proj",
 
 
25
  "wte",
26
+ "fc_out",
27
+ "v_proj",
28
+ "k_proj",
29
+ "fc_in"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:85dfe4c31ce75352ba7da7625b5006dc78d1a2f79e0b277aecc078d20f5b0d6b
3
- size 302016416
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ab0fda06fcdd9a8e16d9a236718f37697611b438e210d287c0a0fcbfea98ee7
3
+ size 18900240
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b3c66e9158b2f49b9b0b295f6b1b5de6c5745a63fb6f87dc5e502958055b3947
3
- size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6177b9cb5727cbe1c1928de9d1c03adfeb642b365c0c8718daea0f8e4049eb85
3
+ size 5560