sheepy928 commited on
Commit
0bc1fa4
·
verified ·
1 Parent(s): e39e7b0

Model save

Browse files
README.md CHANGED
@@ -1,11 +1,9 @@
1
  ---
2
  base_model: Qwen/Qwen2.5-1.5B-Instruct
3
- datasets: HuggingFaceH4/Bespoke-Stratos-17k
4
  library_name: transformers
5
  model_name: qwen2.5_1.5b_ins-pt-bf16
6
  tags:
7
  - generated_from_trainer
8
- - open-r1
9
  - trl
10
  - sft
11
  licence: license
@@ -13,7 +11,7 @@ licence: license
13
 
14
  # Model Card for qwen2.5_1.5b_ins-pt-bf16
15
 
16
- This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct) on the [HuggingFaceH4/Bespoke-Stratos-17k](https://huggingface.co/datasets/HuggingFaceH4/Bespoke-Stratos-17k) dataset.
17
  It has been trained using [TRL](https://github.com/huggingface/trl).
18
 
19
  ## Quick start
@@ -29,7 +27,7 @@ print(output["generated_text"])
29
 
30
  ## Training procedure
31
 
32
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/sheepy928/r1fp4/runs/a00adzvk)
33
 
34
 
35
  This model was trained with SFT.
 
1
  ---
2
  base_model: Qwen/Qwen2.5-1.5B-Instruct
 
3
  library_name: transformers
4
  model_name: qwen2.5_1.5b_ins-pt-bf16
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
  - sft
9
  licence: license
 
11
 
12
  # Model Card for qwen2.5_1.5b_ins-pt-bf16
13
 
14
+ This model is a fine-tuned version of [Qwen/Qwen2.5-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-1.5B-Instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
27
 
28
  ## Training procedure
29
 
30
+
31
 
32
 
33
  This model was trained with SFT.
all_results.json CHANGED
@@ -5,9 +5,9 @@
5
  "eval_samples_per_second": 7.147,
6
  "eval_steps_per_second": 0.893,
7
  "total_flos": 768626978193408.0,
8
- "train_loss": 2.0993992060908564,
9
- "train_runtime": 74825.3444,
10
- "train_samples": 16610,
11
- "train_samples_per_second": 2.888,
12
- "train_steps_per_second": 0.09
13
  }
 
5
  "eval_samples_per_second": 7.147,
6
  "eval_steps_per_second": 0.893,
7
  "total_flos": 768626978193408.0,
8
+ "train_loss": 0.0,
9
+ "train_runtime": 1.3815,
10
+ "train_samples": 1000,
11
+ "train_samples_per_second": 2424.882,
12
+ "train_steps_per_second": 152.008
13
  }
config.json CHANGED
@@ -23,7 +23,7 @@
23
  "tie_word_embeddings": true,
24
  "torch_dtype": "bfloat16",
25
  "transformers_version": "4.49.0.dev0",
26
- "use_cache": true,
27
  "use_sliding_window": false,
28
  "vocab_size": 151665
29
  }
 
23
  "tie_word_embeddings": true,
24
  "torch_dtype": "bfloat16",
25
  "transformers_version": "4.49.0.dev0",
26
+ "use_cache": false,
27
  "use_sliding_window": false,
28
  "vocab_size": 151665
29
  }
train_results.json CHANGED
@@ -1,8 +1,8 @@
1
  {
2
  "total_flos": 768626978193408.0,
3
- "train_loss": 2.0993992060908564,
4
- "train_runtime": 74825.3444,
5
- "train_samples": 16610,
6
- "train_samples_per_second": 2.888,
7
- "train_steps_per_second": 0.09
8
  }
 
1
  {
2
  "total_flos": 768626978193408.0,
3
+ "train_loss": 0.0,
4
+ "train_runtime": 1.3815,
5
+ "train_samples": 1000,
6
+ "train_samples_per_second": 2424.882,
7
+ "train_steps_per_second": 152.008
8
  }
trainer_state.json CHANGED
@@ -571,17 +571,16 @@
571
  },
572
  {
573
  "epoch": 9.985936343449296,
574
- "mean_token_accuracy": 0.638465811965812,
575
  "step": 6750,
576
  "total_flos": 768626978193408.0,
577
- "train_loss": 2.0993992060908564,
578
- "train_runtime": 74825.3444,
579
- "train_samples_per_second": 2.888,
580
- "train_steps_per_second": 0.09
581
  }
582
  ],
583
  "logging_steps": 200,
584
- "max_steps": 6750,
585
  "num_input_tokens_seen": 0,
586
  "num_train_epochs": 10,
587
  "save_steps": 500,
 
571
  },
572
  {
573
  "epoch": 9.985936343449296,
 
574
  "step": 6750,
575
  "total_flos": 768626978193408.0,
576
+ "train_loss": 0.0,
577
+ "train_runtime": 1.3815,
578
+ "train_samples_per_second": 2424.882,
579
+ "train_steps_per_second": 152.008
580
  }
581
  ],
582
  "logging_steps": 200,
583
+ "max_steps": 210,
584
  "num_input_tokens_seen": 0,
585
  "num_train_epochs": 10,
586
  "save_steps": 500,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4f17d98c150bdf9c737a07533dad8e08c6b61155597b70938a02299b0ad1a6d4
3
- size 7160
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:63001b5aa7df486aef58ad93dc19fc01e6610c1d3a104d0728864438cdf62595
3
+ size 5816