End of training
Browse files- README.md +2 -2
- config.json +1 -1
- generation_config.json +1 -1
- model.safetensors +1 -1
- runs/Jan21_07-08-35_rft-0/events.out.tfevents.1737443316.rft-0.52468.0 +3 -0
- runs/Jan21_07-10-23_rft-0/events.out.tfevents.1737443423.rft-0.55658.0 +3 -0
- tokenizer.json +1 -6
- tokenizer_config.json +1 -0
- training_args.bin +2 -2
README.md
CHANGED
@@ -36,10 +36,10 @@ This model was trained with SFT.
|
|
36 |
### Framework versions
|
37 |
|
38 |
- TRL: 0.12.1
|
39 |
-
- Transformers: 4.
|
40 |
- Pytorch: 2.5.1
|
41 |
- Datasets: 3.1.0
|
42 |
-
- Tokenizers: 0.
|
43 |
|
44 |
## Citations
|
45 |
|
|
|
36 |
### Framework versions
|
37 |
|
38 |
- TRL: 0.12.1
|
39 |
+
- Transformers: 4.48.1
|
40 |
- Pytorch: 2.5.1
|
41 |
- Datasets: 3.1.0
|
42 |
+
- Tokenizers: 0.21.0
|
43 |
|
44 |
## Citations
|
45 |
|
config.json
CHANGED
@@ -27,7 +27,7 @@
|
|
27 |
"rope_theta": 100000,
|
28 |
"tie_word_embeddings": true,
|
29 |
"torch_dtype": "float32",
|
30 |
-
"transformers_version": "4.
|
31 |
"use_cache": true,
|
32 |
"vocab_size": 49152
|
33 |
}
|
|
|
27 |
"rope_theta": 100000,
|
28 |
"tie_word_embeddings": true,
|
29 |
"torch_dtype": "float32",
|
30 |
+
"transformers_version": "4.48.1",
|
31 |
"use_cache": true,
|
32 |
"vocab_size": 49152
|
33 |
}
|
generation_config.json
CHANGED
@@ -3,5 +3,5 @@
|
|
3 |
"bos_token_id": 1,
|
4 |
"eos_token_id": 2,
|
5 |
"pad_token_id": 2,
|
6 |
-
"transformers_version": "4.
|
7 |
}
|
|
|
3 |
"bos_token_id": 1,
|
4 |
"eos_token_id": 2,
|
5 |
"pad_token_id": 2,
|
6 |
+
"transformers_version": "4.48.1"
|
7 |
}
|
model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 538090408
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:01a5e83afe9216e394601c17eaa1cc803e78c1ba65e2fe80aa0d7e2bfed91ab9
|
3 |
size 538090408
|
runs/Jan21_07-08-35_rft-0/events.out.tfevents.1737443316.rft-0.52468.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2ea2f00005c2d7244b14bfb1fa49641df96d19951f9af1ee1b7cc507e7859d85
|
3 |
+
size 8567
|
runs/Jan21_07-10-23_rft-0/events.out.tfevents.1737443423.rft-0.55658.0
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:264a41925ba6686845470b15c0bb330b192e54025a3cd943c969319d311e1abf
|
3 |
+
size 32367
|
tokenizer.json
CHANGED
@@ -1,11 +1,6 @@
|
|
1 |
{
|
2 |
"version": "1.0",
|
3 |
-
"truncation":
|
4 |
-
"direction": "Right",
|
5 |
-
"max_length": 1024,
|
6 |
-
"strategy": "LongestFirst",
|
7 |
-
"stride": 0
|
8 |
-
},
|
9 |
"padding": null,
|
10 |
"added_tokens": [
|
11 |
{
|
|
|
1 |
{
|
2 |
"version": "1.0",
|
3 |
+
"truncation": null,
|
|
|
|
|
|
|
|
|
|
|
4 |
"padding": null,
|
5 |
"added_tokens": [
|
6 |
{
|
tokenizer_config.json
CHANGED
@@ -146,6 +146,7 @@
|
|
146 |
"chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
147 |
"clean_up_tokenization_spaces": false,
|
148 |
"eos_token": "<|im_end|>",
|
|
|
149 |
"model_max_length": 8192,
|
150 |
"pad_token": "<|im_end|>",
|
151 |
"tokenizer_class": "GPT2Tokenizer",
|
|
|
146 |
"chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
147 |
"clean_up_tokenization_spaces": false,
|
148 |
"eos_token": "<|im_end|>",
|
149 |
+
"extra_special_tokens": {},
|
150 |
"model_max_length": 8192,
|
151 |
"pad_token": "<|im_end|>",
|
152 |
"tokenizer_class": "GPT2Tokenizer",
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e5cc7232eb61982f2e7634aae20d29ed56e684a0d1b1d5a10c28781cce1bd975
|
3 |
+
size 5624
|