AlexHT_Hung
commited on
Commit
·
5db4fb4
1
Parent(s):
1202640
remove dummies
Browse files- added_tokens.json +2 -2
- tokenizer.json +3 -29
- tokenizer_config.json +2 -2
added_tokens.json
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
{
|
2 |
-
"<|func_end|>":
|
3 |
-
"<|func_start|>":
|
4 |
}
|
|
|
1 |
{
|
2 |
+
"<|func_end|>": 35686,
|
3 |
+
"<|func_start|>": 35687
|
4 |
}
|
tokenizer.json
CHANGED
@@ -31,7 +31,7 @@
|
|
31 |
"special": true
|
32 |
},
|
33 |
{
|
34 |
-
"id":
|
35 |
"content": "<|func_end|>",
|
36 |
"single_word": true,
|
37 |
"lstrip": true,
|
@@ -40,7 +40,7 @@
|
|
40 |
"special": false
|
41 |
},
|
42 |
{
|
43 |
-
"id":
|
44 |
"content": "<|func_start|>",
|
45 |
"single_word": true,
|
46 |
"lstrip": true,
|
@@ -35810,33 +35810,7 @@
|
|
35810 |
"賅": 35682,
|
35811 |
"簞": 35683,
|
35812 |
"鼴": 35684,
|
35813 |
-
"躂": 35685
|
35814 |
-
"<DUMMY_0>": 35686,
|
35815 |
-
"<DUMMY_1>": 35687,
|
35816 |
-
"<DUMMY_2>": 35688,
|
35817 |
-
"<DUMMY_3>": 35689,
|
35818 |
-
"<DUMMY_4>": 35690,
|
35819 |
-
"<DUMMY_5>": 35691,
|
35820 |
-
"<DUMMY_6>": 35692,
|
35821 |
-
"<DUMMY_7>": 35693,
|
35822 |
-
"<DUMMY_8>": 35694,
|
35823 |
-
"<DUMMY_9>": 35695,
|
35824 |
-
"<DUMMY_10>": 35696,
|
35825 |
-
"<DUMMY_11>": 35697,
|
35826 |
-
"<DUMMY_12>": 35698,
|
35827 |
-
"<DUMMY_13>": 35699,
|
35828 |
-
"<DUMMY_14>": 35700,
|
35829 |
-
"<DUMMY_15>": 35701,
|
35830 |
-
"<DUMMY_16>": 35702,
|
35831 |
-
"<DUMMY_17>": 35703,
|
35832 |
-
"<DUMMY_18>": 35704,
|
35833 |
-
"<DUMMY_19>": 35705,
|
35834 |
-
"<DUMMY_20>": 35706,
|
35835 |
-
"<DUMMY_21>": 35707,
|
35836 |
-
"<DUMMY_22>": 35708,
|
35837 |
-
"<DUMMY_23>": 35709,
|
35838 |
-
"<DUMMY_24>": 35710,
|
35839 |
-
"<DUMMY_25>": 35711
|
35840 |
},
|
35841 |
"merges": [
|
35842 |
"▁ t",
|
|
|
31 |
"special": true
|
32 |
},
|
33 |
{
|
34 |
+
"id": 35686,
|
35 |
"content": "<|func_end|>",
|
36 |
"single_word": true,
|
37 |
"lstrip": true,
|
|
|
40 |
"special": false
|
41 |
},
|
42 |
{
|
43 |
+
"id": 35687,
|
44 |
"content": "<|func_start|>",
|
45 |
"single_word": true,
|
46 |
"lstrip": true,
|
|
|
35810 |
"賅": 35682,
|
35811 |
"簞": 35683,
|
35812 |
"鼴": 35684,
|
35813 |
+
"躂": 35685
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
35814 |
},
|
35815 |
"merges": [
|
35816 |
"▁ t",
|
tokenizer_config.json
CHANGED
@@ -26,7 +26,7 @@
|
|
26 |
"single_word": false,
|
27 |
"special": true
|
28 |
},
|
29 |
-
"
|
30 |
"content": "<|func_end|>",
|
31 |
"lstrip": true,
|
32 |
"normalized": false,
|
@@ -34,7 +34,7 @@
|
|
34 |
"single_word": true,
|
35 |
"special": false
|
36 |
},
|
37 |
-
"
|
38 |
"content": "<|func_start|>",
|
39 |
"lstrip": true,
|
40 |
"normalized": false,
|
|
|
26 |
"single_word": false,
|
27 |
"special": true
|
28 |
},
|
29 |
+
"35686": {
|
30 |
"content": "<|func_end|>",
|
31 |
"lstrip": true,
|
32 |
"normalized": false,
|
|
|
34 |
"single_word": true,
|
35 |
"special": false
|
36 |
},
|
37 |
+
"35687": {
|
38 |
"content": "<|func_start|>",
|
39 |
"lstrip": true,
|
40 |
"normalized": false,
|