ahadda5 commited on
Commit
406acb8
·
1 Parent(s): 1b199ca

Upload tokenizer_config.json

Browse files
Files changed (1) hide show
  1. tokenizer_config.json +80 -0
tokenizer_config.json ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "additional_special_tokens": [
4
+ "<present>",
5
+ "<absent>",
6
+ "<category>",
7
+ "<infill>",
8
+ "<seealso>",
9
+ "<header>",
10
+ "<|endoftext|>",
11
+ "<sep>",
12
+ "<mask>",
13
+ "<mixed>",
14
+ "<number>",
15
+ "<phrase>"
16
+ ],
17
+ "bos_token": {
18
+ "__type": "AddedToken",
19
+ "content": "<s>",
20
+ "lstrip": false,
21
+ "normalized": true,
22
+ "rstrip": false,
23
+ "single_word": false
24
+ },
25
+ "cls_token": {
26
+ "__type": "AddedToken",
27
+ "content": "<s>",
28
+ "lstrip": false,
29
+ "normalized": true,
30
+ "rstrip": false,
31
+ "single_word": false
32
+ },
33
+ "eos_token": {
34
+ "__type": "AddedToken",
35
+ "content": "</s>",
36
+ "lstrip": false,
37
+ "normalized": true,
38
+ "rstrip": false,
39
+ "single_word": false
40
+ },
41
+ "errors": "replace",
42
+ "mask_token": {
43
+ "__type": "AddedToken",
44
+ "content": "<mask>",
45
+ "lstrip": true,
46
+ "normalized": true,
47
+ "rstrip": false,
48
+ "single_word": false
49
+ },
50
+ "model_max_length": 512,
51
+ "name_or_path": "ahadda5/bart_wikikp_kp20k_openkp",
52
+ "pad_token": {
53
+ "__type": "AddedToken",
54
+ "content": "<pad>",
55
+ "lstrip": false,
56
+ "normalized": true,
57
+ "rstrip": false,
58
+ "single_word": false
59
+ },
60
+ "sep": "<sep>",
61
+ "sep_token": {
62
+ "__type": "AddedToken",
63
+ "content": "<sep>",
64
+ "lstrip": false,
65
+ "normalized": true,
66
+ "rstrip": false,
67
+ "single_word": false
68
+ },
69
+ "special_tokens_map_file": "/home/ashraf.haddad/.cache/huggingface/transformers/4b7a3619321a39f6780e0c775802e3523e52c1efd3a46e6d1baac9e1e8e234e6.898eb95aac9bb57440c2f57caa963ae18b9b10ba4731cc81020283286b0391fc",
70
+ "tokenizer_class": "BartTokenizer",
71
+ "trim_offsets": true,
72
+ "unk_token": {
73
+ "__type": "AddedToken",
74
+ "content": "<unk>",
75
+ "lstrip": false,
76
+ "normalized": true,
77
+ "rstrip": false,
78
+ "single_word": false
79
+ }
80
+ }