Upload tokenizer_config.json with huggingface_hub
Browse files- tokenizer_config.json +1 -0
tokenizer_config.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"word_index": {"<OOV>": 1, "the": 2, "is": 3, "what": 4, "capital": 5, "of": 6, "france": 7, "continents": 8, "are": 9, "there": 10, "largest": 11, "mammal": 12, "wrote": 13, "hamlet": 14, "how": 15, "many": 16, "who": 17, "play": 18, "paris": 19, "seven": 20, "blue": 21, "whale": 22, "william": 23, "shakespeare": 24}, "index_word": {"1": "<OOV>", "2": "the", "3": "is", "4": "what", "5": "capital", "6": "of", "7": "france", "8": "continents", "9": "are", "10": "there", "11": "largest", "12": "mammal", "13": "wrote", "14": "hamlet", "15": "how", "16": "many", "17": "who", "18": "play", "19": "paris", "20": "seven", "21": "blue", "22": "whale", "23": "william", "24": "shakespeare"}, "num_words": 10000, "filters": "!\"#$%&()*+,-./:;<=>?@[\\]^_`{|}~\t\n", "lower": true, "split": " ", "char_level": false}
|