ayjays132 commited on
Commit
9ae7835
·
verified ·
1 Parent(s): 473d810

Upload 11 files

Browse files
config.json CHANGED
@@ -1,154 +1,158 @@
1
- {
2
- "_name_or_path": "ayjays132/CustomGPT2Conversational",
3
- "activation_function": "gelu_new",
4
- "advanced_model_options": {
5
- "contextual_embeddings": {
6
- "approaches": [
7
- "contextual_attention_mechanisms",
8
- "semantic_embedding_regularization"
9
- ],
10
- "enable": true
11
- },
12
- "dynamic_adaptation": {
13
- "enable": true,
14
- "techniques": [
15
- "adaptive_layer_dropping",
16
- "dynamic_context_window"
17
- ]
18
- },
19
- "innovative_neuron_growth": {
20
- "enable": true,
21
- "strategies": [
22
- "selective_neuron_pruning",
23
- "progressive_neuron_expansion"
24
- ]
25
- },
26
- "memory_optimization": {
27
- "enable": true,
28
- "methods": [
29
- "gradient_checkpointing",
30
- "memory-efficient_attention"
31
- ]
32
- },
33
- "meta_learning": {
34
- "approaches": [
35
- "meta_learning_rate_adjustment",
36
- "online_adaptation"
37
- ],
38
- "enable": true
39
- },
40
- "secret_advanced_options": {
41
- "adaptive_token_embedding": {
42
- "enable": true,
43
- "strategies": [
44
- "dynamic_embedding_resizing",
45
- "contextual_embedding_scaling"
46
- ]
47
- },
48
- "future_context_prediction": {
49
- "enable": true,
50
- "techniques": [
51
- "lookahead_context_integration",
52
- "predictive_attention_mechanisms"
53
- ]
54
- },
55
- "multi_modal_integration": {
56
- "enable": true,
57
- "methods": [
58
- "text_image_alignment",
59
- "cross_modal_attention"
60
- ]
61
- }
62
- }
63
- },
64
- "architectures": [
65
- "GPT2LMHeadModel"
66
- ],
67
- "attn_pdrop": 0.1,
68
- "bos_token_id": 50267,
69
- "context_window": 20,
70
- "contextual_embedding_dim": 1024,
71
- "device": "cuda",
72
- "dropout_rate": 0.1,
73
- "early_stopping": true,
74
- "embd_pdrop": 0.1,
75
- "embedding_dim": 1024,
76
- "eos_token_id": 50267,
77
- "hidden_dim": 1024,
78
- "initializer_range": 0.02,
79
- "innovative_growth_capacity": 50000,
80
- "integration_settings": {
81
- "config_name": "config.json",
82
- "load_from_transformers": true,
83
- "pytorch_dump_folder_path": "./model_save",
84
- "pytorch_model_bin_name": "pytorch_model.bin"
85
- },
86
- "layer_norm_epsilon": 1e-05,
87
- "max_length": 1024,
88
- "max_memory_size": 100000,
89
- "max_neurons": 100,
90
- "meta_learning_rate": 0.001,
91
- "min_length": 50,
92
- "model_type": "gpt2",
93
- "n_ctx": 1024,
94
- "n_embd": 1024,
95
- "n_head": 16,
96
- "n_inner": null,
97
- "n_layer": 24,
98
- "n_positions": 1024,
99
- "no_repeat_ngram_size": 2,
100
- "num_beams": 5,
101
- "num_embeddings": 50257,
102
- "num_heads": 64,
103
- "num_layers": 24,
104
- "output_attentions": true,
105
- "output_hidden_states": true,
106
- "pad_token_id": 50267,
107
- "reorder_and_upcast_attn": false,
108
- "resid_pdrop": 0.1,
109
- "scale_attn_by_inverse_layer_idx": false,
110
- "scale_attn_weights": true,
111
- "sep_token_id": 50267,
112
- "special_tokens": {
113
- "additional_special_tokens": [
114
- "<greeting>",
115
- "<farewell>",
116
- "<thank>",
117
- "<apology>"
118
- ],
119
- "bos_token": "<bos>",
120
- "cls_token": "<cls>",
121
- "eos_token": "<eos>",
122
- "mask_token": "<mask>",
123
- "pad_token": "<pad>",
124
- "sep_token": "<sep>",
125
- "unk_token": "<unk>"
126
- },
127
- "state_shape": null,
128
- "summary_activation": null,
129
- "summary_first_dropout": 0.1,
130
- "summary_proj_to_labels": true,
131
- "summary_type": "cls_index",
132
- "summary_use_proj": true,
133
- "target_q_model": null,
134
- "task_specific_params": {
135
- "text-generation": {
136
- "do_sample": true,
137
- "early_stopping": true,
138
- "length_penalty": 1.0,
139
- "max_length": 2048,
140
- "min_length": 64,
141
- "no_repeat_ngram_size": 2,
142
- "num_beams": 8,
143
- "num_return_sequences": 3,
144
- "repetition_penalty": 1.2,
145
- "temperature": 0.9,
146
- "top_k": 50,
147
- "top_p": 0.95
148
- }
149
- },
150
- "torch_dtype": "float32",
151
- "transformers_version": "4.44.0",
152
- "use_cache": true,
153
- "vocab_size": 50257
154
- }
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "ayjays132/CustomGPT2Conversational",
3
+ "activation_function": "gelu_new",
4
+ "advanced_model_options": {
5
+ "contextual_embeddings": {
6
+ "approaches": [
7
+ "contextual_attention_mechanisms",
8
+ "semantic_embedding_regularization"
9
+ ],
10
+ "enable": true
11
+ },
12
+ "dynamic_adaptation": {
13
+ "enable": true,
14
+ "techniques": [
15
+ "adaptive_layer_dropping",
16
+ "dynamic_context_window"
17
+ ]
18
+ },
19
+ "innovative_neuron_growth": {
20
+ "enable": true,
21
+ "strategies": [
22
+ "selective_neuron_pruning",
23
+ "progressive_neuron_expansion"
24
+ ]
25
+ },
26
+ "memory_optimization": {
27
+ "enable": true,
28
+ "methods": [
29
+ "gradient_checkpointing",
30
+ "memory-efficient_attention"
31
+ ]
32
+ },
33
+ "meta_learning": {
34
+ "approaches": [
35
+ "meta_learning_rate_adjustment",
36
+ "online_adaptation"
37
+ ],
38
+ "enable": true
39
+ },
40
+ "secret_advanced_options": {
41
+ "adaptive_token_embedding": {
42
+ "enable": true,
43
+ "strategies": [
44
+ "dynamic_embedding_resizing",
45
+ "contextual_embedding_scaling"
46
+ ]
47
+ },
48
+ "future_context_prediction": {
49
+ "enable": true,
50
+ "techniques": [
51
+ "lookahead_context_integration",
52
+ "predictive_attention_mechanisms"
53
+ ]
54
+ },
55
+ "multi_modal_integration": {
56
+ "enable": true,
57
+ "methods": [
58
+ "text_image_alignment",
59
+ "cross_modal_attention"
60
+ ]
61
+ }
62
+ }
63
+ },
64
+ "architectures": [
65
+ "GPT2LMHeadModel"
66
+ ],
67
+ "attn_pdrop": 0.1,
68
+ "bos_token_id": 50267,
69
+ "context_window": 20,
70
+ "contextual_embedding_dim": 1024,
71
+ "device": "cuda",
72
+ "dropout_rate": 0.1,
73
+ "early_stopping": true,
74
+ "embd_pdrop": 0.1,
75
+ "embedding_dim": 1024,
76
+ "eos_token_id": 50267,
77
+ "hidden_dim": 1024,
78
+ "initializer_range": 0.02,
79
+ "innovative_growth_capacity": 50000,
80
+ "integration_settings": {
81
+ "config_name": "config.json",
82
+ "load_from_transformers": true,
83
+ "pytorch_dump_folder_path": "./model_save",
84
+ "pytorch_model_bin_name": "pytorch_model.bin"
85
+ },
86
+ "layer_norm_epsilon": 1e-05,
87
+ "max_length": 1024,
88
+ "max_memory_size": 100000,
89
+ "max_neurons": 100,
90
+ "meta_learning_rate": 0.001,
91
+ "min_length": 50,
92
+ "model_type": "gpt2",
93
+ "n_ctx": 1024,
94
+ "n_embd": 1024,
95
+ "n_head": 16,
96
+ "n_inner": null,
97
+ "n_layer": 24,
98
+ "n_positions": 1024,
99
+ "no_repeat_ngram_size": 2,
100
+ "num_beams": 5,
101
+ "num_embeddings": 50257,
102
+ "num_heads": 64,
103
+ "num_layers": 24,
104
+ "output_attentions": true,
105
+ "output_hidden_states": true,
106
+ "pad_token_id": 50267,
107
+ "planning_enabled": true,
108
+ "planning_layers": 2,
109
+ "reasoning_enabled": true,
110
+ "reasoning_layers": 2,
111
+ "reorder_and_upcast_attn": false,
112
+ "resid_pdrop": 0.1,
113
+ "scale_attn_by_inverse_layer_idx": false,
114
+ "scale_attn_weights": true,
115
+ "sep_token_id": 50267,
116
+ "special_tokens": {
117
+ "additional_special_tokens": [
118
+ "<greeting>",
119
+ "<farewell>",
120
+ "<thank>",
121
+ "<apology>"
122
+ ],
123
+ "bos_token": "<bos>",
124
+ "cls_token": "<cls>",
125
+ "eos_token": "<eos>",
126
+ "mask_token": "<mask>",
127
+ "pad_token": "<pad>",
128
+ "sep_token": "<sep>",
129
+ "unk_token": "<unk>"
130
+ },
131
+ "state_shape": null,
132
+ "summary_activation": null,
133
+ "summary_first_dropout": 0.1,
134
+ "summary_proj_to_labels": true,
135
+ "summary_type": "cls_index",
136
+ "summary_use_proj": true,
137
+ "target_q_model": null,
138
+ "task_specific_params": {
139
+ "text-generation": {
140
+ "do_sample": true,
141
+ "early_stopping": true,
142
+ "length_penalty": 1.0,
143
+ "max_length": 2048,
144
+ "min_length": 64,
145
+ "no_repeat_ngram_size": 2,
146
+ "num_beams": 8,
147
+ "num_return_sequences": 3,
148
+ "repetition_penalty": 1.2,
149
+ "temperature": 0.9,
150
+ "top_k": 50,
151
+ "top_p": 0.95
152
+ }
153
+ },
154
+ "torch_dtype": "float32",
155
+ "transformers_version": "4.44.2",
156
+ "use_cache": true,
157
+ "vocab_size": 50257
158
+ }
generation_config.json CHANGED
@@ -10,5 +10,5 @@
10
  "output_attentions": true,
11
  "output_hidden_states": true,
12
  "pad_token_id": 50256,
13
- "transformers_version": "4.44.0"
14
  }
 
10
  "output_attentions": true,
11
  "output_hidden_states": true,
12
  "pad_token_id": 50256,
13
+ "transformers_version": "4.44.2"
14
  }
hd_encoder.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8710306535dcfb1d1ff96a3bcb749e2876994c2eb7f2ac337b0a3f14c6231547
3
+ size 4204329
neuro_symbolic.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf0adeda3a74e513005adc532a5fdfd308f7d4c3743277ac5d70e00d58dc19f2
3
+ size 16792334
special_tokens_map.json CHANGED
@@ -1,25 +1,65 @@
1
  {
2
- "bos_token": {
3
- "content": "<|endoftext|>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "eos_token": {
10
- "content": "<|endoftext|>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "pad_token": {
17
- "content": "<|endoftext|>",
18
- "lstrip": false,
19
- "normalized": true,
20
- "rstrip": false,
21
- "single_word": false
22
- },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  "unk_token": {
24
  "content": "<|endoftext|>",
25
  "lstrip": false,
 
1
  {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<vision>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "<goal>",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ },
17
+ {
18
+ "content": "<strategy>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ {
25
+ "content": "<growth>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ },
31
+ {
32
+ "content": "<challenge>",
33
+ "lstrip": false,
34
+ "normalized": false,
35
+ "rstrip": false,
36
+ "single_word": false
37
+ },
38
+ {
39
+ "content": "<solution>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false
44
+ },
45
+ {
46
+ "content": "<discovery>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false
51
+ },
52
+ {
53
+ "content": "<insight>",
54
+ "lstrip": false,
55
+ "normalized": false,
56
+ "rstrip": false,
57
+ "single_word": false
58
+ }
59
+ ],
60
+ "bos_token": "<|endoftext|>",
61
+ "eos_token": "<|endoftext|>",
62
+ "pad_token": "<|endoftext|>",
63
  "unk_token": {
64
  "content": "<|endoftext|>",
65
  "lstrip": false,
token_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfdc626333fb06bdd23d3a4d053f5b6a7a44c46612e0d860bce2e31fae6adb69
3
+ size 1012
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:247905d7aff427d5340bfce3d0f339e5fc6a02849e5f07081c06ae00f06b59f4
3
- size 2114629
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20c6cae62868b366cf02757f57ba435d9332ccfb63a39c8d481feda15e426dee
3
+ size 2121005
tokenizer_config.json CHANGED
@@ -296,12 +296,295 @@
296
  "rstrip": false,
297
  "single_word": false,
298
  "special": false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
299
  }
300
  },
 
 
 
 
 
 
 
 
 
 
301
  "bos_token": "<|endoftext|>",
302
  "clean_up_tokenization_spaces": true,
303
  "eos_token": "<|endoftext|>",
304
  "max_length": 1024,
 
305
  "model_max_length": 1024,
306
  "pad_to_multiple_of": null,
307
  "pad_token": "<|endoftext|>",
@@ -309,8 +592,9 @@
309
  "padding": "max_length",
310
  "padding_side": "right",
311
  "stride": 0,
312
- "tokenizer_class": "GPT2Tokenizer",
313
  "truncation_side": "right",
314
  "truncation_strategy": "longest_first",
315
- "unk_token": "<|endoftext|>"
 
316
  }
 
296
  "rstrip": false,
297
  "single_word": false,
298
  "special": false
299
+ },
300
+ "50289": {
301
+ "content": "<ENTANGLE>",
302
+ "lstrip": false,
303
+ "normalized": false,
304
+ "rstrip": false,
305
+ "single_word": false,
306
+ "special": true
307
+ },
308
+ "50290": {
309
+ "content": "<SUPERPOSE>",
310
+ "lstrip": false,
311
+ "normalized": false,
312
+ "rstrip": false,
313
+ "single_word": false,
314
+ "special": true
315
+ },
316
+ "50291": {
317
+ "content": "<COLLAPSE>",
318
+ "lstrip": false,
319
+ "normalized": false,
320
+ "rstrip": false,
321
+ "single_word": false,
322
+ "special": true
323
+ },
324
+ "50292": {
325
+ "content": "<TEMPORAL_SHIFT>",
326
+ "lstrip": false,
327
+ "normalized": false,
328
+ "rstrip": false,
329
+ "single_word": false,
330
+ "special": true
331
+ },
332
+ "50293": {
333
+ "content": "<QUANTUM_MEMORY>",
334
+ "lstrip": false,
335
+ "normalized": false,
336
+ "rstrip": false,
337
+ "single_word": false,
338
+ "special": true
339
+ },
340
+ "50294": {
341
+ "content": "<RECURSIVE_START>",
342
+ "lstrip": false,
343
+ "normalized": false,
344
+ "rstrip": false,
345
+ "single_word": false,
346
+ "special": true
347
+ },
348
+ "50295": {
349
+ "content": "<RECURSIVE_END>",
350
+ "lstrip": false,
351
+ "normalized": false,
352
+ "rstrip": false,
353
+ "single_word": false,
354
+ "special": true
355
+ },
356
+ "50296": {
357
+ "content": "<FRACTAL_PATTERN>",
358
+ "lstrip": false,
359
+ "normalized": false,
360
+ "rstrip": false,
361
+ "single_word": false,
362
+ "special": true
363
+ },
364
+ "50297": {
365
+ "content": "<EMERGENT_STATE>",
366
+ "lstrip": false,
367
+ "normalized": false,
368
+ "rstrip": false,
369
+ "single_word": false,
370
+ "special": true
371
+ },
372
+ "50298": {
373
+ "content": "<happy>",
374
+ "lstrip": false,
375
+ "normalized": false,
376
+ "rstrip": false,
377
+ "single_word": false,
378
+ "special": true
379
+ },
380
+ "50299": {
381
+ "content": "<sad>",
382
+ "lstrip": false,
383
+ "normalized": false,
384
+ "rstrip": false,
385
+ "single_word": false,
386
+ "special": true
387
+ },
388
+ "50300": {
389
+ "content": "<angry>",
390
+ "lstrip": false,
391
+ "normalized": false,
392
+ "rstrip": false,
393
+ "single_word": false,
394
+ "special": true
395
+ },
396
+ "50301": {
397
+ "content": "<surprised>",
398
+ "lstrip": false,
399
+ "normalized": false,
400
+ "rstrip": false,
401
+ "single_word": false,
402
+ "special": true
403
+ },
404
+ "50302": {
405
+ "content": "<confused>",
406
+ "lstrip": false,
407
+ "normalized": false,
408
+ "rstrip": false,
409
+ "single_word": false,
410
+ "special": true
411
+ },
412
+ "50303": {
413
+ "content": "<excited>",
414
+ "lstrip": false,
415
+ "normalized": false,
416
+ "rstrip": false,
417
+ "single_word": false,
418
+ "special": true
419
+ },
420
+ "50304": {
421
+ "content": "<neutral>",
422
+ "lstrip": false,
423
+ "normalized": false,
424
+ "rstrip": false,
425
+ "single_word": false,
426
+ "special": true
427
+ },
428
+ "50305": {
429
+ "content": "<fear>",
430
+ "lstrip": false,
431
+ "normalized": false,
432
+ "rstrip": false,
433
+ "single_word": false,
434
+ "special": true
435
+ },
436
+ "50306": {
437
+ "content": "<disgust>",
438
+ "lstrip": false,
439
+ "normalized": false,
440
+ "rstrip": false,
441
+ "single_word": false,
442
+ "special": true
443
+ },
444
+ "50307": {
445
+ "content": "<trust>",
446
+ "lstrip": false,
447
+ "normalized": false,
448
+ "rstrip": false,
449
+ "single_word": false,
450
+ "special": true
451
+ },
452
+ "50308": {
453
+ "content": "<anticipation>",
454
+ "lstrip": false,
455
+ "normalized": false,
456
+ "rstrip": false,
457
+ "single_word": false,
458
+ "special": true
459
+ },
460
+ "50309": {
461
+ "content": "<joy>",
462
+ "lstrip": false,
463
+ "normalized": false,
464
+ "rstrip": false,
465
+ "single_word": false,
466
+ "special": true
467
+ },
468
+ "50310": {
469
+ "content": "<QUANTUM_EMERGENCY>",
470
+ "lstrip": false,
471
+ "normalized": false,
472
+ "rstrip": false,
473
+ "single_word": false,
474
+ "special": true
475
+ },
476
+ "50311": {
477
+ "content": "<QUANTUM_CRITICAL>",
478
+ "lstrip": false,
479
+ "normalized": false,
480
+ "rstrip": false,
481
+ "single_word": false,
482
+ "special": true
483
+ },
484
+ "50312": {
485
+ "content": "<QUANTUM_ALERT>",
486
+ "lstrip": false,
487
+ "normalized": false,
488
+ "rstrip": false,
489
+ "single_word": false,
490
+ "special": true
491
+ },
492
+ "50313": {
493
+ "content": "<TEMPORAL_EMERGENCY>",
494
+ "lstrip": false,
495
+ "normalized": false,
496
+ "rstrip": false,
497
+ "single_word": false,
498
+ "special": true
499
+ },
500
+ "50314": {
501
+ "content": "<ENTANGLED_CRITICAL>",
502
+ "lstrip": false,
503
+ "normalized": false,
504
+ "rstrip": false,
505
+ "single_word": false,
506
+ "special": true
507
+ },
508
+ "50315": {
509
+ "content": "<vision>",
510
+ "lstrip": false,
511
+ "normalized": false,
512
+ "rstrip": false,
513
+ "single_word": false,
514
+ "special": true
515
+ },
516
+ "50316": {
517
+ "content": "<goal>",
518
+ "lstrip": false,
519
+ "normalized": false,
520
+ "rstrip": false,
521
+ "single_word": false,
522
+ "special": true
523
+ },
524
+ "50317": {
525
+ "content": "<strategy>",
526
+ "lstrip": false,
527
+ "normalized": false,
528
+ "rstrip": false,
529
+ "single_word": false,
530
+ "special": true
531
+ },
532
+ "50318": {
533
+ "content": "<growth>",
534
+ "lstrip": false,
535
+ "normalized": false,
536
+ "rstrip": false,
537
+ "single_word": false,
538
+ "special": true
539
+ },
540
+ "50319": {
541
+ "content": "<challenge>",
542
+ "lstrip": false,
543
+ "normalized": false,
544
+ "rstrip": false,
545
+ "single_word": false,
546
+ "special": true
547
+ },
548
+ "50320": {
549
+ "content": "<solution>",
550
+ "lstrip": false,
551
+ "normalized": false,
552
+ "rstrip": false,
553
+ "single_word": false,
554
+ "special": true
555
+ },
556
+ "50321": {
557
+ "content": "<discovery>",
558
+ "lstrip": false,
559
+ "normalized": false,
560
+ "rstrip": false,
561
+ "single_word": false,
562
+ "special": true
563
+ },
564
+ "50322": {
565
+ "content": "<insight>",
566
+ "lstrip": false,
567
+ "normalized": false,
568
+ "rstrip": false,
569
+ "single_word": false,
570
+ "special": true
571
  }
572
  },
573
+ "additional_special_tokens": [
574
+ "<vision>",
575
+ "<goal>",
576
+ "<strategy>",
577
+ "<growth>",
578
+ "<challenge>",
579
+ "<solution>",
580
+ "<discovery>",
581
+ "<insight>"
582
+ ],
583
  "bos_token": "<|endoftext|>",
584
  "clean_up_tokenization_spaces": true,
585
  "eos_token": "<|endoftext|>",
586
  "max_length": 1024,
587
+ "merges_file": null,
588
  "model_max_length": 1024,
589
  "pad_to_multiple_of": null,
590
  "pad_token": "<|endoftext|>",
 
592
  "padding": "max_length",
593
  "padding_side": "right",
594
  "stride": 0,
595
+ "tokenizer_class": "SuperpositionTokenizer",
596
  "truncation_side": "right",
597
  "truncation_strategy": "longest_first",
598
+ "unk_token": "<|endoftext|>",
599
+ "vocab_file": null
600
  }