# Load your data from a text file | |
with open('C:\\Users\\money\\OneDrive\\Pictures\\Blank Model\\untrained\\New folder (3)\\questions.txt', 'r') as f: | |
data = f.read() | |
# Split your data into individual questions | |
questions = data.split('\n') | |
# Tokenize each question using your custom tokenizer | |
tokenized_questions = [] | |
for question in questions: | |
tokens = custom_tokenizer(question) | |
tokenized_questions.append(tokens) | |
# Now you can use tokenized_questions to train your model | |