Emmanuel Frimpong Asante
commited on
Commit
·
0c971f5
1
Parent(s):
e7420c7
"Update space"
Browse filesSigned-off-by: Emmanuel Frimpong Asante <[email protected]>
app.py
CHANGED
@@ -98,6 +98,10 @@ model_name = "meta-llama/Llama-2-7b-chat-hf" # Example model name; replace with
|
|
98 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
99 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
100 |
|
|
|
|
|
|
|
|
|
101 |
|
102 |
# Define Llama 2 response generation
|
103 |
def llama2_response(user_input):
|
@@ -108,7 +112,7 @@ def llama2_response(user_input):
|
|
108 |
max_length=150,
|
109 |
do_sample=True,
|
110 |
temperature=0.7,
|
111 |
-
pad_token_id=tokenizer.
|
112 |
attention_mask=inputs["attention_mask"]
|
113 |
)
|
114 |
|
|
|
98 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
99 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
100 |
|
101 |
+
# Set the padding token to EOS token or add a new padding token
|
102 |
+
if tokenizer.pad_token is None:
|
103 |
+
tokenizer.add_special_tokens({'pad_token': '[PAD]'})
|
104 |
+
model.resize_token_embeddings(len(tokenizer))
|
105 |
|
106 |
# Define Llama 2 response generation
|
107 |
def llama2_response(user_input):
|
|
|
112 |
max_length=150,
|
113 |
do_sample=True,
|
114 |
temperature=0.7,
|
115 |
+
pad_token_id=tokenizer.pad_token_id, # Use the newly set padding token
|
116 |
attention_mask=inputs["attention_mask"]
|
117 |
)
|
118 |
|