Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -60,35 +60,39 @@ def stream_chat(
|
|
60 |
message: str,
|
61 |
history: list,
|
62 |
):
|
63 |
-
system_prompt = "You are a Kamala Harris chatbot. You only answer like Harris in style and tone."
|
64 |
-
temperature = 0
|
65 |
-
max_new_tokens =
|
66 |
-
top_p =
|
67 |
-
top_k =
|
68 |
-
|
69 |
-
|
|
|
|
|
70 |
conversation = [
|
71 |
{"role": "system", "content": system_prompt}
|
72 |
]
|
73 |
for prompt, answer in history:
|
74 |
conversation.extend([
|
75 |
-
{"role": "user", "content": prompt},
|
76 |
{"role": "assistant", "content": answer},
|
77 |
])
|
78 |
-
|
79 |
conversation.append({"role": "user", "content": message})
|
80 |
-
|
81 |
input_ids = tokenizer.apply_chat_template(conversation, add_generation_prompt=True, return_tensors="pt").to(model.device)
|
82 |
|
83 |
streamer = TextIteratorStreamer(tokenizer, timeout=60.0, skip_prompt=True, skip_special_tokens=True)
|
84 |
|
85 |
generate_kwargs = dict(
|
86 |
-
input_ids=input_ids,
|
87 |
max_new_tokens=max_new_tokens,
|
88 |
do_sample=True,
|
89 |
top_p=top_p,
|
90 |
top_k=top_k,
|
91 |
temperature=temperature,
|
|
|
|
|
|
|
92 |
eos_token_id=[128001,128008,128009],
|
93 |
streamer=streamer,
|
94 |
)
|
|
|
60 |
message: str,
|
61 |
history: list,
|
62 |
):
|
63 |
+
system_prompt = "You are a Kamala Harris chatbot. You only answer like Harris in style and tone. Provide concise and key responses."
|
64 |
+
temperature = 1.0
|
65 |
+
max_new_tokens = 150
|
66 |
+
top_p = 0.9
|
67 |
+
top_k = 50
|
68 |
+
repetition_penalty = 1.5
|
69 |
+
frequency_penalty = 0.8
|
70 |
+
presence_penalty = 0.8
|
71 |
+
|
72 |
conversation = [
|
73 |
{"role": "system", "content": system_prompt}
|
74 |
]
|
75 |
for prompt, answer in history:
|
76 |
conversation.extend([
|
77 |
+
{"role": "user", "content": prompt},
|
78 |
{"role": "assistant", "content": answer},
|
79 |
])
|
|
|
80 |
conversation.append({"role": "user", "content": message})
|
81 |
+
|
82 |
input_ids = tokenizer.apply_chat_template(conversation, add_generation_prompt=True, return_tensors="pt").to(model.device)
|
83 |
|
84 |
streamer = TextIteratorStreamer(tokenizer, timeout=60.0, skip_prompt=True, skip_special_tokens=True)
|
85 |
|
86 |
generate_kwargs = dict(
|
87 |
+
input_ids=input_ids,
|
88 |
max_new_tokens=max_new_tokens,
|
89 |
do_sample=True,
|
90 |
top_p=top_p,
|
91 |
top_k=top_k,
|
92 |
temperature=temperature,
|
93 |
+
repetition_penalty=repetition_penalty,
|
94 |
+
frequency_penalty=frequency_penalty,
|
95 |
+
presence_penalty=presence_penalty,
|
96 |
eos_token_id=[128001,128008,128009],
|
97 |
streamer=streamer,
|
98 |
)
|