Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -68,7 +68,7 @@ ds = load_dataset("lamm-mit/protein_secondary_structure_from_PDB",
|
|
68 |
|
69 |
def respond(
|
70 |
message,
|
71 |
-
history
|
72 |
system_message,
|
73 |
max_tokens,
|
74 |
temperature,
|
@@ -76,10 +76,10 @@ def respond(
|
|
76 |
):
|
77 |
messages = [{"role": "system", "content": system_message}]
|
78 |
|
79 |
-
for
|
80 |
-
messages.append({"role": "user", "content":
|
81 |
-
if
|
82 |
-
messages.append({"role": "assistant", "content":
|
83 |
|
84 |
messages.append({"role": "user", "content": message})
|
85 |
|
@@ -97,12 +97,15 @@ def respond(
|
|
97 |
token = chunk.choices[0].delta.content
|
98 |
if token is not None:
|
99 |
response += token
|
100 |
-
yield [
|
|
|
101 |
|
102 |
-
return [
|
|
|
103 |
except Exception as e:
|
104 |
print(f"Error in respond: {str(e)}")
|
105 |
-
return [
|
|
|
106 |
|
107 |
def process_chat(message, history):
|
108 |
try:
|
@@ -111,11 +114,20 @@ def process_chat(message, history):
|
|
111 |
params = extract_parameters(message, relevant_data)
|
112 |
protein_result = generate_protein(params)
|
113 |
explanation = generate_explanation(protein_result, params)
|
114 |
-
return history + [
|
|
|
|
|
|
|
115 |
else:
|
116 |
-
return history + [
|
|
|
|
|
|
|
117 |
except Exception as e:
|
118 |
-
return history + [
|
|
|
|
|
|
|
119 |
|
120 |
def search_protein_data(query):
|
121 |
relevant_entries = []
|
@@ -624,6 +636,7 @@ with gr.Blocks(theme='ParityError/Interstellar') as demo:
|
|
624 |
chatbot = gr.Chatbot(
|
625 |
height=600,
|
626 |
type='messages' # ๋ฉ์์ง ํ์ ์ง์
|
|
|
627 |
)
|
628 |
with gr.Row():
|
629 |
msg = gr.Textbox(
|
|
|
68 |
|
69 |
def respond(
|
70 |
message,
|
71 |
+
history,
|
72 |
system_message,
|
73 |
max_tokens,
|
74 |
temperature,
|
|
|
76 |
):
|
77 |
messages = [{"role": "system", "content": system_message}]
|
78 |
|
79 |
+
for msg in history:
|
80 |
+
messages.append({"role": "user", "content": msg[0]})
|
81 |
+
if msg[1]:
|
82 |
+
messages.append({"role": "assistant", "content": msg[1]})
|
83 |
|
84 |
messages.append({"role": "user", "content": message})
|
85 |
|
|
|
97 |
token = chunk.choices[0].delta.content
|
98 |
if token is not None:
|
99 |
response += token
|
100 |
+
yield [{"role": "user", "content": message},
|
101 |
+
{"role": "assistant", "content": response}]
|
102 |
|
103 |
+
return [{"role": "user", "content": message},
|
104 |
+
{"role": "assistant", "content": response}]
|
105 |
except Exception as e:
|
106 |
print(f"Error in respond: {str(e)}")
|
107 |
+
return [{"role": "user", "content": message},
|
108 |
+
{"role": "assistant", "content": f"์ค๋ฅ๊ฐ ๋ฐ์ํ์ต๋๋ค: {str(e)}"}]
|
109 |
|
110 |
def process_chat(message, history):
|
111 |
try:
|
|
|
114 |
params = extract_parameters(message, relevant_data)
|
115 |
protein_result = generate_protein(params)
|
116 |
explanation = generate_explanation(protein_result, params)
|
117 |
+
return history + [
|
118 |
+
{"role": "user", "content": message},
|
119 |
+
{"role": "assistant", "content": explanation}
|
120 |
+
]
|
121 |
else:
|
122 |
+
return history + [
|
123 |
+
{"role": "user", "content": message},
|
124 |
+
{"role": "assistant", "content": "๋จ๋ฐฑ์ง ์์ฑ ๊ด๋ จ ํค์๋๋ฅผ ํฌํจํด์ฃผ์ธ์."}
|
125 |
+
]
|
126 |
except Exception as e:
|
127 |
+
return history + [
|
128 |
+
{"role": "user", "content": message},
|
129 |
+
{"role": "assistant", "content": f"์ค๋ฅ๊ฐ ๋ฐ์ํ์ต๋๋ค: {str(e)}"}
|
130 |
+
]
|
131 |
|
132 |
def search_protein_data(query):
|
133 |
relevant_entries = []
|
|
|
636 |
chatbot = gr.Chatbot(
|
637 |
height=600,
|
638 |
type='messages' # ๋ฉ์์ง ํ์ ์ง์
|
639 |
+
|
640 |
)
|
641 |
with gr.Row():
|
642 |
msg = gr.Textbox(
|