Spaces:
Running
Running
modify app.py adapt to enter api
Browse files
app.py
CHANGED
@@ -97,18 +97,22 @@ def main():
|
|
97 |
|
98 |
# LLM Settings
|
99 |
st.subheader("LLM Settings")
|
|
|
|
|
|
|
|
|
100 |
opt_model = st.selectbox(
|
101 |
-
"Optimization Model", ["
|
102 |
)
|
103 |
opt_temp = st.slider("Optimization Temperature", 0.0, 1.0, 0.7)
|
104 |
|
105 |
eval_model = st.selectbox(
|
106 |
-
"Evaluation Model", ["gpt-4o-mini", "
|
107 |
)
|
108 |
eval_temp = st.slider("Evaluation Temperature", 0.0, 1.0, 0.3)
|
109 |
|
110 |
exec_model = st.selectbox(
|
111 |
-
"Execution Model", ["gpt-4o-mini", "
|
112 |
)
|
113 |
exec_temp = st.slider("Execution Temperature", 0.0, 1.0, 0.0)
|
114 |
|
@@ -202,9 +206,9 @@ def main():
|
|
202 |
try:
|
203 |
# Initialize LLM
|
204 |
SPO_LLM.initialize(
|
205 |
-
optimize_kwargs={"model": opt_model, "temperature": opt_temp},
|
206 |
-
evaluate_kwargs={"model": eval_model, "temperature": eval_temp},
|
207 |
-
execute_kwargs={"model": exec_model, "temperature": exec_temp},
|
208 |
)
|
209 |
|
210 |
# Create optimizer instance
|
@@ -252,9 +256,12 @@ def main():
|
|
252 |
try:
|
253 |
with st.spinner("Generating response..."):
|
254 |
SPO_LLM.initialize(
|
255 |
-
optimize_kwargs={"model": opt_model, "temperature": opt_temp
|
256 |
-
|
257 |
-
|
|
|
|
|
|
|
258 |
)
|
259 |
|
260 |
llm = SPO_LLM.get_instance()
|
|
|
97 |
|
98 |
# LLM Settings
|
99 |
st.subheader("LLM Settings")
|
100 |
+
|
101 |
+
base_url = st.text_input("Base URL", value="https://api.example.com")
|
102 |
+
api_key = st.text_input("API Key", type="password")
|
103 |
+
|
104 |
opt_model = st.selectbox(
|
105 |
+
"Optimization Model", ["gpt-4o-mini", "gpt-4o", "deepseek-chat"], index=0
|
106 |
)
|
107 |
opt_temp = st.slider("Optimization Temperature", 0.0, 1.0, 0.7)
|
108 |
|
109 |
eval_model = st.selectbox(
|
110 |
+
"Evaluation Model", ["gpt-4o-mini", "gpt-4o", "deepseek-chat"], index=0
|
111 |
)
|
112 |
eval_temp = st.slider("Evaluation Temperature", 0.0, 1.0, 0.3)
|
113 |
|
114 |
exec_model = st.selectbox(
|
115 |
+
"Execution Model", ["gpt-4o-mini", "gpt-4o", "deepseek-chat"], index=0
|
116 |
)
|
117 |
exec_temp = st.slider("Execution Temperature", 0.0, 1.0, 0.0)
|
118 |
|
|
|
206 |
try:
|
207 |
# Initialize LLM
|
208 |
SPO_LLM.initialize(
|
209 |
+
optimize_kwargs={"model": opt_model, "temperature": opt_temp, "base_url": base_url, "api_key": api_key},
|
210 |
+
evaluate_kwargs={"model": eval_model, "temperature": eval_temp, "base_url": base_url, "api_key": api_key},
|
211 |
+
execute_kwargs={"model": exec_model, "temperature": exec_temp, "base_url": base_url, "api_key": api_key},
|
212 |
)
|
213 |
|
214 |
# Create optimizer instance
|
|
|
256 |
try:
|
257 |
with st.spinner("Generating response..."):
|
258 |
SPO_LLM.initialize(
|
259 |
+
optimize_kwargs={"model": opt_model, "temperature": opt_temp, "base_url": base_url,
|
260 |
+
"api_key": api_key},
|
261 |
+
evaluate_kwargs={"model": eval_model, "temperature": eval_temp, "base_url": base_url,
|
262 |
+
"api_key": api_key},
|
263 |
+
execute_kwargs={"model": exec_model, "temperature": exec_temp, "base_url": base_url,
|
264 |
+
"api_key": api_key},
|
265 |
)
|
266 |
|
267 |
llm = SPO_LLM.get_instance()
|