Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
|
@@ -29,7 +29,6 @@ class ChatRequest(BaseModel):
|
|
| 29 |
|
| 30 |
def generate_chat_response(request, llm):
|
| 31 |
try:
|
| 32 |
-
# Normalizaci贸n del mensaje para manejo robusto
|
| 33 |
user_input = normalize_input(request.message)
|
| 34 |
response = llm.create_chat_completion(
|
| 35 |
messages=[{"role": "user", "content": user_input}],
|
|
@@ -43,11 +42,10 @@ def generate_chat_response(request, llm):
|
|
| 43 |
return {"response": f"Error: {str(e)}", "literal": user_input}
|
| 44 |
|
| 45 |
def normalize_input(input_text):
|
| 46 |
-
# Implementar aqu铆 cualquier l贸gica de normalizaci贸n que sea necesaria
|
| 47 |
return input_text.strip()
|
| 48 |
|
| 49 |
def select_best_response(responses, request):
|
| 50 |
-
coherent_responses = filter_by_coherence(
|
| 51 |
best_response = filter_by_similarity(coherent_responses)
|
| 52 |
return best_response
|
| 53 |
|
|
@@ -77,16 +75,22 @@ async def generate_chat(request: ChatRequest):
|
|
| 77 |
response = future.result()
|
| 78 |
responses.append(response)
|
| 79 |
|
|
|
|
| 80 |
if any("Error" in response['response'] for response in responses):
|
| 81 |
error_response = next(response for response in responses if "Error" in response['response'])
|
| 82 |
raise HTTPException(status_code=500, detail=error_response['response'])
|
| 83 |
|
| 84 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 85 |
|
| 86 |
return {
|
| 87 |
"best_response": best_response,
|
| 88 |
-
"all_responses":
|
| 89 |
-
"literal_inputs":
|
| 90 |
}
|
| 91 |
|
| 92 |
if __name__ == "__main__":
|
|
|
|
| 29 |
|
| 30 |
def generate_chat_response(request, llm):
|
| 31 |
try:
|
|
|
|
| 32 |
user_input = normalize_input(request.message)
|
| 33 |
response = llm.create_chat_completion(
|
| 34 |
messages=[{"role": "user", "content": user_input}],
|
|
|
|
| 42 |
return {"response": f"Error: {str(e)}", "literal": user_input}
|
| 43 |
|
| 44 |
def normalize_input(input_text):
|
|
|
|
| 45 |
return input_text.strip()
|
| 46 |
|
| 47 |
def select_best_response(responses, request):
|
| 48 |
+
coherent_responses = filter_by_coherence(responses, request)
|
| 49 |
best_response = filter_by_similarity(coherent_responses)
|
| 50 |
return best_response
|
| 51 |
|
|
|
|
| 75 |
response = future.result()
|
| 76 |
responses.append(response)
|
| 77 |
|
| 78 |
+
# Verifica si alguna respuesta contiene un error y maneja el error si es necesario
|
| 79 |
if any("Error" in response['response'] for response in responses):
|
| 80 |
error_response = next(response for response in responses if "Error" in response['response'])
|
| 81 |
raise HTTPException(status_code=500, detail=error_response['response'])
|
| 82 |
|
| 83 |
+
# Extrae las respuestas y las entradas literales
|
| 84 |
+
response_texts = [resp['response'] for resp in responses]
|
| 85 |
+
literal_inputs = [resp['literal'] for resp in responses]
|
| 86 |
+
|
| 87 |
+
# Selecciona la mejor respuesta
|
| 88 |
+
best_response = select_best_response(response_texts, request)
|
| 89 |
|
| 90 |
return {
|
| 91 |
"best_response": best_response,
|
| 92 |
+
"all_responses": response_texts,
|
| 93 |
+
"literal_inputs": literal_inputs
|
| 94 |
}
|
| 95 |
|
| 96 |
if __name__ == "__main__":
|