Spaces:
Running
Running
refactor
Browse files- app.py +1 -1
- logs/app.log +32 -0
- model_params.cfg +1 -6
app.py
CHANGED
@@ -45,7 +45,7 @@ scheduler = CommitScheduler(
|
|
45 |
repo_id=model_config.get('app','repo_id'),
|
46 |
repo_type="dataset",
|
47 |
folder_path=JSON_DATASET_DIR,
|
48 |
-
path_in_repo=
|
49 |
token=SPACES_LOG,
|
50 |
every=2) # TESTING: every 2 seconds
|
51 |
|
|
|
45 |
repo_id=model_config.get('app','repo_id'),
|
46 |
repo_type="dataset",
|
47 |
folder_path=JSON_DATASET_DIR,
|
48 |
+
path_in_repo="audit_chatbot",
|
49 |
token=SPACES_LOG,
|
50 |
every=2) # TESTING: every 2 seconds
|
51 |
|
logs/app.log
CHANGED
@@ -1374,3 +1374,35 @@ Make sure your token has the correct permissions.
|
|
1374 |
2025-02-13 15:05:59,504 - httpx - INFO - HTTP Request: GET http://127.0.0.1:7860/startup-events "HTTP/1.1 200 OK"
|
1375 |
2025-02-13 15:05:59,515 - httpx - INFO - HTTP Request: HEAD http://127.0.0.1:7860/ "HTTP/1.1 200 OK"
|
1376 |
2025-02-13 15:06:00,079 - httpx - INFO - HTTP Request: GET https://api.gradio.app/pkg-version "HTTP/1.1 200 OK"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1374 |
2025-02-13 15:05:59,504 - httpx - INFO - HTTP Request: GET http://127.0.0.1:7860/startup-events "HTTP/1.1 200 OK"
|
1375 |
2025-02-13 15:05:59,515 - httpx - INFO - HTTP Request: HEAD http://127.0.0.1:7860/ "HTTP/1.1 200 OK"
|
1376 |
2025-02-13 15:06:00,079 - httpx - INFO - HTTP Request: GET https://api.gradio.app/pkg-version "HTTP/1.1 200 OK"
|
1377 |
+
2025-02-13 15:06:18,759 - __main__ - DEBUG - Created new session: d1422289-8e00-4c3c-b29c-24c0b87e0375
|
1378 |
+
2025-02-13 15:06:18,759 - __main__ - DEBUG - Session duration calculated: 1.1e-05
|
1379 |
+
2025-02-13 15:06:18,762 - auditqa.retriever - INFO - Retriever activated
|
1380 |
+
2025-02-13 15:06:19,969 - sentence_transformers.cross_encoder.CrossEncoder - INFO - Use pytorch device: mps
|
1381 |
+
2025-02-13 15:06:22,707 - httpx - INFO - HTTP Request: POST https://ff3f0448-0a00-470e-9956-49efa3071db3.europe-west3-0.gcp.cloud.qdrant.io:6333/collections/allreports/points/search "HTTP/1.1 200 OK"
|
1382 |
+
2025-02-13 15:06:25,002 - auditqa.retriever - INFO - retrieved paragraphs:3
|
1383 |
+
2025-02-13 15:06:25,220 - __main__ - DEBUG - Logs data before save: {
|
1384 |
+
"session_id": "d1422289-8e00-4c3c-b29c-24c0b87e0375",
|
1385 |
+
"client_location": {
|
1386 |
+
"city": null,
|
1387 |
+
"region": null,
|
1388 |
+
"country": null,
|
1389 |
+
"latitude": null,
|
1390 |
+
"longitude": null
|
1391 |
+
},
|
1392 |
+
"session_duration_seconds": 1.1e-05,
|
1393 |
+
"year": [
|
1394 |
+
"2023"
|
1395 |
+
],
|
1396 |
+
"question": "What are the key findings of this audit report?",
|
1397 |
+
"retriever": "BAAI/bge-large-en-v1.5",
|
1398 |
+
"endpoint_type": "DEDICATED",
|
1399 |
+
"reader": "meta-llama/Llama-3.1-8B-Instruct",
|
1400 |
+
"answer": "",
|
1401 |
+
"time": "1739459185.220484"
|
1402 |
+
}
|
1403 |
+
2025-02-13 15:06:25,220 - auditqa.reader - INFO - Serverless endpoint activated
|
1404 |
+
2025-02-13 15:06:25,220 - auditqa.reader - INFO - Initializing InferenceClient with model: meta-llama/Meta-Llama-3-8B-Instruct
|
1405 |
+
2025-02-13 15:06:25,220 - auditqa.reader - INFO - Serverless InferenceClient initialization successful
|
1406 |
+
2025-02-13 15:06:39,048 - auditqa.utils - INFO - logging done
|
1407 |
+
2025-02-13 15:06:39,049 - __main__ - DEBUG - Logs saved successfully
|
1408 |
+
2025-02-13 15:06:41,719 - auditqa.utils - INFO - logging done
|
model_params.cfg
CHANGED
@@ -13,9 +13,4 @@ NVIDIA_MODEL = meta-llama/Llama-3.1-8B-Instruct
|
|
13 |
NVIDIA_ENDPOINT = https://huggingface.co/api/integrations/dgx/v1
|
14 |
MAX_TOKENS = 512
|
15 |
[app]
|
16 |
-
repo_id = mtyrrell/
|
17 |
-
path_in_repo = .
|
18 |
-
|
19 |
-
|
20 |
-
# GIZ/spaces_log
|
21 |
-
# audit_chatbot
|
|
|
13 |
NVIDIA_ENDPOINT = https://huggingface.co/api/integrations/dgx/v1
|
14 |
MAX_TOKENS = 512
|
15 |
[app]
|
16 |
+
repo_id = mtyrrell/spaces_log
|
|
|
|
|
|
|
|
|
|