AurelioAguirre commited on
Commit
b4e70da
·
1 Parent(s): 15c704c

Fixed smaller issues

Browse files
Files changed (6) hide show
  1. =0.45.0 +30 -0
  2. main/api.py +9 -3
  3. main/app.py +19 -3
  4. main/logs/llm_api.log +629 -0
  5. main/utils/logging.py +3 -0
  6. requirements.txt +2 -1
=0.45.0 ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Collecting bitsandbytes
2
+ Downloading bitsandbytes-0.45.0-py3-none-manylinux_2_24_x86_64.whl.metadata (2.9 kB)
3
+ Requirement already satisfied: torch in ./myenv/lib/python3.12/site-packages (from bitsandbytes) (2.5.1)
4
+ Requirement already satisfied: numpy in ./myenv/lib/python3.12/site-packages (from bitsandbytes) (2.2.1)
5
+ Requirement already satisfied: typing_extensions>=4.8.0 in ./myenv/lib/python3.12/site-packages (from bitsandbytes) (4.12.2)
6
+ Requirement already satisfied: filelock in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (3.16.1)
7
+ Requirement already satisfied: networkx in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (3.4.2)
8
+ Requirement already satisfied: jinja2 in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (3.1.5)
9
+ Requirement already satisfied: fsspec in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (2024.12.0)
10
+ Requirement already satisfied: nvidia-cuda-nvrtc-cu12==12.4.127 in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (12.4.127)
11
+ Requirement already satisfied: nvidia-cuda-runtime-cu12==12.4.127 in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (12.4.127)
12
+ Requirement already satisfied: nvidia-cuda-cupti-cu12==12.4.127 in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (12.4.127)
13
+ Requirement already satisfied: nvidia-cudnn-cu12==9.1.0.70 in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (9.1.0.70)
14
+ Requirement already satisfied: nvidia-cublas-cu12==12.4.5.8 in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (12.4.5.8)
15
+ Requirement already satisfied: nvidia-cufft-cu12==11.2.1.3 in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (11.2.1.3)
16
+ Requirement already satisfied: nvidia-curand-cu12==10.3.5.147 in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (10.3.5.147)
17
+ Requirement already satisfied: nvidia-cusolver-cu12==11.6.1.9 in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (11.6.1.9)
18
+ Requirement already satisfied: nvidia-cusparse-cu12==12.3.1.170 in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (12.3.1.170)
19
+ Requirement already satisfied: nvidia-nccl-cu12==2.21.5 in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (2.21.5)
20
+ Requirement already satisfied: nvidia-nvtx-cu12==12.4.127 in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (12.4.127)
21
+ Requirement already satisfied: nvidia-nvjitlink-cu12==12.4.127 in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (12.4.127)
22
+ Requirement already satisfied: triton==3.1.0 in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (3.1.0)
23
+ Requirement already satisfied: setuptools in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (75.8.0)
24
+ Requirement already satisfied: sympy==1.13.1 in ./myenv/lib/python3.12/site-packages (from torch->bitsandbytes) (1.13.1)
25
+ Requirement already satisfied: mpmath<1.4,>=1.1.0 in ./myenv/lib/python3.12/site-packages (from sympy==1.13.1->torch->bitsandbytes) (1.3.0)
26
+ Requirement already satisfied: MarkupSafe>=2.0 in ./myenv/lib/python3.12/site-packages (from jinja2->torch->bitsandbytes) (3.0.2)
27
+ Downloading bitsandbytes-0.45.0-py3-none-manylinux_2_24_x86_64.whl (69.1 MB)
28
+ ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 69.1/69.1 MB 64.0 MB/s eta 0:00:00
29
+ Installing collected packages: bitsandbytes
30
+ Successfully installed bitsandbytes-0.45.0
main/api.py CHANGED
@@ -3,7 +3,7 @@ from pathlib import Path
3
  from threading import Thread
4
  import torch
5
  from typing import Optional, Iterator, List
6
- from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
7
  from .utils.logging import setup_logger
8
 
9
  class LLMApi:
@@ -100,12 +100,18 @@ class LLMApi:
100
  self.logger.info(f"Loading model from source: {model_name}")
101
  model_path = model_name
102
 
 
 
 
 
 
103
  self.generation_model = AutoModelForCausalLM.from_pretrained(
104
  model_path,
105
  device_map="auto",
106
- load_in_8bit=True,
107
  torch_dtype=torch.float16
108
- )
 
109
  self.tokenizer = AutoTokenizer.from_pretrained(model_path)
110
 
111
  # Update generation config with tokenizer-specific values
 
3
  from threading import Thread
4
  import torch
5
  from typing import Optional, Iterator, List
6
+ from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer, BitsAndBytesConfig
7
  from .utils.logging import setup_logger
8
 
9
  class LLMApi:
 
100
  self.logger.info(f"Loading model from source: {model_name}")
101
  model_path = model_name
102
 
103
+ quantization_config = BitsAndBytesConfig(
104
+ load_in_8bit=True,
105
+ llm_int8_threshold=3.0
106
+ )
107
+
108
  self.generation_model = AutoModelForCausalLM.from_pretrained(
109
  model_path,
110
  device_map="auto",
111
+ quantization_config=quantization_config,
112
  torch_dtype=torch.float16
113
+ )
114
+
115
  self.tokenizer = AutoTokenizer.from_pretrained(model_path)
116
 
117
  # Update generation config with tokenizer-specific values
main/app.py CHANGED
@@ -12,9 +12,13 @@ def load_config():
12
  with open("main/config.yaml", "r") as f:
13
  return yaml.safe_load(f)
14
 
 
 
 
15
  def create_app():
16
- config = load_config()
17
- logger = setup_logger(config, "main")
 
18
  validate_hf(setup_logger, config)
19
  logger.info("Starting LLM API server")
20
 
@@ -41,4 +45,16 @@ def create_app():
41
  logger.info("FastAPI application created successfully")
42
  return app
43
 
44
- app = create_app()
 
 
 
 
 
 
 
 
 
 
 
 
 
12
  with open("main/config.yaml", "r") as f:
13
  return yaml.safe_load(f)
14
 
15
+ config = load_config()
16
+ logger = setup_logger(config, "main")
17
+
18
  def create_app():
19
+ global config
20
+ global logger
21
+
22
  validate_hf(setup_logger, config)
23
  logger.info("Starting LLM API server")
24
 
 
45
  logger.info("FastAPI application created successfully")
46
  return app
47
 
48
+ app = create_app()
49
+
50
+ if __name__ == "__main__":
51
+ host = config["server"]["host"]
52
+ port = config["server"]["port"]
53
+ import uvicorn
54
+ uvicorn.run(
55
+ app,
56
+ host=host,
57
+ port=port,
58
+ log_level=config["logging"]["level"].lower()
59
+ )
60
+ logger.info(f"LLM API server started on {host}:{port}")
main/logs/llm_api.log CHANGED
@@ -72,3 +72,632 @@
72
  Dynamo is not supported on Python 3.13+
73
  2025-01-13 16:18:45,982 - api_routes - ERROR - Error initializing model: Failed to import transformers.integrations.bitsandbytes because of the following error (look up to see its traceback):
74
  Dynamo is not supported on Python 3.13+
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
72
  Dynamo is not supported on Python 3.13+
73
  2025-01-13 16:18:45,982 - api_routes - ERROR - Error initializing model: Failed to import transformers.integrations.bitsandbytes because of the following error (look up to see its traceback):
74
  Dynamo is not supported on Python 3.13+
75
+ 2025-01-14 11:41:25,502 - hf_validation - WARNING - No .env file found. Fine if you're on Huggingface, but you need one to run locally on your PC.
76
+ 2025-01-14 11:41:25,502 - hf_validation - ERROR - No HF_TOKEN found in environment variables
77
+ 2025-01-14 11:41:25,502 - main - INFO - Starting LLM API server
78
+ 2025-01-14 11:41:25,503 - llm_api - INFO - Initializing LLM API
79
+ 2025-01-14 11:41:25,503 - llm_api - INFO - LLM API initialized successfully
80
+ 2025-01-14 11:41:25,503 - api_routes - INFO - Router initialized with LLM API instance
81
+ 2025-01-14 11:41:25,509 - main - INFO - FastAPI application created successfully
82
+ 2025-01-14 11:48:33,807 - hf_validation - WARNING - No .env file found. Fine if you're on Huggingface, but you need one to run locally on your PC.
83
+ 2025-01-14 11:48:33,807 - hf_validation - ERROR - No HF_TOKEN found in environment variables
84
+ 2025-01-14 11:48:33,807 - main - INFO - Starting LLM API server
85
+ 2025-01-14 11:48:33,807 - llm_api - INFO - Initializing LLM API
86
+ 2025-01-14 11:48:33,807 - llm_api - INFO - LLM API initialized successfully
87
+ 2025-01-14 11:48:33,807 - api_routes - INFO - Router initialized with LLM API instance
88
+ 2025-01-14 11:48:33,812 - main - INFO - FastAPI application created successfully
89
+ 2025-01-14 11:53:20,777 - hf_validation - WARNING - No .env file found. Fine if you're on Huggingface, but you need one to run locally on your PC.
90
+ 2025-01-14 11:53:20,777 - hf_validation - ERROR - No HF_TOKEN found in environment variables
91
+ 2025-01-14 11:53:20,777 - main - INFO - Starting LLM API server
92
+ 2025-01-14 11:53:20,777 - llm_api - INFO - Initializing LLM API
93
+ 2025-01-14 11:53:20,778 - llm_api - INFO - LLM API initialized successfully
94
+ 2025-01-14 11:53:20,778 - api_routes - INFO - Router initialized with LLM API instance
95
+ 2025-01-14 11:53:20,783 - main - INFO - FastAPI application created successfully
96
+ 2025-01-14 11:54:28,143 - api_routes - INFO - Received request to download model: microsoft/Phi-3.5-mini-instruct
97
+ 2025-01-14 11:54:28,143 - llm_api - INFO - Starting download of model: microsoft/Phi-3.5-mini-instruct
98
+ 2025-01-14 11:54:28,143 - llm_api - INFO - Enabling stdout logging for download
99
+ 2025-01-14 11:54:47,061 - llm_api - INFO - Disabling stdout logging
100
+ 2025-01-14 11:54:47,061 - llm_api - INFO - Saving model to main/models/Phi-3.5-mini-instruct
101
+ 2025-01-14 11:56:40,600 - llm_api - INFO - Successfully downloaded model: microsoft/Phi-3.5-mini-instruct
102
+ 2025-01-14 11:56:41,266 - api_routes - INFO - Successfully downloaded model: microsoft/Phi-3.5-mini-instruct
103
+ 2025-01-14 11:56:41,364 - api_routes - INFO - Received request to initialize model: microsoft/Phi-3.5-mini-instruct
104
+ 2025-01-14 11:56:41,365 - llm_api - INFO - Initializing generation model: microsoft/Phi-3.5-mini-instruct
105
+ 2025-01-14 11:56:41,367 - llm_api - INFO - Loading model from local path: main/models/Phi-3.5-mini-instruct
106
+ 2025-01-14 11:56:45,322 - llm_api - ERROR - Failed to initialize generation model microsoft/Phi-3.5-mini-instruct: /home/aurelio/Desktop/Projects/LLMServer/myenv/lib/python3.12/site-packages/bitsandbytes/libbitsandbytes_cpu.so: undefined symbol: cget_col_row_stats
107
+ 2025-01-14 11:56:45,322 - api_routes - ERROR - Error initializing model: /home/aurelio/Desktop/Projects/LLMServer/myenv/lib/python3.12/site-packages/bitsandbytes/libbitsandbytes_cpu.so: undefined symbol: cget_col_row_stats
108
+ 2025-01-14 12:29:54,971 - main - INFO - LLM API server started on 0.0.0.0:8001
109
+ 2025-01-14 12:30:01,275 - hf_validation - WARNING - No .env file found. Fine if you're on Huggingface, but you need one to run locally on your PC.
110
+ 2025-01-14 12:30:01,275 - hf_validation - ERROR - No HF_TOKEN found in environment variables
111
+ 2025-01-14 12:30:01,275 - main - INFO - Starting LLM API server
112
+ 2025-01-14 12:30:01,275 - llm_api - INFO - Initializing LLM API
113
+ 2025-01-14 12:30:01,275 - llm_api - INFO - LLM API initialized successfully
114
+ 2025-01-14 12:30:01,276 - api_routes - INFO - Router initialized with LLM API instance
115
+ 2025-01-14 12:30:01,280 - main - INFO - FastAPI application created successfully
116
+ 2025-01-14 12:31:15,345 - hf_validation - WARNING - No .env file found. Fine if you're on Huggingface, but you need one to run locally on your PC.
117
+ 2025-01-14 12:31:15,345 - hf_validation - ERROR - No HF_TOKEN found in environment variables
118
+ 2025-01-14 12:31:15,345 - main - INFO - Starting LLM API server
119
+ 2025-01-14 12:31:15,345 - llm_api - INFO - Initializing LLM API
120
+ 2025-01-14 12:31:15,346 - llm_api - INFO - LLM API initialized successfully
121
+ 2025-01-14 12:31:15,346 - api_routes - INFO - Router initialized with LLM API instance
122
+ 2025-01-14 12:31:15,350 - main - INFO - FastAPI application created successfully
123
+ 2025-01-14 12:31:43,376 - hf_validation - WARNING - No .env file found. Fine if you're on Huggingface, but you need one to run locally on your PC.
124
+ 2025-01-14 12:31:43,376 - hf_validation - ERROR - No HF_TOKEN found in environment variables
125
+ 2025-01-14 12:31:43,376 - main - INFO - Starting LLM API server
126
+ 2025-01-14 12:31:43,377 - llm_api - INFO - Initializing LLM API
127
+ 2025-01-14 12:31:43,377 - llm_api - INFO - LLM API initialized successfully
128
+ 2025-01-14 12:31:43,377 - api_routes - INFO - Router initialized with LLM API instance
129
+ 2025-01-14 12:31:43,381 - main - INFO - FastAPI application created successfully
130
+ 2025-01-14 12:31:51,142 - llm_api - INFO - INFO: 127.0.0.1:52554 - "GET /docs HTTP/1.1" 200 OK
131
+ 2025-01-14 12:31:51,311 - llm_api - INFO - INFO: 127.0.0.1:52554 - "GET /openapi.json HTTP/1.1" 200 OK
132
+ 2025-01-14 12:32:10,756 - api_routes - INFO - Received request to initialize model: microsoft/Phi-3.5-mini-instruct
133
+ 2025-01-14 12:32:10,757 - llm_api - INFO - Initializing generation model: microsoft/Phi-3.5-mini-instruct
134
+ 2025-01-14 12:32:10,757 - llm_api - INFO - Loading model from local path: main/models/Phi-3.5-mini-instruct
135
+ 2025-01-14 12:32:26,447 - llm_api - INFO - Successfully initialized generation model: microsoft/Phi-3.5-mini-instruct
136
+ 2025-01-14 12:32:26,448 - api_routes - INFO - Successfully initialized model: microsoft/Phi-3.5-mini-instruct
137
+ 2025-01-14 12:32:26,448 - llm_api - INFO - INFO: 127.0.0.1:34282 - "POST /api/v1/model/initialize?model_name=microsoft%2FPhi-3.5-mini-instruct HTTP/1.1" 200 OK
138
+ 2025-01-14 12:33:13,272 - api_routes - INFO - Received generation request for prompt: Tell me about yourself, and your capabilities...
139
+ 2025-01-14 12:33:13,272 - llm_api - DEBUG - Generating response for prompt: Tell me about yourself, and your capabilities...
140
+ 2025-01-14 12:33:46,448 - llm_api - DEBUG - Generated response: I am Phi, an AI language model developed by Micros...
141
+ 2025-01-14 12:33:46,448 - api_routes - INFO - Successfully generated response
142
+ 2025-01-14 12:33:46,448 - llm_api - INFO - INFO: 127.0.0.1:57442 - "POST /api/v1/generate HTTP/1.1" 200 OK
143
+ 2025-01-14 12:34:41,588 - api_routes - INFO - Received streaming generation request for prompt: Tell me about yourself, and your capabilities...
144
+ 2025-01-14 12:34:41,588 - llm_api - DEBUG - Starting streaming generation for prompt: Tell me about yourself, and your capabilities...
145
+ 2025-01-14 12:34:41,608 - llm_api - DEBUG - Generated chunk: <|system|> You are a helpful assistant<|end|><|use...
146
+ 2025-01-14 12:34:41,689 - llm_api - DEBUG - Generated chunk: capabilities<|end|><|assistant|> ...
147
+ 2025-01-14 12:34:41,757 - llm_api - DEBUG - Generated chunk: I ...
148
+ 2025-01-14 12:34:41,827 - llm_api - DEBUG - Generated chunk: am ...
149
+ 2025-01-14 12:34:41,895 - llm_api - DEBUG - Generated chunk: ...
150
+ 2025-01-14 12:34:41,965 - llm_api - DEBUG - Generated chunk: ...
151
+ 2025-01-14 12:34:42,033 - llm_api - DEBUG - Generated chunk: Phi, ...
152
+ 2025-01-14 12:34:42,102 - llm_api - DEBUG - Generated chunk: an ...
153
+ 2025-01-14 12:34:42,172 - llm_api - DEBUG - Generated chunk: ...
154
+ 2025-01-14 12:34:42,241 - llm_api - DEBUG - Generated chunk: AI ...
155
+ 2025-01-14 12:34:42,309 - llm_api - DEBUG - Generated chunk: language ...
156
+ 2025-01-14 12:34:42,377 - llm_api - DEBUG - Generated chunk: model ...
157
+ 2025-01-14 12:34:42,448 - llm_api - DEBUG - Generated chunk: created ...
158
+ 2025-01-14 12:34:42,521 - llm_api - DEBUG - Generated chunk: by ...
159
+ 2025-01-14 12:34:42,590 - llm_api - DEBUG - Generated chunk: ...
160
+ 2025-01-14 12:34:42,656 - llm_api - DEBUG - Generated chunk: Microsoft. ...
161
+ 2025-01-14 12:34:42,721 - llm_api - DEBUG - Generated chunk: While ...
162
+ 2025-01-14 12:34:42,788 - llm_api - DEBUG - Generated chunk: I ...
163
+ 2025-01-14 12:34:42,854 - llm_api - DEBUG - Generated chunk: ...
164
+ 2025-01-14 12:34:42,925 - llm_api - DEBUG - Generated chunk: ...
165
+ 2025-01-14 12:34:42,991 - llm_api - DEBUG - Generated chunk: don't ...
166
+ 2025-01-14 12:34:43,063 - llm_api - DEBUG - Generated chunk: have ...
167
+ 2025-01-14 12:34:43,131 - llm_api - DEBUG - Generated chunk: personal ...
168
+ 2025-01-14 12:34:43,201 - llm_api - DEBUG - Generated chunk: experiences ...
169
+ 2025-01-14 12:34:43,267 - llm_api - DEBUG - Generated chunk: or ...
170
+ 2025-01-14 12:34:43,334 - llm_api - DEBUG - Generated chunk: feelings ...
171
+ 2025-01-14 12:34:43,402 - llm_api - DEBUG - Generated chunk: like ...
172
+ 2025-01-14 12:34:43,472 - llm_api - DEBUG - Generated chunk: humans ...
173
+ 2025-01-14 12:34:43,537 - llm_api - DEBUG - Generated chunk: ...
174
+ 2025-01-14 12:34:43,600 - llm_api - DEBUG - Generated chunk: do, ...
175
+ 2025-01-14 12:34:43,663 - llm_api - DEBUG - Generated chunk: let ...
176
+ 2025-01-14 12:34:43,729 - llm_api - DEBUG - Generated chunk: me ...
177
+ 2025-01-14 12:34:43,793 - llm_api - DEBUG - Generated chunk: tell ...
178
+ 2025-01-14 12:34:43,859 - llm_api - DEBUG - Generated chunk: you ...
179
+ 2025-01-14 12:34:43,924 - llm_api - DEBUG - Generated chunk: more ...
180
+ 2025-01-14 12:34:43,989 - llm_api - DEBUG - Generated chunk: about ...
181
+ 2025-01-14 12:34:44,053 - llm_api - DEBUG - Generated chunk: my ...
182
+ 2025-01-14 12:34:44,119 - llm_api - DEBUG - Generated chunk: ...
183
+ 2025-01-14 12:34:44,183 - llm_api - DEBUG - Generated chunk: ...
184
+ 2025-01-14 12:34:44,247 - llm_api - DEBUG - Generated chunk: functionalities:
185
+ ...
186
+ 2025-01-14 12:34:44,312 - llm_api - DEBUG - Generated chunk:
187
+ ...
188
+ 2025-01-14 12:34:44,375 - llm_api - DEBUG - Generated chunk: ...
189
+ 2025-01-14 12:34:44,440 - llm_api - DEBUG - Generated chunk: ...
190
+ 2025-01-14 12:34:44,505 - llm_api - DEBUG - Generated chunk: 1. ...
191
+ 2025-01-14 12:34:44,567 - llm_api - DEBUG - Generated chunk: ...
192
+ 2025-01-14 12:34:44,632 - llm_api - DEBUG - Generated chunk: **Language ...
193
+ 2025-01-14 12:34:44,698 - llm_api - DEBUG - Generated chunk: ...
194
+ 2025-01-14 12:34:44,763 - llm_api - DEBUG - Generated chunk: ...
195
+ 2025-01-14 12:34:44,828 - llm_api - DEBUG - Generated chunk: ...
196
+ 2025-01-14 12:34:44,892 - llm_api - DEBUG - Generated chunk: Understanding**: ...
197
+ 2025-01-14 12:34:44,956 - llm_api - DEBUG - Generated chunk: My ...
198
+ 2025-01-14 12:34:45,020 - llm_api - DEBUG - Generated chunk: primary ...
199
+ 2025-01-14 12:34:45,085 - llm_api - DEBUG - Generated chunk: ...
200
+ 2025-01-14 12:34:45,148 - llm_api - DEBUG - Generated chunk: capability ...
201
+ 2025-01-14 12:34:45,211 - llm_api - DEBUG - Generated chunk: is ...
202
+ 2025-01-14 12:34:45,275 - llm_api - DEBUG - Generated chunk: understanding ...
203
+ 2025-01-14 12:34:45,338 - llm_api - DEBUG - Generated chunk: natural ...
204
+ 2025-01-14 12:34:45,401 - llm_api - DEBUG - Generated chunk: human ...
205
+ 2025-01-14 12:34:45,465 - llm_api - DEBUG - Generated chunk: languages ...
206
+ 2025-01-14 12:34:45,529 - llm_api - DEBUG - Generated chunk: to ...
207
+ 2025-01-14 12:34:45,592 - llm_api - DEBUG - Generated chunk: the ...
208
+ 2025-01-14 12:34:45,658 - llm_api - DEBUG - Generated chunk: best ...
209
+ 2025-01-14 12:34:45,728 - llm_api - DEBUG - Generated chunk: of ...
210
+ 2025-01-14 12:34:45,804 - llm_api - DEBUG - Generated chunk: our ...
211
+ 2025-01-14 12:34:45,871 - llm_api - DEBUG - Generated chunk: current ...
212
+ 2025-01-14 12:34:45,938 - llm_api - DEBUG - Generated chunk: technology ...
213
+ 2025-01-14 12:34:46,004 - llm_api - DEBUG - Generated chunk: allows ...
214
+ 2025-01-14 12:34:46,072 - llm_api - DEBUG - Generated chunk: ...
215
+ 2025-01-14 12:34:46,139 - llm_api - DEBUG - Generated chunk: it. ...
216
+ 2025-01-14 12:34:46,207 - llm_api - DEBUG - Generated chunk: This ...
217
+ 2025-01-14 12:34:46,273 - llm_api - DEBUG - Generated chunk: includes ...
218
+ 2025-01-14 12:34:46,341 - llm_api - DEBUG - Generated chunk: ...
219
+ 2025-01-14 12:34:46,408 - llm_api - DEBUG - Generated chunk: interpreting ...
220
+ 2025-01-14 12:34:46,473 - llm_api - DEBUG - Generated chunk: text ...
221
+ 2025-01-14 12:34:46,539 - llm_api - DEBUG - Generated chunk: inputs ...
222
+ 2025-01-14 12:34:46,605 - llm_api - DEBUG - Generated chunk: from ...
223
+ 2025-01-14 12:34:46,670 - llm_api - DEBUG - Generated chunk: various ...
224
+ 2025-01-14 12:34:46,735 - llm_api - DEBUG - Generated chunk: sources ...
225
+ 2025-01-14 12:34:46,801 - llm_api - DEBUG - Generated chunk: such ...
226
+ 2025-01-14 12:34:46,867 - llm_api - DEBUG - Generated chunk: as ...
227
+ 2025-01-14 12:34:46,935 - llm_api - DEBUG - Generated chunk: ...
228
+ 2025-01-14 12:34:47,005 - llm_api - DEBUG - Generated chunk: websites, ...
229
+ 2025-01-14 12:34:47,073 - llm_api - DEBUG - Generated chunk: ...
230
+ 2025-01-14 12:34:47,142 - llm_api - DEBUG - Generated chunk: books, ...
231
+ 2025-01-14 12:34:47,208 - llm_api - DEBUG - Generated chunk: articles ...
232
+ 2025-01-14 12:34:47,275 - llm_api - DEBUG - Generated chunk: ...
233
+ 2025-01-14 12:34:47,341 - llm_api - DEBUG - Generated chunk: etc., ...
234
+ 2025-01-14 12:34:47,408 - llm_api - DEBUG - Generated chunk: in ...
235
+ 2025-01-14 12:34:47,472 - llm_api - DEBUG - Generated chunk: multiple ...
236
+ 2025-01-14 12:34:47,536 - llm_api - DEBUG - Generated chunk: formats ...
237
+ 2025-01-14 12:34:47,600 - llm_api - DEBUG - Generated chunk: including ...
238
+ 2025-01-14 12:34:47,674 - llm_api - DEBUG - Generated chunk: ...
239
+ 2025-01-14 12:34:47,744 - llm_api - DEBUG - Generated chunk: English, ...
240
+ 2025-01-14 12:34:47,814 - llm_api - DEBUG - Generated chunk: ...
241
+ 2025-01-14 12:34:47,901 - llm_api - DEBUG - Generated chunk: Spanish, ...
242
+ 2025-01-14 12:34:47,991 - llm_api - DEBUG - Generated chunk: French ...
243
+ 2025-01-14 12:34:48,066 - llm_api - DEBUG - Generated chunk: among ...
244
+ 2025-01-14 12:34:48,131 - llm_api - DEBUG - Generated chunk: ...
245
+ 2025-01-14 12:34:48,194 - llm_api - DEBUG - Generated chunk: others. ...
246
+ 2025-01-14 12:34:48,259 - llm_api - DEBUG - Generated chunk:
247
+ ...
248
+ 2025-01-14 12:34:48,325 - llm_api - DEBUG - Generated chunk:
249
+ ...
250
+ 2025-01-14 12:34:48,390 - llm_api - DEBUG - Generated chunk: ...
251
+ 2025-01-14 12:34:48,454 - llm_api - DEBUG - Generated chunk: ...
252
+ 2025-01-14 12:34:48,519 - llm_api - DEBUG - Generated chunk: 2. ...
253
+ 2025-01-14 12:34:48,583 - llm_api - DEBUG - Generated chunk: ...
254
+ 2025-01-14 12:34:48,647 - llm_api - DEBUG - Generated chunk: **Text ...
255
+ 2025-01-14 12:34:48,711 - llm_api - DEBUG - Generated chunk: Generation ...
256
+ 2025-01-14 12:34:48,773 - llm_api - DEBUG - Generated chunk: & ...
257
+ 2025-01-14 12:34:48,837 - llm_api - DEBUG - Generated chunk: ...
258
+ 2025-01-14 12:34:48,902 - llm_api - DEBUG - Generated chunk: ...
259
+ 2025-01-14 12:34:48,965 - llm_api - DEBUG - Generated chunk: ...
260
+ 2025-01-14 12:34:49,031 - llm_api - DEBUG - Generated chunk: ...
261
+ 2025-01-14 12:34:49,096 - llm_api - DEBUG - Generated chunk: ...
262
+ 2025-01-14 12:34:49,160 - llm_api - DEBUG - Generated chunk: Comprehension**: ...
263
+ 2025-01-14 12:34:49,224 - llm_api - DEBUG - Generated chunk: Based ...
264
+ 2025-01-14 12:34:49,289 - llm_api - DEBUG - Generated chunk: on ...
265
+ 2025-01-14 12:34:49,353 - llm_api - DEBUG - Generated chunk: patterns ...
266
+ 2025-01-14 12:34:49,419 - llm_api - DEBUG - Generated chunk: learned ...
267
+ 2025-01-14 12:34:49,488 - llm_api - DEBUG - Generated chunk: during ...
268
+ 2025-01-14 12:34:49,555 - llm_api - DEBUG - Generated chunk: training ...
269
+ 2025-01-14 12:34:49,621 - llm_api - DEBUG - Generated chunk: with ...
270
+ 2025-01-14 12:34:49,686 - llm_api - DEBUG - Generated chunk: diverse ...
271
+ 2025-01-14 12:34:49,750 - llm_api - DEBUG - Generated chunk: internet ...
272
+ 2025-01-14 12:34:49,814 - llm_api - DEBUG - Generated chunk: texts ...
273
+ 2025-01-14 12:34:49,879 - llm_api - DEBUG - Generated chunk: data ...
274
+ 2025-01-14 12:34:49,944 - llm_api - DEBUG - Generated chunk: ...
275
+ 2025-01-14 12:34:50,009 - llm_api - DEBUG - Generated chunk: sets, ...
276
+ 2025-01-14 12:34:50,074 - llm_api - DEBUG - Generated chunk: I ...
277
+ 2025-01-14 12:34:50,140 - llm_api - DEBUG - Generated chunk: can ...
278
+ 2025-01-14 12:34:50,204 - llm_api - DEBUG - Generated chunk: generate ...
279
+ 2025-01-14 12:34:50,269 - llm_api - DEBUG - Generated chunk: ...
280
+ 2025-01-14 12:34:50,336 - llm_api - DEBUG - Generated chunk: ...
281
+ 2025-01-14 12:34:50,401 - llm_api - DEBUG - Generated chunk: coherent ...
282
+ 2025-01-14 12:34:50,466 - llm_api - DEBUG - Generated chunk: responses ...
283
+ 2025-01-14 12:34:50,530 - llm_api - DEBUG - Generated chunk: that ...
284
+ 2025-01-14 12:34:50,595 - llm_api - DEBUG - Generated chunk: ...
285
+ 2025-01-14 12:34:50,658 - llm_api - DEBUG - Generated chunk: ...
286
+ 2025-01-14 12:34:50,722 - llm_api - DEBUG - Generated chunk: mimic ...
287
+ 2025-01-14 12:34:50,787 - llm_api - DEBUG - Generated chunk: how ...
288
+ 2025-01-14 12:34:50,852 - llm_api - DEBUG - Generated chunk: real ...
289
+ 2025-01-14 12:34:50,916 - llm_api - DEBUG - Generated chunk: people ...
290
+ 2025-01-14 12:34:50,980 - llm_api - DEBUG - Generated chunk: ...
291
+ 2025-01-14 12:34:51,043 - llm_api - DEBUG - Generated chunk: write. ...
292
+ 2025-01-14 12:34:51,107 - llm_api - DEBUG - Generated chunk: ...
293
+ 2025-01-14 12:34:51,171 - llm_api - DEBUG - Generated chunk: However, ...
294
+ 2025-01-14 12:34:51,236 - llm_api - DEBUG - Generated chunk: please ...
295
+ 2025-01-14 12:34:51,299 - llm_api - DEBUG - Generated chunk: note ...
296
+ 2025-01-14 12:34:51,364 - llm_api - DEBUG - Generated chunk: these ...
297
+ 2025-01-14 12:34:51,428 - llm_api - DEBUG - Generated chunk: generated ...
298
+ 2025-01-14 12:34:51,491 - llm_api - DEBUG - Generated chunk: outputs ...
299
+ 2025-01-14 12:34:51,556 - llm_api - DEBUG - Generated chunk: ...
300
+ 2025-01-14 12:34:51,620 - llm_api - DEBUG - Generated chunk: ...
301
+ 2025-01-14 12:34:51,685 - llm_api - DEBUG - Generated chunk: aren’t ...
302
+ 2025-01-14 12:34:51,748 - llm_api - DEBUG - Generated chunk: perfect ...
303
+ 2025-01-14 12:34:51,813 - llm_api - DEBUG - Generated chunk: nor ...
304
+ 2025-01-14 12:34:51,877 - llm_api - DEBUG - Generated chunk: fully ...
305
+ 2025-01-14 12:34:51,940 - llm_api - DEBUG - Generated chunk: accurate ...
306
+ 2025-01-14 12:34:52,005 - llm_api - DEBUG - Generated chunk: but ...
307
+ 2025-01-14 12:34:52,068 - llm_api - DEBUG - Generated chunk: they ...
308
+ 2025-01-14 12:34:52,131 - llm_api - DEBUG - Generated chunk: often ...
309
+ 2025-01-14 12:34:52,196 - llm_api - DEBUG - Generated chunk: make ...
310
+ 2025-01-14 12:34:52,260 - llm_api - DEBUG - Generated chunk: sense ...
311
+ 2025-01-14 12:34:52,324 - llm_api - DEBUG - Generated chunk: ...
312
+ 2025-01-14 12:34:52,388 - llm_api - DEBUG - Generated chunk: contextually ...
313
+ 2025-01-14 12:34:52,451 - llm_api - DEBUG - Generated chunk: within ...
314
+ 2025-01-14 12:34:52,516 - llm_api - DEBUG - Generated chunk: given ...
315
+ 2025-01-14 12:34:52,579 - llm_api - DEBUG - Generated chunk: ...
316
+ 2025-01-14 12:34:52,643 - llm_api - DEBUG - Generated chunk: ...
317
+ 2025-01-14 12:34:52,707 - llm_api - DEBUG - Generated chunk: prompts.
318
+ ...
319
+ 2025-01-14 12:34:52,771 - llm_api - DEBUG - Generated chunk:
320
+ ...
321
+ 2025-01-14 12:34:52,835 - llm_api - DEBUG - Generated chunk: ...
322
+ 2025-01-14 12:34:52,899 - llm_api - DEBUG - Generated chunk: ...
323
+ 2025-01-14 12:34:52,963 - llm_api - DEBUG - Generated chunk: 3. ...
324
+ 2025-01-14 12:34:53,026 - llm_api - DEBUG - Generated chunk: ...
325
+ 2025-01-14 12:34:53,095 - llm_api - DEBUG - Generated chunk: ...
326
+ 2025-01-14 12:34:53,168 - llm_api - DEBUG - Generated chunk: ...
327
+ 2025-01-14 12:34:53,237 - llm_api - DEBUG - Generated chunk: **Knowledge ...
328
+ 2025-01-14 12:34:53,302 - llm_api - DEBUG - Generated chunk: Base ...
329
+ 2025-01-14 12:34:53,371 - llm_api - DEBUG - Generated chunk: ...
330
+ 2025-01-14 12:34:53,440 - llm_api - DEBUG - Generated chunk: ...
331
+ 2025-01-14 12:34:53,508 - llm_api - DEBUG - Generated chunk: ...
332
+ 2025-01-14 12:34:53,576 - llm_api - DEBUG - Generated chunk: Accessing**: ...
333
+ 2025-01-14 12:34:53,642 - llm_api - DEBUG - Generated chunk: Although ...
334
+ 2025-01-14 12:34:53,710 - llm_api - DEBUG - Generated chunk: not ...
335
+ 2025-01-14 12:34:53,775 - llm_api - DEBUG - Generated chunk: connected ...
336
+ 2025-01-14 12:34:53,838 - llm_api - DEBUG - Generated chunk: live ...
337
+ 2025-01-14 12:34:53,903 - llm_api - DEBUG - Generated chunk: for ...
338
+ 2025-01-14 12:34:53,968 - llm_api - DEBUG - Generated chunk: ...
339
+ 2025-01-14 12:34:54,032 - llm_api - DEBUG - Generated chunk: browsing ...
340
+ 2025-01-14 12:34:54,096 - llm_api - DEBUG - Generated chunk: external ...
341
+ 2025-01-14 12:34:54,159 - llm_api - DEBUG - Generated chunk: databases ...
342
+ 2025-01-14 12:34:54,224 - llm_api - DEBUG - Generated chunk: at ...
343
+ 2025-01-14 12:34:54,287 - llm_api - DEBUG - Generated chunk: this ...
344
+ 2025-01-14 12:34:54,351 - llm_api - DEBUG - Generated chunk: ...
345
+ 2025-01-14 12:34:54,414 - llm_api - DEBUG - Generated chunk: moment, ...
346
+ 2025-01-14 12:34:54,478 - llm_api - DEBUG - Generated chunk: information ...
347
+ 2025-01-14 12:34:54,542 - llm_api - DEBUG - Generated chunk: up ...
348
+ 2025-01-14 12:34:54,606 - llm_api - DEBUG - Generated chunk: until ...
349
+ 2025-01-14 12:34:54,670 - llm_api - DEBUG - Generated chunk: September ...
350
+ 2025-01-14 12:34:54,735 - llm_api - DEBUG - Generated chunk: ...
351
+ 2025-01-14 12:34:54,799 - llm_api - DEBUG - Generated chunk: ...
352
+ 2025-01-14 12:34:54,864 - llm_api - DEBUG - Generated chunk: ...
353
+ 2025-01-14 12:34:54,928 - llm_api - DEBUG - Generated chunk: ...
354
+ 2025-01-14 12:34:54,992 - llm_api - DEBUG - Generated chunk: 2021 ...
355
+ 2025-01-14 12:34:55,056 - llm_api - DEBUG - Generated chunk: has ...
356
+ 2025-01-14 12:34:55,120 - llm_api - DEBUG - Generated chunk: been ...
357
+ 2025-01-14 12:34:55,184 - llm_api - DEBUG - Generated chunk: used ...
358
+ 2025-01-14 12:34:55,249 - llm_api - DEBUG - Generated chunk: when ...
359
+ 2025-01-14 12:34:55,313 - llm_api - DEBUG - Generated chunk: generating ...
360
+ 2025-01-14 12:34:55,376 - llm_api - DEBUG - Generated chunk: answers ...
361
+ 2025-01-14 12:34:55,441 - llm_api - DEBUG - Generated chunk: based ...
362
+ 2025-01-14 12:34:55,504 - llm_api - DEBUG - Generated chunk: upon ...
363
+ 2025-01-14 12:34:55,568 - llm_api - DEBUG - Generated chunk: extensive ...
364
+ 2025-01-14 12:34:55,632 - llm_api - DEBUG - Generated chunk: datasets ...
365
+ 2025-01-14 12:34:55,696 - llm_api - DEBUG - Generated chunk: which ...
366
+ 2025-01-14 12:34:55,760 - llm_api - DEBUG - Generated chunk: include ...
367
+ 2025-01-14 12:34:55,825 - llm_api - DEBUG - Generated chunk: facts ...
368
+ 2025-01-14 12:34:55,889 - llm_api - DEBUG - Generated chunk: known ...
369
+ 2025-01-14 12:34:55,953 - llm_api - DEBUG - Generated chunk: till ...
370
+ 2025-01-14 12:34:56,017 - llm_api - DEBUG - Generated chunk: then ...
371
+ 2025-01-14 12:34:56,082 - llm_api - DEBUG - Generated chunk: across ...
372
+ 2025-01-14 12:34:56,146 - llm_api - DEBUG - Generated chunk: numerous ...
373
+ 2025-01-14 12:34:56,210 - llm_api - DEBUG - Generated chunk: topics ...
374
+ 2025-01-14 12:34:56,281 - llm_api - DEBUG - Generated chunk: ...
375
+ 2025-01-14 12:34:56,346 - llm_api - DEBUG - Generated chunk: ranging ...
376
+ 2025-01-14 12:34:56,410 - llm_api - DEBUG - Generated chunk: from ...
377
+ 2025-01-14 12:34:56,474 - llm_api - DEBUG - Generated chunk: science ...
378
+ 2025-01-14 12:34:56,537 - llm_api - DEBUG - Generated chunk: to ...
379
+ 2025-01-14 12:34:56,631 - llm_api - DEBUG - Generated chunk: ...
380
+ 2025-01-14 12:34:56,730 - llm_api - DEBUG - Generated chunk: arts, ...
381
+ 2025-01-14 12:34:56,818 - llm_api - DEBUG - Generated chunk: ...
382
+ 2025-01-14 12:34:56,886 - llm_api - DEBUG - Generated chunk: history, ...
383
+ 2025-01-14 12:34:56,951 - llm_api - DEBUG - Generated chunk: ...
384
+ 2025-01-14 12:34:57,016 - llm_api - DEBUG - Generated chunk: culture, ...
385
+ 2025-01-14 12:34:57,080 - llm_api - DEBUG - Generated chunk: sports ...
386
+ 2025-01-14 12:34:57,145 - llm_api - DEBUG - Generated chunk: amongst ...
387
+ 2025-01-14 12:34:57,209 - llm_api - DEBUG - Generated chunk: many ...
388
+ 2025-01-14 12:34:57,273 - llm_api - DEBUG - Generated chunk: other ...
389
+ 2025-01-14 12:34:57,337 - llm_api - DEBUG - Generated chunk: ...
390
+ 2025-01-14 12:34:57,401 - llm_api - DEBUG - Generated chunk: fields.
391
+ ...
392
+ 2025-01-14 12:34:57,467 - llm_api - DEBUG - Generated chunk:
393
+ ...
394
+ 2025-01-14 12:34:57,531 - llm_api - DEBUG - Generated chunk: ...
395
+ 2025-01-14 12:34:57,595 - llm_api - DEBUG - Generated chunk: ...
396
+ 2025-01-14 12:34:57,659 - llm_api - DEBUG - Generated chunk: 4. ...
397
+ 2025-01-14 12:34:57,723 - llm_api - DEBUG - Generated chunk: ...
398
+ 2025-01-14 12:34:57,786 - llm_api - DEBUG - Generated chunk: **Problem ...
399
+ 2025-01-14 12:34:57,850 - llm_api - DEBUG - Generated chunk: ...
400
+ 2025-01-14 12:34:57,914 - llm_api - DEBUG - Generated chunk: Solving ...
401
+ 2025-01-14 12:34:57,978 - llm_api - DEBUG - Generated chunk: ...
402
+ 2025-01-14 12:34:58,042 - llm_api - DEBUG - Generated chunk: ...
403
+ 2025-01-14 12:34:58,106 - llm_api - DEBUG - Generated chunk: ...
404
+ 2025-01-14 12:34:58,168 - llm_api - DEBUG - Generated chunk: Skills**: ...
405
+ 2025-01-14 12:34:58,231 - llm_api - DEBUG - Generated chunk: In ...
406
+ 2025-01-14 12:34:58,294 - llm_api - DEBUG - Generated chunk: certain ...
407
+ 2025-01-14 12:34:58,357 - llm_api - DEBUG - Generated chunk: scenarios ...
408
+ 2025-01-14 12:34:58,421 - llm_api - DEBUG - Generated chunk: where ...
409
+ 2025-01-14 12:34:58,484 - llm_api - DEBUG - Generated chunk: logical ...
410
+ 2025-01-14 12:34:58,549 - llm_api - DEBUG - Generated chunk: reasoning ...
411
+ 2025-01-14 12:34:58,612 - llm_api - DEBUG - Generated chunk: might ...
412
+ 2025-01-14 12:34:58,676 - llm_api - DEBUG - Generated chunk: be ...
413
+ 2025-01-14 12:34:58,739 - llm_api - DEBUG - Generated chunk: required ...
414
+ 2025-01-14 12:34:58,802 - llm_api - DEBUG - Generated chunk: ...
415
+ 2025-01-14 12:34:58,867 - llm_api - DEBUG - Generated chunk: (like ...
416
+ 2025-01-14 12:34:58,930 - llm_api - DEBUG - Generated chunk: math ...
417
+ 2025-01-14 12:34:58,994 - llm_api - DEBUG - Generated chunk: ...
418
+ 2025-01-14 12:34:59,057 - llm_api - DEBUG - Generated chunk: problems), ...
419
+ 2025-01-14 12:34:59,120 - llm_api - DEBUG - Generated chunk: algorithms ...
420
+ 2025-01-14 12:34:59,184 - llm_api - DEBUG - Generated chunk: enable ...
421
+ 2025-01-14 12:34:59,248 - llm_api - DEBUG - Generated chunk: ...
422
+ 2025-01-14 12:34:59,312 - llm_api - DEBUG - Generated chunk: ...
423
+ 2025-01-14 12:34:59,376 - llm_api - DEBUG - Generated chunk: ...
424
+ 2025-01-14 12:34:59,439 - llm_api - DEBUG - Generated chunk: problem-solving ...
425
+ 2025-01-14 12:34:59,503 - llm_api - DEBUG - Generated chunk: ...
426
+ 2025-01-14 12:34:59,567 - llm_api - DEBUG - Generated chunk: abilities ...
427
+ 2025-01-14 12:34:59,631 - llm_api - DEBUG - Generated chunk: similar ...
428
+ 2025-01-14 12:34:59,695 - llm_api - DEBUG - Generated chunk: those ...
429
+ 2025-01-14 12:34:59,758 - llm_api - DEBUG - Generated chunk: found ...
430
+ 2025-01-14 12:34:59,822 - llm_api - DEBUG - Generated chunk: commonly ...
431
+ 2025-01-14 12:34:59,884 - llm_api - DEBUG - Generated chunk: seen ...
432
+ 2025-01-14 12:34:59,948 - llm_api - DEBUG - Generated chunk: in ...
433
+ 2025-01-14 12:35:00,011 - llm_api - DEBUG - Generated chunk: ...
434
+ 2025-01-14 12:35:00,075 - llm_api - DEBUG - Generated chunk: calculators ...
435
+ 2025-01-14 12:35:00,139 - llm_api - DEBUG - Generated chunk: yet ...
436
+ 2025-01-14 12:35:00,202 - llm_api - DEBUG - Generated chunk: without ...
437
+ 2025-01-14 12:35:00,266 - llm_api - DEBUG - Generated chunk: any ...
438
+ 2025-01-14 12:35:00,329 - llm_api - DEBUG - Generated chunk: direct ...
439
+ 2025-01-14 12:35:00,393 - llm_api - DEBUG - Generated chunk: interaction ...
440
+ 2025-01-14 12:35:00,458 - llm_api - DEBUG - Generated chunk: beyond ...
441
+ 2025-01-14 12:35:00,523 - llm_api - DEBUG - Generated chunk: what ...
442
+ 2025-01-14 12:35:00,590 - llm_api - DEBUG - Generated chunk: was ...
443
+ 2025-01-14 12:35:00,656 - llm_api - DEBUG - Generated chunk: provided ...
444
+ 2025-01-14 12:35:00,721 - llm_api - DEBUG - Generated chunk: initially ...
445
+ 2025-01-14 12:35:00,786 - llm_api - DEBUG - Generated chunk: into ...
446
+ 2025-01-14 12:35:00,850 - llm_api - DEBUG - Generated chunk: them ...
447
+ 2025-01-14 12:35:00,916 - llm_api - DEBUG - Generated chunk: - ...
448
+ 2025-01-14 12:35:00,980 - llm_api - DEBUG - Generated chunk: no ...
449
+ 2025-01-14 12:35:01,045 - llm_api - DEBUG - Generated chunk: memory ...
450
+ 2025-01-14 12:35:01,109 - llm_api - DEBUG - Generated chunk: ...
451
+ 2025-01-14 12:35:01,173 - llm_api - DEBUG - Generated chunk: retention ...
452
+ 2025-01-14 12:35:01,239 - llm_api - DEBUG - Generated chunk: after ...
453
+ 2025-01-14 12:35:01,305 - llm_api - DEBUG - Generated chunk: each ...
454
+ 2025-01-14 12:35:01,372 - llm_api - DEBUG - Generated chunk: session ...
455
+ 2025-01-14 12:35:01,438 - llm_api - DEBUG - Generated chunk: ends ...
456
+ 2025-01-14 12:35:01,501 - llm_api - DEBUG - Generated chunk: due ...
457
+ 2025-01-14 12:35:01,567 - llm_api - DEBUG - Generated chunk: to ...
458
+ 2025-01-14 12:35:01,630 - llm_api - DEBUG - Generated chunk: design ...
459
+ 2025-01-14 12:35:01,694 - llm_api - DEBUG - Generated chunk: ...
460
+ 2025-01-14 12:35:01,759 - llm_api - DEBUG - Generated chunk: considerations ...
461
+ 2025-01-14 12:35:01,823 - llm_api - DEBUG - Generated chunk: around ...
462
+ 2025-01-14 12:35:01,887 - llm_api - DEBUG - Generated chunk: ...
463
+ 2025-01-14 12:35:01,951 - llm_api - DEBUG - Generated chunk: privacy ...
464
+ 2025-01-14 12:35:02,015 - llm_api - DEBUG - Generated chunk: protection ...
465
+ 2025-01-14 12:35:02,079 - llm_api - DEBUG - Generated chunk: policies ...
466
+ 2025-01-14 12:35:02,142 - llm_api - DEBUG - Generated chunk: followed ...
467
+ 2025-01-14 12:35:02,207 - llm_api - DEBUG - Generated chunk: strictly ...
468
+ 2025-01-14 12:35:02,271 - llm_api - DEBUG - Generated chunk: ...
469
+ 2025-01-14 12:35:02,334 - llm_api - DEBUG - Generated chunk: ...
470
+ 2025-01-14 12:35:02,399 - llm_api - DEBUG - Generated chunk: adhered ...
471
+ 2025-01-14 12:35:02,462 - llm_api - DEBUG - Generated chunk: ...
472
+ 2025-01-14 12:35:02,527 - llm_api - DEBUG - Generated chunk: too.
473
+ ...
474
+ 2025-01-14 12:35:02,592 - llm_api - DEBUG - Generated chunk:
475
+ ...
476
+ 2025-01-14 12:35:02,654 - llm_api - DEBUG - Generated chunk: ...
477
+ 2025-01-14 12:35:02,716 - llm_api - DEBUG - Generated chunk: ...
478
+ 2025-01-14 12:35:02,781 - llm_api - DEBUG - Generated chunk: 5. ...
479
+ 2025-01-14 12:35:02,844 - llm_api - DEBUG - Generated chunk: ...
480
+ 2025-01-14 12:35:02,908 - llm_api - DEBUG - Generated chunk: ...
481
+ 2025-01-14 12:35:02,974 - llm_api - DEBUG - Generated chunk: ...
482
+ 2025-01-14 12:35:03,043 - llm_api - DEBUG - Generated chunk: **Learning ...
483
+ 2025-01-14 12:35:03,108 - llm_api - DEBUG - Generated chunk: ...
484
+ 2025-01-14 12:35:03,173 - llm_api - DEBUG - Generated chunk: ...
485
+ 2025-01-14 12:35:03,239 - llm_api - DEBUG - Generated chunk: ...
486
+ 2025-01-14 12:35:03,303 - llm_api - DEBUG - Generated chunk: Capabilities**: ...
487
+ 2025-01-14 12:35:03,369 - llm_api - DEBUG - Generated chunk: It ...
488
+ 2025-01-14 12:35:03,433 - llm_api - DEBUG - Generated chunk: should ...
489
+ 2025-01-14 12:35:03,498 - llm_api - DEBUG - Generated chunk: also ...
490
+ 2025-01-14 12:35:03,562 - llm_api - DEBUG - Generated chunk: be ...
491
+ 2025-01-14 12:35:03,626 - llm_api - DEBUG - Generated chunk: noted ...
492
+ 2025-01-14 12:35:03,690 - llm_api - DEBUG - Generated chunk: though ...
493
+ 2025-01-14 12:35:03,755 - llm_api - DEBUG - Generated chunk: there ...
494
+ 2025-01-14 12:35:03,819 - llm_api - DEBUG - Generated chunk: ...
495
+ 2025-01-14 12:35:03,884 - llm_api - DEBUG - Generated chunk: ...
496
+ 2025-01-14 12:35:03,949 - llm_api - DEBUG - Generated chunk: isn't ...
497
+ 2025-01-14 12:35:04,013 - llm_api - DEBUG - Generated chunk: learning ...
498
+ 2025-01-14 12:35:04,078 - llm_api - DEBUG - Generated chunk: happening ...
499
+ 2025-01-14 12:35:04,142 - llm_api - DEBUG - Generated chunk: per ...
500
+ 2025-01-14 12:35:04,206 - llm_api - DEBUG - Generated chunk: ...
501
+ 2025-01-14 12:35:04,270 - llm_api - DEBUG - Generated chunk: se; ...
502
+ 2025-01-14 12:35:04,334 - llm_api - DEBUG - Generated chunk: continuous ...
503
+ 2025-01-14 12:35:04,399 - llm_api - DEBUG - Generated chunk: improvements ...
504
+ 2025-01-14 12:35:04,463 - llm_api - DEBUG - Generated chunk: occur ...
505
+ 2025-01-14 12:35:04,528 - llm_api - DEBUG - Generated chunk: through ...
506
+ 2025-01-14 12:35:04,593 - llm_api - DEBUG - Generated chunk: updates ...
507
+ 2025-01-14 12:35:04,659 - llm_api - DEBUG - Generated chunk: made ...
508
+ 2025-01-14 12:35:04,724 - llm_api - DEBUG - Generated chunk: ...
509
+ 2025-01-14 12:35:04,790 - llm_api - DEBUG - Generated chunk: periodically ...
510
+ 2025-01-14 12:35:04,855 - llm_api - DEBUG - Generated chunk: ...
511
+ 2025-01-14 12:35:04,923 - llm_api - DEBUG - Generated chunk: reflective ...
512
+ 2025-01-14 12:35:04,993 - llm_api - DEBUG - Generated chunk: ...
513
+ 2025-01-14 12:35:05,061 - llm_api - DEBUG - Generated chunk: ...
514
+ 2025-01-14 12:35:05,129 - llm_api - DEBUG - Generated chunk: advancements ...
515
+ 2025-01-14 12:35:05,197 - llm_api - DEBUG - Generated chunk: achieved ...
516
+ 2025-01-14 12:35:05,265 - llm_api - DEBUG - Generated chunk: over ...
517
+ 2025-01-14 12:35:05,333 - llm_api - DEBUG - Generated chunk: time ...
518
+ 2025-01-14 12:35:05,400 - llm_api - DEBUG - Generated chunk: via ...
519
+ 2025-01-14 12:35:05,467 - llm_api - DEBUG - Generated chunk: machine ...
520
+ 2025-01-14 12:35:05,535 - llm_api - DEBUG - Generated chunk: learning ...
521
+ 2025-01-14 12:35:05,603 - llm_api - DEBUG - Generated chunk: techniques ...
522
+ 2025-01-14 12:35:05,672 - llm_api - DEBUG - Generated chunk: applied ...
523
+ 2025-01-14 12:35:05,742 - llm_api - DEBUG - Generated chunk: ...
524
+ 2025-01-14 12:35:05,812 - llm_api - DEBUG - Generated chunk: systematically ...
525
+ 2025-01-14 12:35:05,881 - llm_api - DEBUG - Generated chunk: throughout ...
526
+ 2025-01-14 12:35:05,951 - llm_api - DEBUG - Generated chunk: development ...
527
+ 2025-01-14 12:35:06,022 - llm_api - DEBUG - Generated chunk: phases ...
528
+ 2025-01-14 12:35:06,091 - llm_api - DEBUG - Generated chunk: ...
529
+ 2025-01-14 12:35:06,162 - llm_api - DEBUG - Generated chunk: aimed ...
530
+ 2025-01-14 12:35:06,231 - llm_api - DEBUG - Generated chunk: towards ...
531
+ 2025-01-14 12:35:06,301 - llm_api - DEBUG - Generated chunk: ...
532
+ 2025-01-14 12:35:06,370 - llm_api - DEBUG - Generated chunk: ...
533
+ 2025-01-14 12:35:06,439 - llm_api - DEBUG - Generated chunk: enhancing ...
534
+ 2025-01-14 12:35:06,505 - llm_api - DEBUG - Generated chunk: performance ...
535
+ 2025-01-14 12:35:06,569 - llm_api - DEBUG - Generated chunk: ...
536
+ 2025-01-14 12:35:06,634 - llm_api - DEBUG - Generated chunk: consistently ...
537
+ 2025-01-14 12:35:06,697 - llm_api - DEBUG - Generated chunk: while ...
538
+ 2025-01-14 12:35:06,761 - llm_api - DEBUG - Generated chunk: ...
539
+ 2025-01-14 12:35:06,825 - llm_api - DEBUG - Generated chunk: maintaining ...
540
+ 2025-01-14 12:35:06,888 - llm_api - DEBUG - Generated chunk: user ...
541
+ 2025-01-14 12:35:06,952 - llm_api - DEBUG - Generated chunk: trust ...
542
+ 2025-01-14 12:35:07,017 - llm_api - DEBUG - Generated chunk: simultaneously ...
543
+ 2025-01-14 12:35:07,081 - llm_api - DEBUG - Generated chunk: ...
544
+ 2025-01-14 12:35:07,145 - llm_api - DEBUG - Generated chunk: ensuring ...
545
+ 2025-01-14 12:35:07,210 - llm_api - DEBUG - Generated chunk: ...
546
+ 2025-01-14 12:35:07,273 - llm_api - DEBUG - Generated chunk: ethical ...
547
+ 2025-01-14 12:35:07,339 - llm_api - DEBUG - Generated chunk: ...
548
+ 2025-01-14 12:35:07,406 - llm_api - DEBUG - Generated chunk: ...
549
+ 2025-01-14 12:35:07,472 - llm_api - DEBUG - Generated chunk: guidelines ...
550
+ 2025-01-14 12:35:07,539 - llm_api - DEBUG - Generated chunk: remain ...
551
+ 2025-01-14 12:35:07,606 - llm_api - DEBUG - Generated chunk: ...
552
+ 2025-01-14 12:35:07,673 - llm_api - DEBUG - Generated chunk: ...
553
+ 2025-01-14 12:35:07,740 - llm_api - DEBUG - Generated chunk: ...
554
+ 2025-01-14 12:35:07,806 - llm_api - DEBUG - Generated chunk: uncompromised ...
555
+ 2025-01-14 12:35:07,871 - llm_api - DEBUG - Generated chunk: always ...
556
+ 2025-01-14 12:35:07,938 - llm_api - DEBUG - Generated chunk: ...
557
+ 2025-01-14 12:35:08,004 - llm_api - DEBUG - Generated chunk: ...
558
+ 2025-01-14 12:35:08,070 - llm_api - DEBUG - Generated chunk: prioritized ...
559
+ 2025-01-14 12:35:08,136 - llm_api - DEBUG - Generated chunk: above ...
560
+ 2025-01-14 12:35:08,203 - llm_api - DEBUG - Generated chunk: all ...
561
+ 2025-01-14 12:35:08,270 - llm_api - DEBUG - Generated chunk: ...
562
+ 2025-01-14 12:35:08,337 - llm_api - DEBUG - Generated chunk: else!
563
+ ...
564
+ 2025-01-14 12:35:08,404 - llm_api - DEBUG - Generated chunk:
565
+ ...
566
+ 2025-01-14 12:35:08,470 - llm_api - DEBUG - Generated chunk: ...
567
+ 2025-01-14 12:35:08,535 - llm_api - DEBUG - Generated chunk: ...
568
+ 2025-01-14 12:35:08,602 - llm_api - DEBUG - Generated chunk: 6. ...
569
+ 2025-01-14 12:35:08,667 - llm_api - DEBUG - Generated chunk: ...
570
+ 2025-01-14 12:35:08,731 - llm_api - DEBUG - Generated chunk: ...
571
+ 2025-01-14 12:35:08,796 - llm_api - DEBUG - Generated chunk: ...
572
+ 2025-01-14 12:35:08,860 - llm_api - DEBUG - Generated chunk: **Multilingual ...
573
+ 2025-01-14 12:35:08,927 - llm_api - DEBUG - Generated chunk: ...
574
+ 2025-01-14 12:35:08,990 - llm_api - DEBUG - Generated chunk: ...
575
+ 2025-01-14 12:35:09,054 - llm_api - DEBUG - Generated chunk: Support**: ...
576
+ 2025-01-14 12:35:09,119 - llm_api - DEBUG - Generated chunk: As ...
577
+ 2025-01-14 12:35:09,184 - llm_api - DEBUG - Generated chunk: mentioned ...
578
+ 2025-01-14 12:35:09,248 - llm_api - DEBUG - Generated chunk: earlier ...
579
+ 2025-01-14 12:35:09,311 - llm_api - DEBUG - Generated chunk: regarding ...
580
+ 2025-01-14 12:35:09,375 - llm_api - DEBUG - Generated chunk: Language ...
581
+ 2025-01-14 12:35:09,439 - llm_api - DEBUG - Generated chunk: ...
582
+ 2025-01-14 12:35:09,503 - llm_api - DEBUG - Generated chunk: comprehension ...
583
+ 2025-01-14 12:35:09,569 - llm_api - DEBUG - Generated chunk: skills ...
584
+ 2025-01-14 12:35:09,633 - llm_api - DEBUG - Generated chunk: – ...
585
+ 2025-01-14 12:35:09,698 - llm_api - DEBUG - Generated chunk: one ...
586
+ 2025-01-14 12:35:09,762 - llm_api - DEBUG - Generated chunk: significant ...
587
+ 2025-01-14 12:35:09,825 - llm_api - DEBUG - Generated chunk: advantage ...
588
+ 2025-01-14 12:35:09,890 - llm_api - DEBUG - Generated chunk: here ...
589
+ 2025-01-14 12:35:09,953 - llm_api - DEBUG - Generated chunk: lies ...
590
+ 2025-01-14 12:35:10,019 - llm_api - DEBUG - Generated chunk: ...
591
+ 2025-01-14 12:35:10,084 - llm_api - DEBUG - Generated chunk: ...
592
+ 2025-01-14 12:35:10,149 - llm_api - DEBUG - Generated chunk: multilanguage ...
593
+ 2025-01-14 12:35:10,213 - llm_api - DEBUG - Generated chunk: support ...
594
+ 2025-01-14 12:35:10,278 - llm_api - DEBUG - Generated chunk: allowing ...
595
+ 2025-01-14 12:35:10,342 - llm_api - DEBUG - Generated chunk: users ...
596
+ 2025-01-14 12:35:10,408 - llm_api - DEBUG - Generated chunk: ...
597
+ 2025-01-14 12:35:10,472 - llm_api - DEBUG - Generated chunk: worldwide ...
598
+ 2025-01-14 12:35:10,537 - llm_api - DEBUG - Generated chunk: ...
599
+ 2025-01-14 12:35:10,603 - llm_api - DEBUG - Generated chunk: ...
600
+ 2025-01-14 12:35:10,669 - llm_api - DEBUG - Generated chunk: irrespective ...
601
+ 2025-01-14 12:35:10,735 - llm_api - DEBUG - Generated chunk: ...
602
+ 2025-01-14 12:35:10,801 - llm_api - DEBUG - Generated chunk: geographical ...
603
+ 2025-01-14 12:35:10,866 - llm_api - DEBUG - Generated chunk: location ...
604
+ 2025-01-14 12:35:10,931 - llm_api - DEBUG - Generated chunk: ...
605
+ 2025-01-14 12:35:10,996 - llm_api - DEBUG - Generated chunk: accessibility ...
606
+ 2025-01-14 12:35:11,061 - llm_api - DEBUG - Generated chunk: ...
607
+ 2025-01-14 12:35:11,125 - llm_api - DEBUG - Generated chunk: facilitated ...
608
+ 2025-01-14 12:35:11,190 - llm_api - DEBUG - Generated chunk: smooth ...
609
+ 2025-01-14 12:35:11,255 - llm_api - DEBUG - Generated chunk: communication ...
610
+ 2025-01-14 12:35:11,327 - llm_api - DEBUG - Generated chunk: experience ...
611
+ 2025-01-14 12:35:11,393 - llm_api - DEBUG - Generated chunk: ...
612
+ 2025-01-14 12:35:11,460 - llm_api - DEBUG - Generated chunk: fostering ...
613
+ 2025-01-14 12:35:11,527 - llm_api - DEBUG - Generated chunk: ...
614
+ 2025-01-14 12:35:11,593 - llm_api - DEBUG - Generated chunk: inclusivity ...
615
+ 2025-01-14 12:35:11,699 - llm_api - DEBUG - Generated chunk: ...
616
+ 2025-01-14 12:35:11,794 - llm_api - DEBUG - Generated chunk: globally ...
617
+ 2025-01-14 12:35:11,887 - llm_api - DEBUG - Generated chunk: ...
618
+ 2025-01-14 12:35:11,960 - llm_api - DEBUG - Generated chunk: promoting ...
619
+ 2025-01-14 12:35:12,034 - llm_api - DEBUG - Generated chunk: cross ...
620
+ 2025-01-14 12:35:12,113 - llm_api - DEBUG - Generated chunk: cultural ...
621
+ 2025-01-14 12:35:12,184 - llm_api - DEBUG - Generated chunk: exchange ...
622
+ 2025-01-14 12:35:12,252 - llm_api - DEBUG - Generated chunk: effectively ...
623
+ 2025-01-14 12:35:12,331 - llm_api - DEBUG - Generated chunk: ...
624
+ 2025-01-14 12:35:12,401 - llm_api - DEBUG - Generated chunk: bridging ...
625
+ 2025-01-14 12:35:12,467 - llm_api - DEBUG - Generated chunk: ...
626
+ 2025-01-14 12:35:12,532 - llm_api - DEBUG - Generated chunk: gaps ...
627
+ 2025-01-14 12:35:12,598 - llm_api - DEBUG - Generated chunk: between ...
628
+ 2025-01-14 12:35:12,662 - llm_api - DEBUG - Generated chunk: different ...
629
+ 2025-01-14 12:35:12,726 - llm_api - DEBUG - Generated chunk: communities ...
630
+ 2025-01-14 12:35:12,792 - llm_api - DEBUG - Generated chunk: everywhere ...
631
+ 2025-01-14 12:35:12,857 - llm_api - DEBUG - Generated chunk: creating ...
632
+ 2025-01-14 12:35:12,922 - llm_api - DEBUG - Generated chunk: ...
633
+ 2025-01-14 12:35:12,986 - llm_api - DEBUG - Generated chunk: opportunities ...
634
+ 2025-01-14 12:35:13,053 - llm_api - DEBUG - Generated chunk: connecting ...
635
+ 2025-01-14 12:35:13,120 - llm_api - DEBUG - Generated chunk: hearts ...
636
+ 2025-01-14 12:35:13,186 - llm_api - DEBUG - Generated chunk: minds ...
637
+ 2025-01-14 12:35:13,252 - llm_api - DEBUG - Generated chunk: together ...
638
+ 2025-01-14 12:35:13,319 - llm_api - DEBUG - Generated chunk: ...
639
+ 2025-01-14 12:35:13,385 - llm_api - DEBUG - Generated chunk: ...
640
+ 2025-01-14 12:35:13,450 - llm_api - DEBUG - Generated chunk: harmoniously ...
641
+ 2025-01-14 12:35:13,516 - llm_api - DEBUG - Generated chunk: ...
642
+ 2025-01-14 12:35:13,583 - llm_api - DEBUG - Generated chunk: ...
643
+ 2025-01-14 12:35:13,648 - llm_api - DEBUG - Generated chunk: transcending ...
644
+ 2025-01-14 12:35:13,713 - llm_api - DEBUG - Generated chunk: boundaries ...
645
+ 2025-01-14 12:35:13,779 - llm_api - DEBUG - Generated chunk: ...
646
+ 2025-01-14 12:35:13,844 - llm_api - DEBUG - Generated chunk: effortlessly ...
647
+ 2025-01-14 12:35:13,909 - llm_api - DEBUG - Generated chunk: breaking ...
648
+ 2025-01-14 12:35:13,975 - llm_api - DEBUG - Generated chunk: down ...
649
+ 2025-01-14 12:35:14,042 - llm_api - DEBUG - Generated chunk: walls ...
650
+ 2025-01-14 12:35:14,111 - llm_api - DEBUG - Generated chunk: ...
651
+ 2025-01-14 12:35:14,176 - llm_api - DEBUG - Generated chunk: silencing ...
652
+ 2025-01-14 12:35:14,242 - llm_api - DEBUG - Generated chunk: voices ...
653
+ 2025-01-14 12:35:14,307 - llm_api - DEBUG - Generated chunk: ...
654
+ 2025-01-14 12:35:14,372 - llm_api - DEBUG - Generated chunk: suppressions ...
655
+ 2025-01-14 12:35:14,438 - llm_api - DEBUG - Generated chunk: fear ...
656
+ 2025-01-14 12:35:14,503 - llm_api - DEBUG - Generated chunk: ...
657
+ 2025-01-14 12:35:14,569 - llm_api - DEBUG - Generated chunk: dividing ...
658
+ 2025-01-14 12:35:14,570 - llm_api - DEBUG - Generated chunk: us...
659
+ 2025-01-14 12:35:14,570 - llm_api - INFO - INFO: 127.0.0.1:37118 - "POST /api/v1/generate/stream HTTP/1.1" 200 OK
660
+ 2025-01-14 12:47:53,791 - api_routes - INFO - Received generation request for prompt: Tell me about yourself and your capabilities...
661
+ 2025-01-14 12:47:53,791 - llm_api - DEBUG - Generating response for prompt: Tell me about yourself and your capabilities...
662
+ 2025-01-14 12:48:27,656 - llm_api - DEBUG - Generated response: I'm Phi, an AI developed by Microsoft. While I don...
663
+ 2025-01-14 12:48:27,656 - api_routes - INFO - Successfully generated response
664
+ 2025-01-14 12:48:27,656 - llm_api - INFO - INFO: 127.0.0.1:43528 - "POST /api/v1/generate HTTP/1.1" 200 OK
665
+ 2025-01-14 12:50:17,735 - api_routes - INFO - Received generation request for prompt: Please analyze this query and create a JSON respon...
666
+ 2025-01-14 12:50:17,735 - llm_api - DEBUG - Generating response for prompt: Please analyze this query and create a JSON respon...
667
+ 2025-01-14 12:50:31,906 - llm_api - DEBUG - Generated response: ```json
668
+ {
669
+ "original_query": "Who is Djengis Khan...
670
+ 2025-01-14 12:50:31,906 - api_routes - INFO - Successfully generated response
671
+ 2025-01-14 12:50:31,906 - llm_api - INFO - INFO: 127.0.0.1:50042 - "POST /api/v1/generate HTTP/1.1" 200 OK
672
+ 2025-01-14 13:08:38,951 - api_routes - INFO - Received request to download model: PowerInfer/SmallThinker-3B-Preview
673
+ 2025-01-14 13:08:38,951 - llm_api - INFO - Starting download of model: PowerInfer/SmallThinker-3B-Preview
674
+ 2025-01-14 13:08:38,951 - llm_api - INFO - Enabling stdout logging for download
675
+ 2025-01-14 13:11:52,350 - llm_api - INFO - Disabling stdout logging
676
+ 2025-01-14 13:11:52,350 - llm_api - INFO - Saving model to main/models/SmallThinker-3B-Preview
677
+ 2025-01-14 13:13:04,420 - llm_api - INFO - Successfully downloaded model: PowerInfer/SmallThinker-3B-Preview
678
+ 2025-01-14 13:13:05,175 - api_routes - INFO - Successfully downloaded model: PowerInfer/SmallThinker-3B-Preview
679
+ 2025-01-14 13:13:31,469 - api_routes - INFO - Received request to initialize model: PowerInfer/SmallThinker-3B-Preview
680
+ 2025-01-14 13:13:31,469 - llm_api - INFO - Initializing generation model: PowerInfer/SmallThinker-3B-Preview
681
+ 2025-01-14 13:13:31,472 - llm_api - INFO - Loading model from local path: main/models/SmallThinker-3B-Preview
682
+ 2025-01-14 13:13:31,909 - llm_api - ERROR - Failed to initialize generation model PowerInfer/SmallThinker-3B-Preview: Some modules are dispatched on the CPU or the disk. Make sure you have enough GPU RAM to fit the quantized model. If you want to dispatch the model on the CPU or the disk while keeping these modules in 32-bit, you need to set `llm_int8_enable_fp32_cpu_offload=True` and pass a custom `device_map` to `from_pretrained`. Check https://huggingface.co/docs/transformers/main/en/main_classes/quantization#offload-between-cpu-and-gpu for more details.
683
+ 2025-01-14 13:13:31,909 - api_routes - ERROR - Error initializing model: Some modules are dispatched on the CPU or the disk. Make sure you have enough GPU RAM to fit the quantized model. If you want to dispatch the model on the CPU or the disk while keeping these modules in 32-bit, you need to set `llm_int8_enable_fp32_cpu_offload=True` and pass a custom `device_map` to `from_pretrained`. Check https://huggingface.co/docs/transformers/main/en/main_classes/quantization#offload-between-cpu-and-gpu for more details.
684
+ 2025-01-14 13:14:36,924 - main - INFO - LLM API server started on 0.0.0.0:8001
685
+ 2025-01-14 13:14:49,486 - hf_validation - WARNING - No .env file found. Fine if you're on Huggingface, but you need one to run locally on your PC.
686
+ 2025-01-14 13:14:49,486 - hf_validation - ERROR - No HF_TOKEN found in environment variables
687
+ 2025-01-14 13:14:49,486 - main - INFO - Starting LLM API server
688
+ 2025-01-14 13:14:49,486 - llm_api - INFO - Initializing LLM API
689
+ 2025-01-14 13:14:49,486 - llm_api - INFO - LLM API initialized successfully
690
+ 2025-01-14 13:14:49,486 - api_routes - INFO - Router initialized with LLM API instance
691
+ 2025-01-14 13:14:49,490 - main - INFO - FastAPI application created successfully
692
+ 2025-01-14 13:14:56,382 - api_routes - INFO - Received request to initialize model: PowerInfer/SmallThinker-3B-Preview
693
+ 2025-01-14 13:14:56,383 - llm_api - INFO - Initializing generation model: PowerInfer/SmallThinker-3B-Preview
694
+ 2025-01-14 13:14:56,383 - llm_api - INFO - Loading model from local path: main/models/SmallThinker-3B-Preview
695
+ 2025-01-14 13:15:07,065 - llm_api - INFO - Successfully initialized generation model: PowerInfer/SmallThinker-3B-Preview
696
+ 2025-01-14 13:15:07,065 - api_routes - INFO - Successfully initialized model: PowerInfer/SmallThinker-3B-Preview
697
+ 2025-01-14 13:15:07,065 - llm_api - INFO - INFO: 127.0.0.1:40472 - "POST /api/v1/model/initialize?model_name=PowerInfer%2FSmallThinker-3B-Preview HTTP/1.1" 200 OK
698
+ 2025-01-14 13:16:09,874 - api_routes - INFO - Received generation request for prompt: Tell me about yourself and your capabilities...
699
+ 2025-01-14 13:16:09,874 - llm_api - DEBUG - Generating response for prompt: Tell me about yourself and your capabilities...
700
+ 2025-01-14 13:17:03,595 - llm_api - DEBUG - Generated response: I'm an AI developed by Alibaba, designed to assist...
701
+ 2025-01-14 13:17:03,595 - api_routes - INFO - Successfully generated response
702
+ 2025-01-14 13:17:03,595 - llm_api - INFO - INFO: 127.0.0.1:44786 - "POST /api/v1/generate HTTP/1.1" 200 OK
703
+ 2025-01-14 13:18:04,891 - main - INFO - LLM API server started on 0.0.0.0:8001
main/utils/logging.py CHANGED
@@ -23,6 +23,9 @@ class StreamToLogger:
23
  def enable(self):
24
  self.enabled = True
25
 
 
 
 
26
  def disable(self):
27
  self.enabled = False
28
 
 
23
  def enable(self):
24
  self.enabled = True
25
 
26
+ def isatty(self):
27
+ return False
28
+
29
  def disable(self):
30
  self.enabled = False
31
 
requirements.txt CHANGED
@@ -1,7 +1,7 @@
1
  accelerate==1.2.1
2
  annotated-types==0.7.0
3
  anyio==4.8.0
4
- bitsandbytes @ https://github.com/bitsandbytes-foundation/bitsandbytes/releases/download/continuous-release_multi-backend-refactor/bitsandbytes-0.44.1.dev0-py3-none-manylinux_2_24_x86_64.whl#sha256=66deda2b99cee0d4e52a183d9bac5c8e8618cd9b4d4933ccf23b908622d6b879
5
  certifi==2024.12.14
6
  charset-normalizer==3.4.1
7
  click==8.1.8
@@ -45,6 +45,7 @@ tokenizers==0.21.0
45
  torch==2.5.1
46
  tqdm==4.67.1
47
  transformers==4.47.1
 
48
  typing_extensions==4.12.2
49
  urllib3==2.3.0
50
  uvicorn==0.34.0
 
1
  accelerate==1.2.1
2
  annotated-types==0.7.0
3
  anyio==4.8.0
4
+ bitsandbytes==0.45.0
5
  certifi==2024.12.14
6
  charset-normalizer==3.4.1
7
  click==8.1.8
 
45
  torch==2.5.1
46
  tqdm==4.67.1
47
  transformers==4.47.1
48
+ triton==3.1.0
49
  typing_extensions==4.12.2
50
  urllib3==2.3.0
51
  uvicorn==0.34.0