Spaces:
Running
on
T4
Running
on
T4
sparkleman
commited on
Commit
·
efc2316
1
Parent(s):
2dd12cc
UPDATE: [0] prefill
Browse files
app.py
CHANGED
|
@@ -333,7 +333,7 @@ async def chatResponse(
|
|
| 333 |
)
|
| 334 |
logger.info(f"[REQ] {completionId} - prompt - {prompt}")
|
| 335 |
|
| 336 |
-
out, model_tokens, model_state = await runPrefill(request, prompt, [], model_state)
|
| 337 |
|
| 338 |
prefillTime = time.time()
|
| 339 |
promptTokenCount = len(model_tokens)
|
|
@@ -415,9 +415,9 @@ async def chatResponseStream(
|
|
| 415 |
else request.prompt.strip()
|
| 416 |
)
|
| 417 |
|
| 418 |
-
logger.info(f"[REQ] {completionId} - context
|
| 419 |
|
| 420 |
-
out, model_tokens, model_state = await runPrefill(request, prompt, [], model_state)
|
| 421 |
|
| 422 |
prefillTime = time.time()
|
| 423 |
promptTokenCount = len(model_tokens)
|
|
|
|
| 333 |
)
|
| 334 |
logger.info(f"[REQ] {completionId} - prompt - {prompt}")
|
| 335 |
|
| 336 |
+
out, model_tokens, model_state = await runPrefill(request, prompt, [0], model_state)
|
| 337 |
|
| 338 |
prefillTime = time.time()
|
| 339 |
promptTokenCount = len(model_tokens)
|
|
|
|
| 415 |
else request.prompt.strip()
|
| 416 |
)
|
| 417 |
|
| 418 |
+
logger.info(f"[REQ] {completionId} - context\n```{prompt}```")
|
| 419 |
|
| 420 |
+
out, model_tokens, model_state = await runPrefill(request, prompt, [0], model_state)
|
| 421 |
|
| 422 |
prefillTime = time.time()
|
| 423 |
promptTokenCount = len(model_tokens)
|