Spaces:
Running
on
Zero
Running
on
Zero
Update llava_llama3/serve/cli.py
Browse files
llava_llama3/serve/cli.py
CHANGED
@@ -70,7 +70,7 @@ def chat_llava(args, image_file, text, tokenizer, model, image_processor, contex
|
|
70 |
do_sample=True if args.temperature > 0 else False,
|
71 |
temperature=args.temperature,
|
72 |
max_new_tokens=args.max_new_tokens,
|
73 |
-
|
74 |
use_cache=True)
|
75 |
|
76 |
outputs = tokenizer.decode(output_ids[0]).strip()
|
|
|
70 |
do_sample=True if args.temperature > 0 else False,
|
71 |
temperature=args.temperature,
|
72 |
max_new_tokens=args.max_new_tokens,
|
73 |
+
streamer=streamer,
|
74 |
use_cache=True)
|
75 |
|
76 |
outputs = tokenizer.decode(output_ids[0]).strip()
|