TobyYang7 commited on
Commit
315417c
·
verified ·
1 Parent(s): ca30e4f

Update llava_llama3/serve/cli.py

Browse files
Files changed (1) hide show
  1. llava_llama3/serve/cli.py +1 -1
llava_llama3/serve/cli.py CHANGED
@@ -70,7 +70,7 @@ def chat_llava(args, image_file, text, tokenizer, model, image_processor, contex
70
  do_sample=True if args.temperature > 0 else False,
71
  temperature=args.temperature,
72
  max_new_tokens=args.max_new_tokens,
73
- # streamer=streamer,
74
  use_cache=True)
75
 
76
  outputs = tokenizer.decode(output_ids[0]).strip()
 
70
  do_sample=True if args.temperature > 0 else False,
71
  temperature=args.temperature,
72
  max_new_tokens=args.max_new_tokens,
73
+ streamer=streamer,
74
  use_cache=True)
75
 
76
  outputs = tokenizer.decode(output_ids[0]).strip()