VatsalPatel18 commited on
Commit
a39a4c5
·
1 Parent(s): 35b43b4

Add Phi-3-mini-4k generator option

Browse files
Files changed (1) hide show
  1. app.py +15 -1
app.py CHANGED
@@ -226,8 +226,22 @@ def load_tinyllama():
226
  )
227
 
228
 
 
 
 
 
 
 
 
 
 
 
 
 
 
229
  GENERATORS = {
230
  "tinyllama-1.1b-chat": GeneratorWrapper("tinyllama-1.1b-chat", load_tinyllama),
 
231
  }
232
 
233
 
@@ -403,7 +417,7 @@ with gr.Blocks(title="MedDiscover") as demo:
403
  model_dd = gr.Dropdown(
404
  label="Generator Model",
405
  choices=list(GENERATORS.keys()),
406
- value="tinyllama-1.1b-chat",
407
  interactive=True,
408
  )
409
  k_slider = gr.Slider(1, 10, value=3, step=1, label="Top-k chunks")
 
226
  )
227
 
228
 
229
+ def load_phi3_mini():
230
+ use_cuda = torch.cuda.is_available()
231
+ device_map = "cuda" if use_cuda else "cpu"
232
+ dtype = torch.float16 if use_cuda else torch.float32
233
+ return pipeline(
234
+ "text-generation",
235
+ model="microsoft/Phi-3-mini-4k-instruct",
236
+ device_map=device_map,
237
+ torch_dtype=dtype,
238
+ trust_remote_code=True,
239
+ )
240
+
241
+
242
  GENERATORS = {
243
  "tinyllama-1.1b-chat": GeneratorWrapper("tinyllama-1.1b-chat", load_tinyllama),
244
+ "phi-3-mini-4k": GeneratorWrapper("phi-3-mini-4k", load_phi3_mini),
245
  }
246
 
247
 
 
417
  model_dd = gr.Dropdown(
418
  label="Generator Model",
419
  choices=list(GENERATORS.keys()),
420
+ value="phi-3-mini-4k",
421
  interactive=True,
422
  )
423
  k_slider = gr.Slider(1, 10, value=3, step=1, label="Top-k chunks")