Spaces:
Sleeping
Sleeping
Commit
·
a39a4c5
1
Parent(s):
35b43b4
Add Phi-3-mini-4k generator option
Browse files
app.py
CHANGED
|
@@ -226,8 +226,22 @@ def load_tinyllama():
|
|
| 226 |
)
|
| 227 |
|
| 228 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 229 |
GENERATORS = {
|
| 230 |
"tinyllama-1.1b-chat": GeneratorWrapper("tinyllama-1.1b-chat", load_tinyllama),
|
|
|
|
| 231 |
}
|
| 232 |
|
| 233 |
|
|
@@ -403,7 +417,7 @@ with gr.Blocks(title="MedDiscover") as demo:
|
|
| 403 |
model_dd = gr.Dropdown(
|
| 404 |
label="Generator Model",
|
| 405 |
choices=list(GENERATORS.keys()),
|
| 406 |
-
value="
|
| 407 |
interactive=True,
|
| 408 |
)
|
| 409 |
k_slider = gr.Slider(1, 10, value=3, step=1, label="Top-k chunks")
|
|
|
|
| 226 |
)
|
| 227 |
|
| 228 |
|
| 229 |
+
def load_phi3_mini():
|
| 230 |
+
use_cuda = torch.cuda.is_available()
|
| 231 |
+
device_map = "cuda" if use_cuda else "cpu"
|
| 232 |
+
dtype = torch.float16 if use_cuda else torch.float32
|
| 233 |
+
return pipeline(
|
| 234 |
+
"text-generation",
|
| 235 |
+
model="microsoft/Phi-3-mini-4k-instruct",
|
| 236 |
+
device_map=device_map,
|
| 237 |
+
torch_dtype=dtype,
|
| 238 |
+
trust_remote_code=True,
|
| 239 |
+
)
|
| 240 |
+
|
| 241 |
+
|
| 242 |
GENERATORS = {
|
| 243 |
"tinyllama-1.1b-chat": GeneratorWrapper("tinyllama-1.1b-chat", load_tinyllama),
|
| 244 |
+
"phi-3-mini-4k": GeneratorWrapper("phi-3-mini-4k", load_phi3_mini),
|
| 245 |
}
|
| 246 |
|
| 247 |
|
|
|
|
| 417 |
model_dd = gr.Dropdown(
|
| 418 |
label="Generator Model",
|
| 419 |
choices=list(GENERATORS.keys()),
|
| 420 |
+
value="phi-3-mini-4k",
|
| 421 |
interactive=True,
|
| 422 |
)
|
| 423 |
k_slider = gr.Slider(1, 10, value=3, step=1, label="Top-k chunks")
|