Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -22,7 +22,7 @@ from torch import nn
|
|
22 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
23 |
|
24 |
tokenizer = AutoTokenizer.from_pretrained("tmberooney/medllama-merged")
|
25 |
-
model = AutoModelForCausalLM.from_pretrained("tmberooney/medllama-merged"
|
26 |
'''device_map = {"transformer.word_embeddings": "cpu",
|
27 |
"transformer.word_embeddings_layernorm": "cpu",
|
28 |
"lm_head": "cpu",
|
@@ -48,7 +48,7 @@ llama_pipeline = pipeline(
|
|
48 |
"text-generation", # LLM task
|
49 |
model=model,
|
50 |
torch_dtype=torch.float16,
|
51 |
-
device_map="
|
52 |
tokenizer=tokenizer
|
53 |
)
|
54 |
|
|
|
22 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
23 |
|
24 |
tokenizer = AutoTokenizer.from_pretrained("tmberooney/medllama-merged")
|
25 |
+
model = AutoModelForCausalLM.from_pretrained("tmberooney/medllama-merged")
|
26 |
'''device_map = {"transformer.word_embeddings": "cpu",
|
27 |
"transformer.word_embeddings_layernorm": "cpu",
|
28 |
"lm_head": "cpu",
|
|
|
48 |
"text-generation", # LLM task
|
49 |
model=model,
|
50 |
torch_dtype=torch.float16,
|
51 |
+
device_map="cpu",
|
52 |
tokenizer=tokenizer
|
53 |
)
|
54 |
|