fdfsdfs
Browse files
app.py
CHANGED
@@ -16,11 +16,16 @@ before = datetime.datetime.now()
|
|
16 |
# Load model directly
|
17 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
18 |
|
19 |
-
tokenizer = AutoTokenizer.from_pretrained("01-ai/Yi-1.5-6B-Chat")
|
20 |
-
model = AutoModelForCausalLM.from_pretrained("01-ai/Yi-1.5-6B-Chat")
|
|
|
|
|
|
|
|
|
|
|
21 |
|
22 |
st.write('tokenizando...')
|
23 |
-
prompt = "
|
24 |
inputs = tokenizer(prompt, return_tensors="pt")
|
25 |
|
26 |
# Generate
|
|
|
16 |
# Load model directly
|
17 |
from transformers import AutoTokenizer, AutoModelForCausalLM
|
18 |
|
19 |
+
# tokenizer = AutoTokenizer.from_pretrained("01-ai/Yi-1.5-6B-Chat")
|
20 |
+
# model = AutoModelForCausalLM.from_pretrained("01-ai/Yi-1.5-6B-Chat")
|
21 |
+
|
22 |
+
from transformers import AutoTokenizer, AutoModelForQuestionAnswering
|
23 |
+
|
24 |
+
tokenizer = AutoTokenizer.from_pretrained("deepset/roberta-base-squad2")
|
25 |
+
model = AutoModelForQuestionAnswering.from_pretrained("deepset/roberta-base-squad2")
|
26 |
|
27 |
st.write('tokenizando...')
|
28 |
+
prompt = "Qual é o maior planeta do sistema solar ?"
|
29 |
inputs = tokenizer(prompt, return_tensors="pt")
|
30 |
|
31 |
# Generate
|