liweixin commited on
Commit
41dd48c
·
1 Parent(s): b170295

Create finetuning.py

Browse files
Files changed (1) hide show
  1. finetuning.py +16 -0
finetuning.py ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ from transformers import AutoTokenizer, AutoModelForSequenceClassification
2
+
3
+ # 加载预训练的模型和分词器
4
+ tokenizer = AutoTokenizer.from_pretrained("bert-base-uncased-finetuned-sst-2-english")
5
+ model = AutoModelForSequenceClassification.from_pretrained("bert-base-uncased-finetuned-sst-2-english")
6
+
7
+ # 对输入文本进行分词和编码
8
+ inputs = tokenizer("I love Transformers!", return_tensors="pt")
9
+
10
+ # 用模型进行推理,得到预测的标签和分数
11
+ outputs = model(**inputs)
12
+ label = outputs.logits.argmax(-1).item()
13
+ score = outputs.logits.max(-1).values.item()
14
+
15
+ # 打印结果
16
+ print(f"Label: {label}, Score: {score}")