juyoung-trl commited on
Commit
7f64c04
ยท
verified ยท
1 Parent(s): b3b8c5a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +27 -7
README.md CHANGED
@@ -48,13 +48,33 @@ Where `{lang_tag} โˆˆ { en, ko, ja, zh }`.
48
  ```python
49
  from transformers import AutoTokenizer, AutoModelForCausalLM
50
 
51
- tok = AutoTokenizer.from_pretrained("trillionlabs/Tri-1.8B-Translation")
52
- model = AutoModelForCausalLM.from_pretrained("trillionlabs/Tri-1.8B-Translation", device_map="auto")
53
-
54
- prompt = "Translate the following Korean text into English:\n์•ˆ๋…•ํ•˜์„ธ์š” <en>"
55
- inputs = tok(prompt, return_tensors="pt").to(model.device)
56
- out = model.generate(**inputs, max_new_tokens=128)
57
- print(tok.decode(out[0], skip_special_tokens=True))
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
58
  ```
59
 
60
  ---
 
48
  ```python
49
  from transformers import AutoTokenizer, AutoModelForCausalLM
50
 
51
+ model_name = "trillionlabs/Tri-1.8B-Translation"
52
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
53
+ model = AutoModelForCausalLM.from_pretrained(model_name, device_map="auto")
54
+
55
+ text = "์•ˆ๋…•ํ•˜์„ธ์š”"
56
+ messages = [
57
+ {"role": "user", "content": f"Translate the following Korean text into English:\n{text} <en>"}
58
+ ]
59
+
60
+ inputs = tokenizer.apply_chat_template(
61
+ messages,
62
+ return_tensors="pt",
63
+ add_generation_prompt=True
64
+ ).to(model.device)
65
+
66
+ outputs = model.generate(
67
+ inputs,
68
+ max_new_tokens=256,
69
+ do_sample=False,
70
+ pad_token_id=tokenizer.eos_token_id
71
+ )
72
+
73
+ full_output = tokenizer.decode(outputs[0], skip_special_tokens=True)
74
+ translation = tokenizer.decode(outputs[0][len(inputs[0]):], skip_special_tokens=True)
75
+
76
+ print(f"Korean: {text}")
77
+ print(f"English: {translation}")
78
  ```
79
 
80
  ---