johnrachwanpruna commited on
Commit
4e59697
·
verified ·
1 Parent(s): 88db0e1

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -3
README.md CHANGED
@@ -45,17 +45,18 @@ import transformers
45
  import torch
46
 
47
  model = "PrunaAI/mattshumer-Hermes-2-Pro-11B-bnb-4bit"
48
- tokenizer = "mattshumer/Hermes-2-Pro-11B"
49
  messages = [{"role": "user", "content": "What is a large language model?"}]
50
 
51
- tokenizer = AutoTokenizer.from_pretrained(tokenizer)
52
  prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
53
  pipeline = transformers.pipeline(
54
  "text-generation",
55
  model=model,
56
  device_map="auto",
57
- cache_dir='/nfs/students/rachwan/delete_me/hermes'
58
  )
 
 
 
59
  ```
60
 
61
  outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
 
45
  import torch
46
 
47
  model = "PrunaAI/mattshumer-Hermes-2-Pro-11B-bnb-4bit"
 
48
  messages = [{"role": "user", "content": "What is a large language model?"}]
49
 
50
+ tokenizer = AutoTokenizer.from_pretrained(model)
51
  prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
52
  pipeline = transformers.pipeline(
53
  "text-generation",
54
  model=model,
55
  device_map="auto",
 
56
  )
57
+
58
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
59
+ print(outputs[0]["generated_text"])
60
  ```
61
 
62
  outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)