Update README.md
Browse files
README.md
CHANGED
@@ -180,7 +180,7 @@ import torch
|
|
180 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
181 |
|
182 |
# Load model and tokenizer
|
183 |
-
model_name = "
|
184 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
185 |
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, device_map="auto")
|
186 |
|
@@ -221,7 +221,7 @@ import torch
|
|
221 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
222 |
|
223 |
class CharacterChat:
|
224 |
-
def __init__(self, model_path="
|
225 |
print(f"Loading model from {model_path}...")
|
226 |
self.tokenizer = AutoTokenizer.from_pretrained(model_path)
|
227 |
self.model = AutoModelForCausalLM.from_pretrained(
|
|
|
180 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
181 |
|
182 |
# Load model and tokenizer
|
183 |
+
model_name = "luvgpt/phi3-uncensored-chat"
|
184 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
185 |
model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.float16, device_map="auto")
|
186 |
|
|
|
221 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
222 |
|
223 |
class CharacterChat:
|
224 |
+
def __init__(self, model_path="luvgpt/phi3-uncensored-chat", persona=None):
|
225 |
print(f"Loading model from {model_path}...")
|
226 |
self.tokenizer = AutoTokenizer.from_pretrained(model_path)
|
227 |
self.model = AutoModelForCausalLM.from_pretrained(
|