Update services/llama_service.py
Browse files
services/llama_service.py
CHANGED
@@ -5,8 +5,8 @@ from transformers import AutoTokenizer, AutoModelForCausalLM
|
|
5 |
# Load Llama model and tokenizer for text generation
|
6 |
|
7 |
|
8 |
-
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.2-
|
9 |
-
model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-3.2-
|
10 |
|
11 |
# Ensure the tokenizer has a padding token; if not, add a new padding token
|
12 |
if tokenizer.pad_token is None:
|
|
|
5 |
# Load Llama model and tokenizer for text generation
|
6 |
|
7 |
|
8 |
+
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-3.2-8B")
|
9 |
+
model = AutoModelForCausalLM.from_pretrained("meta-llama/Llama-3.2-8B")
|
10 |
|
11 |
# Ensure the tokenizer has a padding token; if not, add a new padding token
|
12 |
if tokenizer.pad_token is None:
|