Update qwen_model.py
Browse files- qwen_model.py +2 -2
qwen_model.py
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
|
| 2 |
|
| 3 |
# Replace with your target Qwen model on Hugging Face
|
| 4 |
-
MODEL_NAME = "Qwen/Qwen2.5-7B-Instruct
|
| 5 |
|
| 6 |
# Initialize tokenizer and model
|
| 7 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, trust_remote_code=True)
|
|
@@ -18,7 +18,7 @@ qwen_pipeline = pipeline(
|
|
| 18 |
tokenizer=tokenizer
|
| 19 |
)
|
| 20 |
|
| 21 |
-
def generate_response(retrieved_texts, query, max_new_tokens=
|
| 22 |
"""
|
| 23 |
Generates a response based on the retrieved texts and query using Qwen.
|
| 24 |
Args:
|
|
|
|
| 1 |
from transformers import AutoTokenizer, AutoModelForCausalLM, pipeline
|
| 2 |
|
| 3 |
# Replace with your target Qwen model on Hugging Face
|
| 4 |
+
MODEL_NAME = "Qwen/Qwen2.5-7B-Instruct"
|
| 5 |
|
| 6 |
# Initialize tokenizer and model
|
| 7 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, trust_remote_code=True)
|
|
|
|
| 18 |
tokenizer=tokenizer
|
| 19 |
)
|
| 20 |
|
| 21 |
+
def generate_response(retrieved_texts, query, max_new_tokens=500):
|
| 22 |
"""
|
| 23 |
Generates a response based on the retrieved texts and query using Qwen.
|
| 24 |
Args:
|