Update README.md
Browse files
README.md
CHANGED
@@ -282,9 +282,9 @@ class LLM:
|
|
282 |
add_generation_prompt=True, # Must add for generation
|
283 |
return_tensors="pt",
|
284 |
).to("cuda")
|
285 |
-
|
286 |
# Increase the max_new_tokens to allow more detailed responses
|
287 |
-
output_ids = self.model.generate(inputs, max_new_tokens=self.max_new_tokens, pad_token_id=self.
|
288 |
output_ids = output_ids.tolist()[0] if output_ids.size(0) == 1 else output_ids.tolist()
|
289 |
|
290 |
output_text = self.__tokenizer.decode(output_ids, skip_special_tokens=True)
|
@@ -320,4 +320,7 @@ class LLM:
|
|
320 |
},
|
321 |
"methods": ["invoke", "extract_answer"]
|
322 |
}
|
|
|
|
|
|
|
323 |
```
|
|
|
282 |
add_generation_prompt=True, # Must add for generation
|
283 |
return_tensors="pt",
|
284 |
).to("cuda")
|
285 |
+
|
286 |
# Increase the max_new_tokens to allow more detailed responses
|
287 |
+
output_ids = self.model.generate(inputs, max_new_tokens=self.max_new_tokens, pad_token_id=self.__tokenizer.pad_token_id)
|
288 |
output_ids = output_ids.tolist()[0] if output_ids.size(0) == 1 else output_ids.tolist()
|
289 |
|
290 |
output_text = self.__tokenizer.decode(output_ids, skip_special_tokens=True)
|
|
|
320 |
},
|
321 |
"methods": ["invoke", "extract_answer"]
|
322 |
}
|
323 |
+
|
324 |
+
|
325 |
+
llm = LLM()
|
326 |
```
|