Spaces:
Running
Running
File size: 943 Bytes
372531f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 |
import asyncio
from gpt_researcher.utils.llm import get_llm
from gpt_researcher import GPTResearcher
from dotenv import load_dotenv
load_dotenv()
async def main():
# Example usage of get_llm function
llm_provider = "openai"
model = "gpt-3.5-turbo"
temperature = 0.7
max_tokens = 1000
llm = get_llm(llm_provider, model=model, temperature=temperature, max_tokens=max_tokens)
print(f"LLM Provider: {llm_provider}, Model: {model}, Temperature: {temperature}, Max Tokens: {max_tokens}")
print('llm: ',llm)
await test_llm(llm=llm)
async def test_llm(llm):
# Test the connection with a simple query
messages = [{"role": "user", "content": "sup?"}]
try:
response = await llm.get_chat_response(messages, stream=False)
print("LLM response:", response)
except Exception as e:
print(f"Error: {e}")
# Run the async function
asyncio.run(main()) |