Spaces:
Running
Running
import asyncio | |
from gpt_researcher.utils.llm import get_llm | |
from gpt_researcher import GPTResearcher | |
from dotenv import load_dotenv | |
load_dotenv() | |
async def main(): | |
# Example usage of get_llm function | |
llm_provider = "openai" | |
model = "gpt-3.5-turbo" | |
temperature = 0.7 | |
max_tokens = 1000 | |
llm = get_llm(llm_provider, model=model, temperature=temperature, max_tokens=max_tokens) | |
print(f"LLM Provider: {llm_provider}, Model: {model}, Temperature: {temperature}, Max Tokens: {max_tokens}") | |
print('llm: ',llm) | |
await test_llm(llm=llm) | |
async def test_llm(llm): | |
# Test the connection with a simple query | |
messages = [{"role": "user", "content": "sup?"}] | |
try: | |
response = await llm.get_chat_response(messages, stream=False) | |
print("LLM response:", response) | |
except Exception as e: | |
print(f"Error: {e}") | |
# Run the async function | |
asyncio.run(main()) |