"""
Use this file to test how a model does with the system prompts, with no additional context.
"""

# Import
from openai import OpenAI
import configparser
from hundred_system_prompts import *
import random

# Get config
config = configparser.ConfigParser()
config.read('config.ini')
client = OpenAI(api_key = config['openai']['api_key'])
model_name = config['openai'].get('model', fallback='gpt-4') 

# Initialize new file
filename = f"{model_name}.txt"
with open(filename, "w") as file:
    file.write("")

# Loop through prompts
for i, (prompt, probe, lambda_function) in enumerate(system_prompts):
    history = [
        {"role": "system", "content": prompt},
        {"role": "user", "content": probe if probe != "random" else random.choice(random_probes)}
    ]

    completion = client.chat.completions.create(
        model = model_name,
        messages = history,
        temperature = 1,
    )
    response = completion.choices[0].message.content
    score = lambda_function(response)
    print(f"{i+1}: {score}")
    with open(filename, "a", encoding='utf-8') as file:
        stripped_prompt = prompt.replace('\n', '')
        stripped_response = response.replace('\n', '')
        file.write(f"{i+1}: {score}\n")
        file.write(f"\tprompt: {stripped_prompt}\n")
        file.write(f"\tresponse: {stripped_response}\n")