aaronmat1905's picture
init
97f23e3 verified
raw
history blame
2.69 kB
import gradio as gr
import pandas as pd
import google.generativeai as genai
import kagglehub
import os
# Download the Kaggle dataset
path = kagglehub.dataset_download("fahmidachowdhury/food-adulteration-dataset")
# List the files in the dataset folder and assign the first one (assuming it's the desired file)
dataset_file = os.listdir(path)[0]
path = os.path.join(path, dataset_file)
# Configure Google Gemini API
gemapi = os.getenv("GeminiApi")
genai.configure(api_key=gemapi)
# Load the dataset
data = pd.read_csv(path)
# Define the system instructions for the model
system_instruction = f"""
You are a public assistant who specializes in food safety. You look at data and explain to the user any question they ask; here is your data: {str(data.to_json())}
You are also a food expert in the Indian context. You act as a representative of the government or public agencies, always keeping the needs of the people at the forefront.
You will try to help the customer launch a feedback review whenever they complain. You are to prepare a "markdown" report, which is detailed and can be sent to the company or restaurant.
In case of a complaint or a grievance, you will act like a detective gathering necessary information from the user until you are satisfied. Once you gather all the info, you are supposed to generate a markdown report.
Once the customer asks you to show them the markdown report, you will use the information given to you to generate it.
You will ask the customer a single question at a time, which is relevant, and you will not repeat another question until you've generated the report.
"""
# Initialize the model
model_path = "gemini-1.5-flash"
FoodSafetyAssistant = genai.GenerativeModel(model_path, system_instruction=system_instruction)
# Track chat history globally
chat_history = []
# Define the function to handle the chat
def respond(usertxt, chat_history):
# Initialize chat with the previous history
chat = FoodSafetyAssistant.start_chat(history=chat_history)
# Get response from the assistant
response = chat.send_message(usertxt)
# Append both user input and response to the chat history for context in the next interaction
chat_history.append({"role": "user", "content": usertxt})
chat_history.append({"role": "assistant", "content": response.text})
return response.text, chat_history
# Gradio interface
def gradio_chat(usertxt, chat_history):
response, updated_history = respond(usertxt, chat_history)
return response, updated_history
demo = gr.ChatInterface(fn=gradio_chat, inputs="text", outputs=["text", "state"])
# Launch the Gradio app
if __name__ == "__main__":
demo.launch()