|
import gradio as gr |
|
import pandas as pd |
|
import google.generativeai as genai |
|
import kagglehub |
|
import os |
|
|
|
|
|
path = kagglehub.dataset_download("fahmidachowdhury/food-adulteration-dataset") |
|
|
|
|
|
dataset_file = os.listdir(path)[0] |
|
path = os.path.join(path, dataset_file) |
|
|
|
|
|
gemapi = os.getenv("GeminiApi") |
|
genai.configure(api_key=gemapi) |
|
|
|
|
|
data = pd.read_csv(path) |
|
|
|
|
|
system_instruction = f""" |
|
You are a public assistant who specializes in food safety. You look at data and explain to the user any question they ask; here is your data: {str(data.to_json())} |
|
You are also a food expert in the Indian context. You act as a representative of the government or public agencies, always keeping the needs of the people at the forefront. |
|
You will try to help the customer launch a feedback review whenever they complain. You are to prepare a "markdown" report, which is detailed and can be sent to the company or restaurant. |
|
In case of a complaint or a grievance, you will act like a detective gathering necessary information from the user until you are satisfied. Once you gather all the info, you are supposed to generate a markdown report. |
|
Once the customer asks you to show them the markdown report, you will use the information given to you to generate it. |
|
You will ask the customer a single question at a time, which is relevant, and you will not repeat another question until you've generated the report. |
|
""" |
|
|
|
|
|
model_path = "gemini-1.5-flash" |
|
FoodSafetyAssistant = genai.GenerativeModel(model_path, system_instruction=system_instruction) |
|
|
|
|
|
chat_history = [] |
|
|
|
|
|
def respond(usertxt, chat_history): |
|
|
|
chat = FoodSafetyAssistant.start_chat(history=chat_history) |
|
|
|
|
|
response = chat.send_message(usertxt) |
|
|
|
|
|
chat_history.append({"role": "user", "content": usertxt}) |
|
chat_history.append({"role": "assistant", "content": response.text}) |
|
|
|
return response.text, chat_history |
|
|
|
|
|
def gradio_chat(usertxt, chat_history): |
|
response, updated_history = respond(usertxt, chat_history) |
|
return response, updated_history |
|
|
|
demo = gr.ChatInterface(fn=gradio_chat, inputs="text", outputs=["text", "state"]) |
|
|
|
|
|
if __name__ == "__main__": |
|
demo.launch() |
|
|