import gradio as gr from huggingface_hub import InferenceClient import time # Inference client setup client = InferenceClient("HuggingFaceH4/zephyr-7b-beta") # Global flag to handle cancellation stop_inference = False def respond( message, history: list[tuple[str, str]], system_message, max_tokens, temperature, top_p, use_local_model, ): global stop_inference stop_inference = False # Reset cancellation flag if use_local_model: # Simulate local inference time.sleep(2) # simulate a delay response = "This is a response from the local model." history.append((message, response)) yield history else: # API-based inference messages = [{"role": "system", "content": system_message}] for val in history: if val[0]: messages.append({"role": "user", "content": val[0]}) if val[1]: messages.append({"role": "assistant", "content": val[1]}) messages.append({"role": "user", "content": message}) response = "" for message in client.chat_completion( messages, max_tokens=max_tokens, stream=True, temperature=temperature, top_p=top_p, ): if stop_inference: history.append((message, "Inference cancelled.")) yield history break token = message.choices[0].delta.content response += token history.append((message, response)) yield history def cancel_inference(): global stop_inference stop_inference = True # Custom CSS for a fancy look custom_css = """ #main-container { background-color: #f0f0f0; font-family: 'Arial', sans-serif; } .gradio-container { max-width: 700px; margin: 0 auto; padding: 20px; background: white; box-shadow: 0 4px 8px rgba(0, 0, 0, 0.1); border-radius: 10px; } .gr-button { background-color: #4CAF50; color: white; border: none; border-radius: 5px; padding: 10px 20px; cursor: pointer; transition: background-color 0.3s ease; } .gr-button:hover { background-color: #45a049; } .gr-slider input { color: #4CAF50; } .gr-chat { font-size: 16px; } #title { text-align: center; font-size: 2em; margin-bottom: 20px; color: #333; } """ # Define the interface with gr.Blocks(css=custom_css) as demo: gr.Markdown("