import gradio as gr from huggingface_hub import InferenceClient client = InferenceClient(model="https://9bsneexhe83nu6-80.proxy.runpod.net") def inference(message, history): partial_message = "" for token in client.text_generation(prompt=message, max_new_tokens=512, stream=True, best_of=1, temperature=0.3, top_p=0.99, do_sample=True, repetition_penalty=1.0): if token.startswith("<s>"): return partial_message partial_message += token yield partial_message gr.ChatInterface( inference, chatbot=gr.Chatbot(height=300, scale=7), textbox=gr.Textbox(placeholder="你可以问我任何关于SequioaDB的问题!", container=False, scale=7), description="这是SequioaDB旗下的AI智能大语言模型,训练超过上万条真实数据和7亿参数。", title="ChatSDB", examples=["SequoiaDB巨杉数据库是什么?", "SequoiaDB巨杉数据库支持哪些类型的数据库实例?"], retry_btn="重试", undo_btn="撤销", clear_btn="清除", submit_btn="提问", ).queue().launch()