|
import gradio as gr |
|
from huggingface_hub import InferenceClient |
|
|
|
client = InferenceClient(model="https://9bsneexhe83nu6-80.proxy.runpod.net") |
|
|
|
def inference(message, history): |
|
partial_message = "" |
|
for token in client.text_generation(prompt=message, max_new_tokens=512, stream=True, best_of=1, temperature=0.3, |
|
top_p=0.99, do_sample=True, repetition_penalty=1.0): |
|
if token.startswith("<s>"): |
|
return partial_message |
|
partial_message += token |
|
yield partial_message |
|
|
|
gr.ChatInterface( |
|
inference, |
|
chatbot=gr.Chatbot(height=300, scale=7), |
|
textbox=gr.Textbox(placeholder="你可以问我任何关于SequioaDB的问题!", container=False, scale=7), |
|
description="这是SequioaDB旗下的AI智能大语言模型,训练超过上万条真实数据和7亿参数。", |
|
title="ChatSDB", |
|
examples=["SequoiaDB巨杉数据库是什么?", "SequoiaDB巨杉数据库支持哪些类型的数据库实例?"], |
|
retry_btn="重试", |
|
undo_btn="撤销", |
|
clear_btn="清除", |
|
submit_btn="提问", |
|
).queue().launch() |