Spaces:
Sleeping
Sleeping
File size: 1,831 Bytes
e88637e 17da9d8 e88637e 2990eac 17da9d8 e88637e d1881be e88637e 2041cec e88637e 2041cec e88637e 14e3057 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 |
import torch
from transformers import pipeline
import gradio as gr
# ๋ชจ๋ธ ๋ก๋
def load_model(model_name="deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B"):
pipe = pipeline(
"text-generation",
model=model_name,
device_map="sequential",
torch_dtype=torch.float16,
trust_remote_code=True,
truncation=True,
max_new_tokens=2048,
model_kwargs={
"low_cpu_mem_usage": True,
"offload_folder": "offload"
}
)
return pipe
# ์ฑ๋ด ์๋ต ์์ฑ ํจ์
def chat(message, history):
pipe = load_model()
prompt = f"Human: {message}\n\nAssistant:"
# ๋ชจ๋ธ ์์ฑ
response = pipe(
prompt,
max_new_tokens=2048,
temperature=0.7,
do_sample=True,
truncation=True,
pad_token_id=50256
)
# ์๋ต ์ถ์ถ ๋ฐ ์ฒ๋ฆฌ
try:
bot_text = response[0]["generated_text"]
bot_text = bot_text.split("Assistant:")[-1].strip()
if "</think>" in bot_text:
bot_text = bot_text.split("</think>")[-1].strip()
except:
bot_text = "Sorry, there was a problem generating the response."
return bot_text
# Gradio ์ธํฐํ์ด์ค ์ค์
demo = gr.ChatInterface(
chat,
chatbot=gr.Chatbot(
height=600,
type="messages" # ๋ฉ์์ง ํ์ ์ง์
),
textbox=gr.Textbox(placeholder="Enter your message...", container=False, scale=7),
title="DeepSeek-R1 Chatbot",
description="DeepSeek-R1-Distill-Qwen-1.5B ๋ชจ๋ธ์ ์ฌ์ฉํ ๋ํ ํ
์คํธ์ฉ ๋ฐ๋ชจ์
๋๋ค.",
examples=["Hello", "Who are you?", "What can you do?"],
theme=gr.themes.Soft(),
type="messages" # ChatInterface์ type๋ ์ง์
)
# ์๋ฒ ์คํ
if __name__ == "__main__":
demo.launch(server_name="0.0.0.0") |