oguzhandoganoglu commited on
Commit
f26facc
·
verified ·
1 Parent(s): 6337b9d

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +18 -22
app.py CHANGED
@@ -9,24 +9,25 @@ from llama_cpp_agent.chat_history.messages import Roles
9
  import gradio as gr
10
  from huggingface_hub import hf_hub_download
11
 
 
12
  hf_hub_download(
13
  repo_id="CerebrumTech/cere-gemma-2-9b-tr",
14
  filename="unsloth.Q4_K_M.gguf",
15
  local_dir="./models"
16
  )
17
 
18
- # 推論関数
19
  @spaces.GPU(duration=120)
20
  def respond(
21
  message,
22
  history: list[tuple[str, str]],
23
- model,
24
  system_message,
 
25
  max_tokens,
26
  temperature,
27
  top_p,
28
  top_k,
29
- repeat_penalty,
30
  ):
31
  chat_template = MessagesFormatterType.GEMMA_2
32
 
@@ -41,7 +42,7 @@ def respond(
41
 
42
  agent = LlamaCppAgent(
43
  provider,
44
- system_prompt=f"{system_message}",
45
  predefined_messages_formatter_type=chat_template,
46
  debug_output=True
47
  )
@@ -51,19 +52,19 @@ def respond(
51
  settings.top_k = top_k
52
  settings.top_p = top_p
53
  settings.max_tokens = max_tokens
54
- settings.repeat_penalty = repeat_penalty
55
  settings.stream = True
56
 
57
  messages = BasicChatHistory()
58
 
59
- for msn in history:
60
  user = {
61
  'role': Roles.user,
62
- 'content': msn[0]
63
  }
64
  assistant = {
65
  'role': Roles.assistant,
66
- 'content': msn[1]
67
  }
68
  messages.add_message(user)
69
  messages.add_message(assistant)
@@ -81,32 +82,27 @@ def respond(
81
  outputs += output
82
  yield outputs
83
 
84
-
85
  def create_interface(model_name, description):
86
- return gr.Interface(
87
- fn=respond,
88
- inputs=[
89
  gr.Textbox(value="You are a helpful assistant.", label="System message"),
90
  gr.Textbox(value=model_name, label="Model", interactive=False),
91
  gr.Slider(minimum=1, maximum=4096, value=2048, step=1, label="Max tokens"),
92
  gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
93
  gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p"),
94
  gr.Slider(minimum=0, maximum=100, value=40, step=1, label="Top-k"),
95
- gr.Slider(minimum=0.0, maximum=2.0, value=1.1, step=0.1, label="Repetition penalty"),
96
  ],
97
- outputs=gr.Chatbot(show_copy_button=True),
98
  title=model_name,
99
  description=description,
100
- live=True
101
  )
102
 
103
- description = """<p align="center"cCerebrumTech/cere-gemma-2-9b-tr/p>"""
 
104
  interface = create_interface('unsloth.Q4_K_M.gguf', description)
105
 
106
- demo = gr.Blocks()
107
-
108
- with demo:
109
- interface.render()
110
-
111
  if __name__ == "__main__":
112
- demo.launch()
 
9
  import gradio as gr
10
  from huggingface_hub import hf_hub_download
11
 
12
+ # Modeli indirme
13
  hf_hub_download(
14
  repo_id="CerebrumTech/cere-gemma-2-9b-tr",
15
  filename="unsloth.Q4_K_M.gguf",
16
  local_dir="./models"
17
  )
18
 
19
+ # Yanıt üretme fonksiyonu
20
  @spaces.GPU(duration=120)
21
  def respond(
22
  message,
23
  history: list[tuple[str, str]],
 
24
  system_message,
25
+ model,
26
  max_tokens,
27
  temperature,
28
  top_p,
29
  top_k,
30
+ repetition_penalty,
31
  ):
32
  chat_template = MessagesFormatterType.GEMMA_2
33
 
 
42
 
43
  agent = LlamaCppAgent(
44
  provider,
45
+ system_prompt=system_message,
46
  predefined_messages_formatter_type=chat_template,
47
  debug_output=True
48
  )
 
52
  settings.top_k = top_k
53
  settings.top_p = top_p
54
  settings.max_tokens = max_tokens
55
+ settings.repeat_penalty = repetition_penalty
56
  settings.stream = True
57
 
58
  messages = BasicChatHistory()
59
 
60
+ for user_msg, assistant_msg in history:
61
  user = {
62
  'role': Roles.user,
63
+ 'content': user_msg
64
  }
65
  assistant = {
66
  'role': Roles.assistant,
67
+ 'content': assistant_msg
68
  }
69
  messages.add_message(user)
70
  messages.add_message(assistant)
 
82
  outputs += output
83
  yield outputs
84
 
85
+ # Arayüz oluşturma fonksiyonu
86
  def create_interface(model_name, description):
87
+ return gr.ChatInterface(
88
+ fn=respond,
89
+ additional_inputs=[
90
  gr.Textbox(value="You are a helpful assistant.", label="System message"),
91
  gr.Textbox(value=model_name, label="Model", interactive=False),
92
  gr.Slider(minimum=1, maximum=4096, value=2048, step=1, label="Max tokens"),
93
  gr.Slider(minimum=0.1, maximum=4.0, value=0.7, step=0.1, label="Temperature"),
94
  gr.Slider(minimum=0.1, maximum=1.0, value=0.95, step=0.05, label="Top-p"),
95
  gr.Slider(minimum=0, maximum=100, value=40, step=1, label="Top-k"),
96
+ gr.Slider(minimum=0.0, maximum 2.0, value=1.1, step=0.1, label="Repetition penalty"),
97
  ],
 
98
  title=model_name,
99
  description=description,
 
100
  )
101
 
102
+ # Açıklama ve arayüz oluşturma
103
+ description = """<p align="center">CerebrumTech/cere-gemma-2-9b-tr</p>"""
104
  interface = create_interface('unsloth.Q4_K_M.gguf', description)
105
 
106
+ # Gradio uygulamasını başlatma
 
 
 
 
107
  if __name__ == "__main__":
108
+ interface.launch()