MatteoScript commited on
Commit
eb8cfbd
·
1 Parent(s): 283fd24

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +11 -24
app.py CHANGED
@@ -3,40 +3,27 @@ import gradio as gr
3
 
4
  client = InferenceClient("mistralai/Mixtral-8x7B-Instruct-v0.1")
5
 
6
- def format_prompt(message, history):
7
- prompt = "<s>"
8
- for user_prompt, bot_response in history:
9
- prompt += f"[INST] {user_prompt} [/INST]"
10
- prompt += f" {bot_response}</s> "
11
- prompt += f"[INST] {message} [/INST]"
12
- return prompt
13
-
14
- def generate(
15
- prompt, history, temperature=0.2, max_new_tokens=30000, top_p=0.9, repetition_penalty=1.0,
16
- ):
17
  temperature = max(float(temperature), 0.01)
18
  top_p = max(min(float(top_p), 1.0), 0.0)
19
  repetition_penalty = max(float(repetition_penalty), 0.01)
20
 
21
- generate_kwargs = dict(
22
- temperature=temperature,
23
- max_new_tokens=max_new_tokens,
24
- top_p=top_p,
25
- repetition_penalty=repetition_penalty,
26
- do_sample=True,
27
- seed=42,
28
- )
29
-
30
- formatted_prompt = format_prompt(prompt, history)
31
 
32
- # Generate text
33
- response = client.text_generation(formatted_prompt, **generate_kwargs)
34
  generated_text = response["generated_text"]
35
  return generated_text
36
 
37
  iface = gr.Interface(
38
  fn=generate,
39
- inputs=["text", "text", gr.inputs.Slider(0.1, 2.0), gr.inputs.Slider(100, 50000), gr.inputs.Slider(0.1, 1.0)],
40
  outputs="text",
41
  title="Text Generation"
42
  )
 
3
 
4
  client = InferenceClient("mistralai/Mixtral-8x7B-Instruct-v0.1")
5
 
6
+ def generate(prompt, temperature=0.2, max_new_tokens=30000, top_p=0.9, repetition_penalty=1.0):
 
 
 
 
 
 
 
 
 
 
7
  temperature = max(float(temperature), 0.01)
8
  top_p = max(min(float(top_p), 1.0), 0.0)
9
  repetition_penalty = max(float(repetition_penalty), 0.01)
10
 
11
+ generate_kwargs = {
12
+ "temperature": temperature,
13
+ "max_new_tokens": max_new_tokens,
14
+ "top_p": top_p,
15
+ "repetition_penalty": repetition_penalty,
16
+ "do_sample": True,
17
+ "seed": 42,
18
+ }
 
 
19
 
20
+ response = client.text_generation(prompt, **generate_kwargs)
 
21
  generated_text = response["generated_text"]
22
  return generated_text
23
 
24
  iface = gr.Interface(
25
  fn=generate,
26
+ inputs=["text", gr.inputs.Slider(0.1, 2.0), gr.inputs.Slider(100, 50000), gr.inputs.Slider(0.1, 1.0)],
27
  outputs="text",
28
  title="Text Generation"
29
  )