Spaces:
vericava
/
Running on Zero

metastable-void commited on
Commit
45c9a85
·
1 Parent(s): 4a30925
Files changed (1) hide show
  1. app.py +14 -14
app.py CHANGED
@@ -10,20 +10,20 @@ import torch
10
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer, pipeline
11
  from peft import PeftModel
12
 
13
- DESCRIPTION = "# 真空ジェネレータ\n<p>Imitate 真空 (@vericava)'s posts interactively</p>"
14
 
15
  if not torch.cuda.is_available():
16
  DESCRIPTION += "\n<p>Running on CPU 🥶 This demo does not work on CPU.</p>"
17
 
18
- MAX_MAX_NEW_TOKENS = 768
19
- DEFAULT_MAX_NEW_TOKENS = 512
20
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "32768"))
21
 
22
 
23
  if torch.cuda.is_available():
24
  my_pipeline=pipeline(
25
  task="text-generation",
26
- model="vericava/gpt2-medium-vericava-posts-v3",
27
  do_sample=True,
28
  num_beams=1,
29
  )
@@ -59,8 +59,8 @@ def generate(
59
  )[-1]["generated_text"]
60
  print(output)
61
  gen_text = output[len(user_input):]
62
- gen_text = gen_text[:gen_text.find("\n")] if "\n" in gen_text else gen_text
63
- gen_text = gen_text[:(gen_text.rfind("。") + 1)] if "。" in gen_text else gen_text
64
  yield gen_text
65
 
66
  demo = gr.ChatInterface(
@@ -87,29 +87,29 @@ demo = gr.ChatInterface(
87
  minimum=0.05,
88
  maximum=1.0,
89
  step=0.05,
90
- value=0.95,
91
  ),
92
  gr.Slider(
93
  label="Top-k",
94
  minimum=1,
95
- maximum=1000,
96
  step=1,
97
- value=50,
98
  ),
99
  gr.Slider(
100
  label="Repetition penalty",
101
  minimum=1.0,
102
- maximum=2.0,
103
  step=0.05,
104
- value=1.5,
105
  ),
106
  ],
107
  stop_btn=None,
108
  examples=[
109
- ["サマリーを作る男の人,サマリーマン。"],
110
- ["やばい場所にクリティカルな配線ができてしまったので掲示した。"],
111
  ["にゃん"],
112
- ["Wikipedia の情報は入っているのかもしれない"],
113
  ],
114
  description=DESCRIPTION,
115
  css_paths="style.css",
 
10
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer, pipeline
11
  from peft import PeftModel
12
 
13
+ DESCRIPTION = "# 真空ジェネレータ (v3)\n<p>Imitate 真空 (@vericava)'s posts interactively</p>"
14
 
15
  if not torch.cuda.is_available():
16
  DESCRIPTION += "\n<p>Running on CPU 🥶 This demo does not work on CPU.</p>"
17
 
18
+ MAX_MAX_NEW_TOKENS = 128
19
+ DEFAULT_MAX_NEW_TOKENS = 64
20
  MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "32768"))
21
 
22
 
23
  if torch.cuda.is_available():
24
  my_pipeline=pipeline(
25
  task="text-generation",
26
+ model="vericava/llm-jp-3-vericava-posts-v1",
27
  do_sample=True,
28
  num_beams=1,
29
  )
 
59
  )[-1]["generated_text"]
60
  print(output)
61
  gen_text = output[len(user_input):]
62
+ #gen_text = gen_text[:gen_text.find("\n")] if "\n" in gen_text else gen_text
63
+ #gen_text = gen_text[:(gen_text.rfind("。") + 1)] if "。" in gen_text else gen_text
64
  yield gen_text
65
 
66
  demo = gr.ChatInterface(
 
87
  minimum=0.05,
88
  maximum=1.0,
89
  step=0.05,
90
+ value=0.90,
91
  ),
92
  gr.Slider(
93
  label="Top-k",
94
  minimum=1,
95
+ maximum=100,
96
  step=1,
97
+ value=20,
98
  ),
99
  gr.Slider(
100
  label="Repetition penalty",
101
  minimum=1.0,
102
+ maximum=4.0,
103
  step=0.05,
104
+ value=2.0,
105
  ),
106
  ],
107
  stop_btn=None,
108
  examples=[
109
+ ["おはよ"],
110
+ ["えらいね"],
111
  ["にゃん"],
112
+ ["よしよし"],
113
  ],
114
  description=DESCRIPTION,
115
  css_paths="style.css",