Spaces:
Sleeping
Sleeping
Updates of defaults and GPU resource change
Browse files
app.py
CHANGED
@@ -8,8 +8,10 @@ import torch
|
|
8 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
9 |
|
10 |
MAX_MAX_NEW_TOKENS = 2048
|
11 |
-
DEFAULT_MAX_NEW_TOKENS =
|
12 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
|
|
|
|
13 |
|
14 |
DESCRIPTION = """\
|
15 |
# Test Chat Information System for MEPO Summer Bridge 2024 courtesy of Dr. Dancy & THiCC Lab
|
@@ -25,6 +27,8 @@ As a derivate work of [Llama-2-7b-chat](https://huggingface.co/meta-llama/Llama-
|
|
25 |
this demo is governed by the original [license](https://huggingface.co/spaces/huggingface-projects/llama-2-7b-chat/blob/main/LICENSE.txt) and [acceptable use policy](https://huggingface.co/spaces/huggingface-projects/llama-2-7b-chat/blob/main/USE_POLICY.md).
|
26 |
"""
|
27 |
|
|
|
|
|
28 |
if not torch.cuda.is_available():
|
29 |
DESCRIPTION += "We won't be able to run this space! We need GPU processing"
|
30 |
|
@@ -36,12 +40,12 @@ if torch.cuda.is_available():
|
|
36 |
tokenizer.use_default_system_prompt = False
|
37 |
|
38 |
|
39 |
-
@spaces.GPU
|
40 |
def generate(
|
41 |
message: str,
|
42 |
chat_history: list[tuple[str, str]],
|
43 |
system_prompt: str,
|
44 |
-
max_new_tokens: int =
|
45 |
temperature: float = 0.6,
|
46 |
top_p: float = 0.9,
|
47 |
top_k: int = 50,
|
|
|
8 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
9 |
|
10 |
MAX_MAX_NEW_TOKENS = 2048
|
11 |
+
DEFAULT_MAX_NEW_TOKENS = 512
|
12 |
MAX_INPUT_TOKEN_LENGTH = int(os.getenv("MAX_INPUT_TOKEN_LENGTH", "4096"))
|
13 |
+
DEFAULT_SYS_PROMPT = """\
|
14 |
+
"""
|
15 |
|
16 |
DESCRIPTION = """\
|
17 |
# Test Chat Information System for MEPO Summer Bridge 2024 courtesy of Dr. Dancy & THiCC Lab
|
|
|
27 |
this demo is governed by the original [license](https://huggingface.co/spaces/huggingface-projects/llama-2-7b-chat/blob/main/LICENSE.txt) and [acceptable use policy](https://huggingface.co/spaces/huggingface-projects/llama-2-7b-chat/blob/main/USE_POLICY.md).
|
28 |
"""
|
29 |
|
30 |
+
default
|
31 |
+
|
32 |
if not torch.cuda.is_available():
|
33 |
DESCRIPTION += "We won't be able to run this space! We need GPU processing"
|
34 |
|
|
|
40 |
tokenizer.use_default_system_prompt = False
|
41 |
|
42 |
|
43 |
+
@spaces.GPU(duration=120)
|
44 |
def generate(
|
45 |
message: str,
|
46 |
chat_history: list[tuple[str, str]],
|
47 |
system_prompt: str,
|
48 |
+
max_new_tokens: int = MAX_MAX_NEW_TOKENS,
|
49 |
temperature: float = 0.6,
|
50 |
top_p: float = 0.9,
|
51 |
top_k: int = 50,
|