Spaces:
Sleeping
Sleeping
test
Browse files
app.py
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
import gradio as gr
|
2 |
from huggingface_hub import InferenceClient
|
3 |
-
from unsloth import FastLanguageModel
|
4 |
from peft import AutoPeftModelForCausalLM
|
5 |
from transformers import AutoTokenizer
|
6 |
|
@@ -44,7 +44,7 @@ def respond(message, history: list[tuple[str, str]], system_message, max_tokens,
|
|
44 |
)
|
45 |
tokenizer = AutoTokenizer.from_pretrained("halme/id2223_lora_model")
|
46 |
|
47 |
-
FastLanguageModel.for_inference(model) # Enable native 2x faster inference
|
48 |
|
49 |
"""messages = [
|
50 |
{"role": "user", "content": "Continue the fibonnaci sequence: 1, 1, 2, 3, 5, 8,"},
|
|
|
1 |
import gradio as gr
|
2 |
from huggingface_hub import InferenceClient
|
3 |
+
#from unsloth import FastLanguageModel
|
4 |
from peft import AutoPeftModelForCausalLM
|
5 |
from transformers import AutoTokenizer
|
6 |
|
|
|
44 |
)
|
45 |
tokenizer = AutoTokenizer.from_pretrained("halme/id2223_lora_model")
|
46 |
|
47 |
+
#FastLanguageModel.for_inference(model) # Enable native 2x faster inference
|
48 |
|
49 |
"""messages = [
|
50 |
{"role": "user", "content": "Continue the fibonnaci sequence: 1, 1, 2, 3, 5, 8,"},
|