halme commited on
Commit
acbac18
·
1 Parent(s): bcb8b37
Files changed (1) hide show
  1. app.py +2 -2
app.py CHANGED
@@ -1,6 +1,6 @@
1
  import gradio as gr
2
  from huggingface_hub import InferenceClient
3
- from unsloth import FastLanguageModel
4
  from peft import AutoPeftModelForCausalLM
5
  from transformers import AutoTokenizer
6
 
@@ -44,7 +44,7 @@ def respond(message, history: list[tuple[str, str]], system_message, max_tokens,
44
  )
45
  tokenizer = AutoTokenizer.from_pretrained("halme/id2223_lora_model")
46
 
47
- FastLanguageModel.for_inference(model) # Enable native 2x faster inference
48
 
49
  """messages = [
50
  {"role": "user", "content": "Continue the fibonnaci sequence: 1, 1, 2, 3, 5, 8,"},
 
1
  import gradio as gr
2
  from huggingface_hub import InferenceClient
3
+ #from unsloth import FastLanguageModel
4
  from peft import AutoPeftModelForCausalLM
5
  from transformers import AutoTokenizer
6
 
 
44
  )
45
  tokenizer = AutoTokenizer.from_pretrained("halme/id2223_lora_model")
46
 
47
+ #FastLanguageModel.for_inference(model) # Enable native 2x faster inference
48
 
49
  """messages = [
50
  {"role": "user", "content": "Continue the fibonnaci sequence: 1, 1, 2, 3, 5, 8,"},