MakiAi commited on
Commit
c06b9fa
·
verified ·
1 Parent(s): 62780ba

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +24 -6
app.py CHANGED
@@ -1,12 +1,30 @@
1
-
2
  import gradio as gr
3
- import spaces
4
- import transformers_gradio
 
 
 
 
 
 
 
 
 
 
 
 
5
 
6
- #demo = gr.load(name="MakiAi/Llama-3.2-3B-Instruct-bnb-4bit-OKU_wiki_llama3.1_8b_inst_Reflexive_chunk200_overlap700-10epochs", src=transformers_gradio.registry)
7
- demo = gr.load(name="unsloth/Llama-3.2-1B", src=transformers_gradio.registry)
 
 
 
8
 
9
- demo.fn = spaces.GPU()(demo.fn)
 
 
 
 
10
 
11
  if __name__ == "__main__":
12
  demo.launch()
 
 
1
  import gradio as gr
2
+ from transformers import AutoModelForCausalLM, AutoTokenizer
3
+ import torch
4
+
5
+ # Base modelとadapterの設定
6
+ base_model_id = "unsloth/Llama-3.2-3B" # または適切なbase model
7
+ adapter_model_id = "MakiAi/Llama-3.2-3B-Instruct-bnb-4bit-OKU_wiki_llama3.1_8b_inst_Reflexive_chunk200_overlap700-10epochs"
8
+
9
+ # モデルとトークナイザーの読み込み
10
+ model = AutoModelForCausalLM.from_pretrained(
11
+ adapter_model_id,
12
+ torch_dtype=torch.float16,
13
+ device_map="auto"
14
+ )
15
+ tokenizer = AutoTokenizer.from_pretrained(adapter_model_id)
16
 
17
+ # Gradioインターフェースの作成
18
+ def generate_text(prompt):
19
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
20
+ outputs = model.generate(**inputs, max_new_tokens=100)
21
+ return tokenizer.decode(outputs[0], skip_special_tokens=True)
22
 
23
+ demo = gr.Interface(
24
+ fn=generate_text,
25
+ inputs="text",
26
+ outputs="text"
27
+ )
28
 
29
  if __name__ == "__main__":
30
  demo.launch()