osmankoc commited on
Commit
06a7d62
·
1 Parent(s): 0c527d7
Files changed (1) hide show
  1. app.py +25 -6
app.py CHANGED
@@ -1,8 +1,27 @@
 
 
 
1
  import torch
2
 
3
- print("CUDA mevcut mu?:", torch.cuda.is_available())
4
- print("GPU Sayısı:", torch.cuda.device_count())
5
- if torch.cuda.is_available():
6
- print("GPU Adı:", torch.cuda.get_device_name(0))
7
- else:
8
- print("CUDA GPU ALGILANMADI! Hugging Face Spaces'in GPU kullandığını kontrol et.")
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import spaces
2
+ import gradio as gr
3
+ from transformers import AutoModelForCausalLM, AutoTokenizer
4
  import torch
5
 
6
+ MODEL_NAME = "osmankoc/llama-2-7b-zoa"
7
+
8
+ # ZeroGPU için model GPU'ya sadece gerektiğinde yüklenecek
9
+ @spaces.GPU
10
+ def generate(prompt):
11
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
12
+ model = AutoModelForCausalLM.from_pretrained(
13
+ MODEL_NAME, torch_dtype=torch.float16, device_map="auto"
14
+ )
15
+ inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
16
+ output = model.generate(**inputs, max_length=500)
17
+ response = tokenizer.decode(output[0], skip_special_tokens=True)
18
+ return response
19
+
20
+ # Gradio UI (Basit bir API arayüzü gibi çalışacak)
21
+ demo = gr.Interface(
22
+ fn=generate,
23
+ inputs=gr.Textbox(placeholder="Enter prompt..."),
24
+ outputs=gr.Textbox(),
25
+ )
26
+
27
+ demo.launch()