Manju017 commited on
Commit
98c19b4
·
verified ·
1 Parent(s): 0bbbd1f

Updated to load the model in 8-bit precision to reduce memory usage

Browse files
Files changed (1) hide show
  1. app.py +7 -1
app.py CHANGED
@@ -2,7 +2,13 @@ import gradio as gr
2
  from transformers import AutoModelForCausalLM, AutoTokenizer
3
 
4
  model_name = "ai4bharat/Airavata"
5
- model = AutoModelForCausalLM.from_pretrained(model_name)
 
 
 
 
 
 
6
  tokenizer = AutoTokenizer.from_pretrained(model_name)
7
 
8
  def generate_text(prompt, max_length):
 
2
  from transformers import AutoModelForCausalLM, AutoTokenizer
3
 
4
  model_name = "ai4bharat/Airavata"
5
+
6
+ # Load the model in 8-bit precision to reduce memory usage
7
+ model = AutoModelForCausalLM.from_pretrained(
8
+ model_name,
9
+ device_map="auto",
10
+ load_in_8bit=True
11
+ )
12
  tokenizer = AutoTokenizer.from_pretrained(model_name)
13
 
14
  def generate_text(prompt, max_length):