SPBhai commited on
Commit
6d76007
·
verified ·
1 Parent(s): 909c589

Loading the model without quantization

Browse files
Files changed (1) hide show
  1. app.py +2 -3
app.py CHANGED
@@ -31,11 +31,10 @@ similarity_retriever = vector_store.as_retriever(
31
  )
32
 
33
  # Load the LLM
34
- quantization_config = BitsAndBytesConfig(load_in_8bit=True, device = "cpu")
35
  tokenizer = AutoTokenizer.from_pretrained("google/gemma-2-9b-it")
36
  llm_model = AutoModelForCausalLM.from_pretrained(
37
- "google/gemma-2-9b-it",
38
- quantization_config=quantization_config,
39
  )
40
  text_generation_pipeline = pipeline(
41
  model=llm_model,
 
31
  )
32
 
33
  # Load the LLM
34
+ # quantization_config = BitsAndBytesConfig(load_in_8bit=True, device = "cpu")
35
  tokenizer = AutoTokenizer.from_pretrained("google/gemma-2-9b-it")
36
  llm_model = AutoModelForCausalLM.from_pretrained(
37
+ "google/gemma-2-9b-it"
 
38
  )
39
  text_generation_pipeline = pipeline(
40
  model=llm_model,