MISHANM commited on
Commit
f9daeca
·
verified ·
1 Parent(s): f0a4fdd

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -5
README.md CHANGED
@@ -27,15 +27,12 @@ This model is an advanced fp8 quantized version of google/gemma-3-27b-it, meticu
27
  from transformers import AutoProcessor, Gemma3ForConditionalGeneration, BitsAndBytesConfig
28
  from PIL import Image
29
  import torch
30
-
31
- # Define configuration for 8-bit quantization
32
- quantization_config = BitsAndBytesConfig(load_in_8bit=True)
33
-
34
  model_id = "MISHANM/google-gemma-3-27b-it-fp8"
35
 
36
  # Load the model with 8-bit quantization
37
  model = Gemma3ForConditionalGeneration.from_pretrained(
38
- model_id, device_map="auto", quantization_config=quantization_config
39
  ).eval()
40
 
41
  processor = AutoProcessor.from_pretrained(model_id)
 
27
  from transformers import AutoProcessor, Gemma3ForConditionalGeneration, BitsAndBytesConfig
28
  from PIL import Image
29
  import torch
30
+
 
 
 
31
  model_id = "MISHANM/google-gemma-3-27b-it-fp8"
32
 
33
  # Load the model with 8-bit quantization
34
  model = Gemma3ForConditionalGeneration.from_pretrained(
35
+ model_id, device_map="auto"
36
  ).eval()
37
 
38
  processor = AutoProcessor.from_pretrained(model_id)