Update README.md
Browse files
README.md
CHANGED
@@ -27,15 +27,12 @@ This model is an advanced fp8 quantized version of google/gemma-3-27b-it, meticu
|
|
27 |
from transformers import AutoProcessor, Gemma3ForConditionalGeneration, BitsAndBytesConfig
|
28 |
from PIL import Image
|
29 |
import torch
|
30 |
-
|
31 |
-
# Define configuration for 8-bit quantization
|
32 |
-
quantization_config = BitsAndBytesConfig(load_in_8bit=True)
|
33 |
-
|
34 |
model_id = "MISHANM/google-gemma-3-27b-it-fp8"
|
35 |
|
36 |
# Load the model with 8-bit quantization
|
37 |
model = Gemma3ForConditionalGeneration.from_pretrained(
|
38 |
-
model_id, device_map="auto"
|
39 |
).eval()
|
40 |
|
41 |
processor = AutoProcessor.from_pretrained(model_id)
|
|
|
27 |
from transformers import AutoProcessor, Gemma3ForConditionalGeneration, BitsAndBytesConfig
|
28 |
from PIL import Image
|
29 |
import torch
|
30 |
+
|
|
|
|
|
|
|
31 |
model_id = "MISHANM/google-gemma-3-27b-it-fp8"
|
32 |
|
33 |
# Load the model with 8-bit quantization
|
34 |
model = Gemma3ForConditionalGeneration.from_pretrained(
|
35 |
+
model_id, device_map="auto"
|
36 |
).eval()
|
37 |
|
38 |
processor = AutoProcessor.from_pretrained(model_id)
|