English
yintongl commited on
Commit
5a5cf71
·
verified ·
1 Parent(s): 809c657

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -15
README.md CHANGED
@@ -9,24 +9,10 @@ language:
9
  ## Model Details
10
 
11
  This model is an int4 model with group_size 128 of [facebook/opt-13b](https://huggingface.co/facebook/opt-13b) generated by [intel/auto-round](https://github.com/intel/auto-round).
 
12
 
13
 
14
 
15
- ### INT4 Inference with AutoGPTQ's Kernel
16
-
17
- ```python
18
- ##pip install auto-gptq[triton]
19
- ##pip install triton==2.2.0
20
- from transformers import AutoModelForCausalLM, AutoTokenizer
21
- quantized_model_dir = "Intel/opt-13b-int4-inc"
22
- model = AutoModelForCausalLM.from_pretrained(quantized_model_dir,
23
- device_map="auto",
24
- trust_remote_code=False,
25
- )
26
- tokenizer = AutoTokenizer.from_pretrained(quantized_model_dir, use_fast=True)
27
- print(tokenizer.decode(model.generate(**tokenizer("There is a girl who likes adventure,", return_tensors="pt").to(model.device),max_new_tokens=50)[0]))
28
- ```
29
-
30
 
31
 
32
  ### Evaluate the model
 
9
  ## Model Details
10
 
11
  This model is an int4 model with group_size 128 of [facebook/opt-13b](https://huggingface.co/facebook/opt-13b) generated by [intel/auto-round](https://github.com/intel/auto-round).
12
+ Inference of this model is compatible with AutoGPTQ's Kernel.
13
 
14
 
15
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
16
 
17
 
18
  ### Evaluate the model