John Ho commited on
Commit
4e1e198
·
1 Parent(s): a83f12f

try quantization again

Browse files
Files changed (2) hide show
  1. app.py +5 -5
  2. pyproject.toml +1 -0
app.py CHANGED
@@ -110,11 +110,11 @@ def inference(
110
  apply_quantization: bool = True,
111
  ):
112
  # default processor
113
- processor, model = PROCESSOR, MODEL
114
- # processor = load_processor()
115
- # model = load_model(
116
- # use_flash_attention=use_flash_attention, apply_quantization=apply_quantization
117
- # )
118
 
119
  # The model is trained on 8.0 FPS which we recommend for optimal inference
120
  fps = get_fps_ffmpeg(video_path)
 
110
  apply_quantization: bool = True,
111
  ):
112
  # default processor
113
+ # processor, model = PROCESSOR, MODEL
114
+ processor = load_processor()
115
+ model = load_model(
116
+ use_flash_attention=use_flash_attention, apply_quantization=apply_quantization
117
+ )
118
 
119
  # The model is trained on 8.0 FPS which we recommend for optimal inference
120
  fps = get_fps_ffmpeg(video_path)
pyproject.toml CHANGED
@@ -14,4 +14,5 @@ dependencies = [
14
  "ffmpeg-python>=0.2.0",
15
  "accelerate==0.32.1",
16
  "bitsandbytes==0.41.1",
 
17
  ]
 
14
  "ffmpeg-python>=0.2.0",
15
  "accelerate==0.32.1",
16
  "bitsandbytes==0.41.1",
17
+ "scipy==1.11.3",
18
  ]