mdanish commited on
Commit
fd8d179
·
verified ·
1 Parent(s): 9cefacb

Upload app.py with huggingface_hub

Browse files
Files changed (1) hide show
  1. app.py +4 -4
app.py CHANGED
@@ -8,6 +8,7 @@ import open_clip
8
 
9
  knnpath = '20241204-ams-no-env-open_clip_ViT-H-14-378-quickgelu.npz'
10
  clip_model_name = 'ViT-H-14-378-quickgelu'
 
11
 
12
 
13
  # Set page config
@@ -22,16 +23,15 @@ st.set_page_config(
22
  #model, preprocess = open_clip.create_model_from_pretrained(clip_model_name)
23
  #tokenizer = open_clip.get_tokenizer(clip_model_name)
24
 
25
- st.write("Available models:", open_clip.list_models())
26
 
27
  @st.cache_resource
28
  def load_model():
29
  """Load the OpenCLIP model and return model and processor"""
30
  model, _, preprocess = open_clip.create_model_and_transforms(
31
- 'ViT-H-14-378-quickgelu',
32
- pretrained='laion2b_s32b_b79k'
33
  )
34
- tokenizer = open_clip.get_tokenizer('ViT-H-14')
35
  return model, preprocess, tokenizer
36
 
37
  def process_image(image, preprocess):
 
8
 
9
  knnpath = '20241204-ams-no-env-open_clip_ViT-H-14-378-quickgelu.npz'
10
  clip_model_name = 'ViT-H-14-378-quickgelu'
11
+ pretrained_name = 'dfn5b'
12
 
13
 
14
  # Set page config
 
23
  #model, preprocess = open_clip.create_model_from_pretrained(clip_model_name)
24
  #tokenizer = open_clip.get_tokenizer(clip_model_name)
25
 
26
+ #st.write("Available models:", open_clip.list_models())
27
 
28
  @st.cache_resource
29
  def load_model():
30
  """Load the OpenCLIP model and return model and processor"""
31
  model, _, preprocess = open_clip.create_model_and_transforms(
32
+ clip_model_name, pretrained=pretrained_name
 
33
  )
34
+ tokenizer = open_clip.get_tokenizer(clip_model_name)
35
  return model, preprocess, tokenizer
36
 
37
  def process_image(image, preprocess):