drlon commited on
Commit
6a2dd2b
·
1 Parent(s): c44028f

root logger

Browse files
Files changed (2) hide show
  1. app.py +9 -8
  2. util/utils.py +4 -1
app.py CHANGED
@@ -1,13 +1,5 @@
1
  import traceback
2
  import logging
3
- logger = logging.getLogger()
4
- logger.setLevel(logging.WARNING)
5
- if not logger.handlers:
6
- handler = logging.StreamHandler()
7
- handler.setFormatter(logging.Formatter("%(asctime)s [%(levelname)s] %(name)s: %(message)s"))
8
- logger.addHandler(handler)
9
- logger.warning("here")
10
-
11
  from typing import Optional
12
  import spaces
13
  import gradio as gr
@@ -30,6 +22,14 @@ import torch
30
  from transformers import AutoModelForCausalLM
31
  from transformers import AutoProcessor
32
 
 
 
 
 
 
 
 
 
33
  # Define repository and local directory
34
  repo_id = "microsoft/OmniParser-v2.0" # HF repo
35
  local_dir = "weights" # Target local directory
@@ -105,6 +105,7 @@ def get_som_response(instruction, image_som):
105
  inputs = magma_processor(images=[image_som], texts=prompt, return_tensors="pt")
106
  inputs['pixel_values'] = inputs['pixel_values'].unsqueeze(0).to(torch.bfloat16) # Add .to(torch.bfloat16) here for explicit casting
107
  inputs['image_sizes'] = inputs['image_sizes'].unsqueeze(0)
 
108
  # inputs = inputs.to("cuda")
109
  inputs = inputs.to("cuda", dtype=torch.bfloat16)
110
 
 
1
  import traceback
2
  import logging
 
 
 
 
 
 
 
 
3
  from typing import Optional
4
  import spaces
5
  import gradio as gr
 
22
  from transformers import AutoModelForCausalLM
23
  from transformers import AutoProcessor
24
 
25
+ logger = logging.getLogger()
26
+ logger.setLevel(logging.WARNING)
27
+ if not logger.handlers:
28
+ handler = logging.StreamHandler()
29
+ handler.setFormatter(logging.Formatter("%(asctime)s [%(levelname)s] %(name)s: %(message)s"))
30
+ logger.addHandler(handler)
31
+ logger.warning("here")
32
+
33
  # Define repository and local directory
34
  repo_id = "microsoft/OmniParser-v2.0" # HF repo
35
  local_dir = "weights" # Target local directory
 
105
  inputs = magma_processor(images=[image_som], texts=prompt, return_tensors="pt")
106
  inputs['pixel_values'] = inputs['pixel_values'].unsqueeze(0).to(torch.bfloat16) # Add .to(torch.bfloat16) here for explicit casting
107
  inputs['image_sizes'] = inputs['image_sizes'].unsqueeze(0)
108
+ logger.warning(inputs['pixel_values'].dtype)
109
  # inputs = inputs.to("cuda")
110
  inputs = inputs.to("cuda", dtype=torch.bfloat16)
111
 
util/utils.py CHANGED
@@ -18,7 +18,8 @@ import numpy as np
18
  from matplotlib import pyplot as plt
19
  import easyocr
20
  from paddleocr import PaddleOCR
21
- reader = easyocr.Reader(['en'])
 
22
  paddle_ocr = PaddleOCR(
23
  lang='en', # other lang also available
24
  use_angle_cls=False,
@@ -520,6 +521,8 @@ def check_ocr_box(image_source: Union[str, Image.Image], display_img = True, out
520
  else: # EasyOCR
521
  if easyocr_args is None:
522
  easyocr_args = {}
 
 
523
  result = reader.readtext(image_np, **easyocr_args)
524
  coord = [item[0] for item in result]
525
  text = [item[1] for item in result]
 
18
  from matplotlib import pyplot as plt
19
  import easyocr
20
  from paddleocr import PaddleOCR
21
+ # reader = easyocr.Reader(['en'])
22
+ reader = None
23
  paddle_ocr = PaddleOCR(
24
  lang='en', # other lang also available
25
  use_angle_cls=False,
 
521
  else: # EasyOCR
522
  if easyocr_args is None:
523
  easyocr_args = {}
524
+ if reader is None:
525
+ reader = easyocr.Reader(['en'])
526
  result = reader.readtext(image_np, **easyocr_args)
527
  coord = [item[0] for item in result]
528
  text = [item[1] for item in result]