Jabrain commited on
Commit
0f72fdd
·
1 Parent(s): 2ec0cea

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +37 -37
app.py CHANGED
@@ -4,7 +4,7 @@ import gradio as gr
4
  import torch
5
  import transformers
6
  import librosa
7
- import cv2
8
  import numpy as np
9
 
10
  # Load models
@@ -63,46 +63,46 @@ def audio_to_video(input):
63
  output = gr.outputs.Video.from_str(output)
64
  return output
65
 
66
- def image_to_text(input):
67
- input = cv2.imread(input)
68
- input = cv2.cvtColor(input, cv2.COLOR_BGR2RGB)
69
- input = np.expand_dims(input, axis=0)
70
- output = image_model(input)
71
- return output[0]["label"]
72
-
73
- def image_to_audio(input):
74
- input = cv2.imread(input)
75
- input = cv2.cvtColor(input, cv2.COLOR_BGR2RGB)
76
- input = np.expand_dims(input, axis=0)
77
- output = image_model(input)
78
- output = gr.outputs.Audio.from_str(output[0]["label"])
79
- return output
80
 
81
  def image_to_image(input):
82
  return input
83
 
84
- def image_to_video(input):
85
- input = cv2.imread(input)
86
- input = cv2.cvtColor(input, cv2.COLOR_BGR2RGB)
87
- input = np.expand_dims(input, axis=0)
88
- output = image_model(input)
89
- output = gr.outputs.Video.from_str(output[0]["label"])
90
- return output
91
-
92
- def video_to_text(input):
93
- input = cv2.VideoCapture(input)
94
- frames = []
95
- while input.isOpened():
96
- ret, frame = input.read()
97
- if ret:
98
- frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB)
99
- frames.append(frame)
100
- else:
101
- breakpoint
102
- input.release()
103
- frames = np.array(frames)
104
- output = video_model(frames)
105
- return output[0]["label"]
106
 
107
  #Output switches
108
 
 
4
  import torch
5
  import transformers
6
  import librosa
7
+ #import cv2
8
  import numpy as np
9
 
10
  # Load models
 
63
  output = gr.outputs.Video.from_str(output)
64
  return output
65
 
66
+ #def image_to_text(input):
67
+ # input = cv2.imread(input)
68
+ # input = cv2.cvtColor(input, cv2.COLOR_BGR2RGB)
69
+ # input = np.expand_dims(input, axis=0)
70
+ # output = image_model(input)
71
+ # return output[0]["label"]
72
+
73
+ #def image_to_audio(input):
74
+ # input = cv2.imread(input)
75
+ # input = cv2.cvtColor(input, cv2.COLOR_BGR2RGB)
76
+ # input = np.expand_dims(input, axis=0)
77
+ # output = image_model(input)
78
+ # output = gr.outputs.Audio.from_str(output[0]["label"])
79
+ # return output
80
 
81
  def image_to_image(input):
82
  return input
83
 
84
+ #def image_to_video(input):
85
+ # input = cv2.imread(input)
86
+ # input = cv2.cvtColor(input, cv2.COLOR_BGR2RGB)
87
+ # input = np.expand_dims(input, axis=0)
88
+ # output = image_model(input)
89
+ # output = gr.outputs.Video.from_str(output[0]["label"])
90
+ # return output
91
+
92
+ #def video_to_text(input):
93
+ # input = cv2.VideoCapture(input)
94
+ # frames = []
95
+ # while input.isOpened():
96
+ # ret, frame = input.read()
97
+ # if ret:
98
+ # frame = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB)
99
+ # frames.append(frame)
100
+ # else:
101
+ # breakpoint
102
+ # input.release()
103
+ # frames = np.array(frames)
104
+ # output = video_model(frames)
105
+ # return output[0]["label"]
106
 
107
  #Output switches
108