|
import gradio as gr |
|
from fastai.vision.all import * |
|
from transformers import AutoImageProcessor, AutoModelForImageClassification |
|
from PIL import Image |
|
import requests |
|
|
|
learn_inf = load_learner("export.pkl") |
|
processor = AutoImageProcessor.from_pretrained("dima806/facial_emotions_image_detection") |
|
model = AutoModelForImageClassification.from_pretrained("dima806/facial_emotions_image_detection") |
|
|
|
|
|
def predict(value) -> str: |
|
image = Image.fromarray(value).convert("L").convert("RGB") |
|
inputs = processor(images=image, return_tensors="pt") |
|
outputs = model(**inputs) |
|
logits = outputs.logits |
|
predicted_class_idx = logits.argmax(-1).item() |
|
return model.config.id2label[predicted_class_idx] |
|
|
|
with gr.Blocks() as demo: |
|
with gr.Row(): |
|
with gr.Column(): |
|
input_img = gr.Image(label="Input", sources="webcam") |
|
with gr.Column(): |
|
output_lbl = gr.Label(value="Output", label="Expression Prediction") |
|
input_img.stream(fn=predict, inputs=input_img, outputs=output_lbl,concurrency_limit=20,time_limit=20,stream_every=0.1) |
|
|
|
if __name__ == "__main__": |
|
|
|
demo.launch() |