File size: 2,989 Bytes
5971329
88142f2
5971329
 
b125ffc
 
 
5971329
b125ffc
5971329
136ac77
5971329
 
b125ffc
5971329
 
136ac77
 
132b5af
cdadc4f
 
 
2b8f874
 
 
5971329
cdadc4f
 
 
 
 
 
 
 
 
 
 
 
 
136ac77
cdadc4f
 
136ac77
cdadc4f
 
 
 
 
136ac77
cdadc4f
 
136ac77
c99ac98
b207167
136ac77
 
2b8f874
136ac77
 
 
c430995
b03b518
 
136ac77
 
99b7bd9
 
 
 
 
 
 
 
136ac77
 
 
 
 
fdc7694
136ac77
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
import gradio as gr
import os
import cv2
import numpy as np
import imutils
from keras.preprocessing.image import img_to_array
from keras.models import load_model

# Load the pre-trained models and define parameters
detection_model_path = 'haarcascade_files/haarcascade_frontalface_default.xml'
emotion_model_path = 'model4_0.83/model4_entire_model.h5'
face_detection = cv2.CascadeClassifier(detection_model_path)
emotion_classifier = load_model(emotion_model_path, compile=False)
EMOTIONS = ['neutral', 'happiness', 'surprise', 'sadness', 'anger', 'disgust', 'fear', 'contempt', 'unknown']


# face_detector_mtcnn = MTCNN()
classifier = load_model(emotion_model_path)

def predict_emotion(frame):
    frame = imutils.resize(frame, width=300)
    gray = cv2.cvtColor(frame, cv2.COLOR_RGB2GRAY)
    faces = face_detection.detectMultiScale(gray, scaleFactor=1.1,
                                            minNeighbors=5, minSize=(30, 30),
                                            flags=cv2.CASCADE_SCALE_IMAGE)

    frame_clone = frame.copy()
    if len(faces) > 0:
        faces = sorted(faces, reverse=True,
                       key=lambda x: (x[2] - x[0]) * (x[3] - x[1]))[0]
        (fX, fY, fW, fH) = faces
        
        # Extract the ROI of the face from the grayscale image, resize it to a fixed 28x28 pixels, and then prepare
        # the ROI for classification via the CNN
        roi = gray[fY:fY + fH, fX:fX + fW]
        roi = cv2.resize(roi, (48, 48))
        roi = roi.astype("float") / 255.0
        roi = img_to_array(roi)
        roi = np.expand_dims(roi, axis=0)

        preds = emotion_classifier.predict(roi)[0]
        label = EMOTIONS[preds.argmax()]

        # Overlay a box over the detected face
        cv2.putText(frame_clone, label, (fX, fY - 10),
                    cv2.FONT_HERSHEY_DUPLEX, 1, (238, 164, 64), 1)
        cv2.rectangle(frame_clone, (fX, fY), (fX + fW, fY + fH),
                      (238, 164, 64), 2)

    else:
        label = "Can't find your face"

    return frame_clone




demo = gr.Interface(
    fn = predict_emotion,
    inputs = gr.Image(type="numpy"),
    outputs = gr.Image(),
    # gr.components.Image(label="Predicted Emotion"),
    # gr.components.Label(num_top_classes=2, label="Top 2 Probabilities")
    #flagging_options=["blurry", "incorrect", "other"],
    examples = [
        
        os.path.join(os.path.dirname(__file__), "images/chandler.jpeg"),
        os.path.join(os.path.dirname(__file__), "images/janice.jpeg"),
        os.path.join(os.path.dirname(__file__), "images/joey.jpeg"),
        os.path.join(os.path.dirname(__file__), "images/phoebe.jpeg"),
        os.path.join(os.path.dirname(__file__), "images/rachel_monica.jpeg"),
        os.path.join(os.path.dirname(__file__), "images/ross.jpeg"),
        os.path.join(os.path.dirname(__file__), "images/gunther.jpeg")
     
    ],
    title = "Whatchu feeling?",
    theme = "shivi/calm_seafoam"
)
    


if __name__ == "__main__":
    demo.launch()