import PIL.Image import PIL.ImageOps import gradio as gr import numpy as np import tensorflow as tf from poser import draw_bones, movenet def predict(image: PIL.Image): input_size = 256 size = (1280, 1280) image = PIL.ImageOps.fit(image, size, PIL.Image.LANCZOS) # image = PIL.ImageOps.contain(image, size) image_tf = tf.keras.preprocessing.image.img_to_array(image) # Resize and pad the image to keep the aspect ratio and fit the expected size. input_image = tf.expand_dims(image_tf, axis=0) input_image = tf.image.resize_with_pad(input_image, input_size, input_size) keypoints = movenet(input_image) keypoints = np.array(keypoints) image = tf.keras.preprocessing.image.array_to_img(image_tf) joints = draw_bones(image, keypoints) points = [f"{x}#{y}" for p, x, y in joints] return image, joints, points footer = r"""
Demo for MoveNet
""" with gr.Blocks(title="MoveNet") as app: gr.HTML("

Human Pose Estimation with MoveNet

") gr.HTML("

MoveNet: Ultra fast and accurate pose detection model

") with gr.Row(equal_height=False): with gr.Column(): input_img = gr.Image(type="pil", label="Input image") run_btn = gr.Button(variant="primary") with gr.Column(): output_img = gr.Image(type="numpy", label="Output image") with gr.Accordion("See Positions", open=False): positions = gr.Dataframe( interactive=True, headers=["x", "y", "label"], datatype=["str", "number", "number"], row_count=16, col_count=(3, "fixed"), ) data = gr.Textbox(label="Positions", lines=17) gr.ClearButton(components=[input_img, output_img, positions, data], variant="stop") run_btn.click(predict, [input_img], [output_img, positions, data]) with gr.Row(): blobs = [[f"examples/{x:02d}.jpg"] for x in range(1, 4)] examples = gr.Dataset(components=[input_img], samples=blobs) examples.click(lambda x: x[0], [examples], [input_img]) with gr.Row(): gr.HTML(footer) app.launch(share=False, debug=True, show_error=True) app.queue()