/posenet-python

A port of Google TensorFlow.js PoseNet to Python

Primary LanguagePythonApache License 2.0Apache-2.0

PoseNet Python

This repository contains a pure Python implementation (multi-pose only) of the Google TensorFlow.js Posenet model.

I first adapted the JS code more or less verbatim and found the performance was low so made some vectorized numpy/scipy version of a few key functions (named _fast).

Further optimization is possible. The MobileNet base models have a throughput of 200-300 fps on a GTX 1080 Ti (or better). The fast post processing code limits this to about 80-100fps if all file io and drawing is removed from the loop. A Cython or pure C++ port would be ideal.

Install

A suitable Python 3.x environment with a recent version of Tensorflow is required. Development and testing was done with Python 3.6.8 and Tensorflow 1.12.0 from Conda.

A conda environment with these packages should suffice: conda install tensorflow-gpu scipy pyyaml opencv

Note: If you want to use the webcam demo, a pip version of opencv (pip install python-opencv) is required instead of the conda version. Anaconda's default opencv does not include ffpmeg/VideoCapture support.

Using pip 8.1.1 and Python 3.5.2 on Ubuntu 16.04

  1. clone repository
  2. cd into root directory of repository
  3. $ python3 -m venv venv
  4. $ source venv/bin/activate
  5. $ pip install -r requirements.txt
Using Intel RealSense camera (for 3D estimation)

You need librealsense installed on your machine and additionally install python bindings via pip install pyrealsense2

Usage

There are three demo apps in the root that utilize the PoseNet model. They are very basic and could definitely be improved.

The first time these apps are run (or the library is used) model weights will be downloaded from the TensorFlow.js version and converted on the fly.

For all demos, the model can be specified with the '--model` argument by using its ordinal id (0-3) or integer depth multiplier (50, 75, 100, 101). The default is the 101 model.

image_demo.py

Image demo runs inference on an input folder of images and outputs those images with the keypoints and skeleton overlayed.

python image_demo.py --model 101 --image_dir ./images --output_dir ./output

A folder of suitable test images can be downloaded by first running the get_test_images.py script.

benchmark.py

A minimal performance benchmark based on image_demo. Images in --image_dir are pre-loaded and inference is run --num_images times with no drawing and no text output.

webcam_demo.py

The webcam demo uses OpenCV to capture images from a connected webcam. The result is overlayed with the keypoints and skeletons and rendered to the screen. The default args for the webcam_demo assume device_id=0 for the camera and that 1280x720 resolution is possible.

Credits

The original model, weights, code, etc. was created by Google and can be found at https://github.com/tensorflow/tfjs-models/tree/master/posenet

This port and my work is in no way related to Google.

The Python conversion code that started me on my way was adapted from the CoreML port at https://github.com/infocom-tpo/PoseNet-CoreML

TODO (someday, maybe)

  • More stringent verification of correctness against the original implementation
  • Performance improvements (especially edge loops in 'decode.py')
  • OpenGL rendering/drawing
  • Comment interfaces, tensor dimensions, etc
  • Implement batch inference for image_demo

TODO (soon, definitely)

  • save converted models somewhere (gdrive?)