/mediapipe

MediaPipe is a cross-platform framework for building multimodal applied machine learning pipelines

Primary LanguageC++Apache License 2.0Apache-2.0

MediaPipe

MediaPipe is a framework for building multimodal (eg. video, audio, any time series data), cross platform (i.e Android, iOS, web, edge devices) applied ML pipelines. With MediaPipe, a perception pipeline can be built as a graph of modular components, including, for instance, inference models (e.g., TensorFlow, TFLite) and media processing functions.

Real-time Face Detection

"MediaPipe has made it extremely easy to build our 3D person pose reconstruction demo app, facilitating accelerated neural network inference on device and synchronization of our result visualization with the video capture stream. Highly recommended!" - George Papandreou, CTO, Ariel AI

ML Solutions in MediaPipe

face_detection face_mesh hand_tracking multi-hand_tracking hair_segmentation object_detection object_tracking objectron_shoes objectron_chair

Installation

Follow these instructions.

Getting started

See mobile, desktop, web and Google Coral examples.

Check out some web demos [Edge detection] [Face detection] [Hand Tracking]

Documentation

MediaPipe Read-the-Docs or docs.mediapipe.dev

Check out the Examples page for tutorials on how to use MediaPipe. Concepts page for basic definitions

Visualizing MediaPipe graphs

A web-based visualizer is hosted on viz.mediapipe.dev. Please also see instructions here.

Google Open Source Code search

Search MediaPipe Github repository using Google Open Source code search

Videos

Publications

Events

Community forum

  • Discuss - General community discussion around MediaPipe

Alpha Disclaimer

MediaPipe is currently in alpha for v0.7. We are still making breaking API changes and expect to get to stable API by v1.0.

Contributing

We welcome contributions. Please follow these guidelines.

We use GitHub issues for tracking requests and bugs. Please post questions to the MediaPipe Stack Overflow with a 'mediapipe' tag.