Camillo Lugaresi, Jiuqiang Tang, Hadon Nash, Chris McClanahan, Esha Uboweja, Michael Hays, Fan Zhang, Chuo-Ling Chang, Ming Guang Yong, Juhyun Lee, Wan-Teh Chang, Wei Hua, Manfred Georg and Matthias Grundmann
Google has long used MediaPipe in its products and services. Since 2012, it has been used for real-time analysis of video and audio on YouTube. Over time MediaPipe has been incorporated into many more products such as Gmail, Google Home, etc.[20]
MediaPipe's first stable release was version 0.5.0.[21] It was made open source in June 2019 at the Conference on Computer Vision and Pattern Recognition in Long Beach, California, by Google Research. This initial release included only five pipelines examples: Object Detection, Face Detection, Hand Tracking, Multi-hand Tracking, and Hair Segmentation.[22] From its initial release to April 2023, numerous pipelines have been made. In May 2025, MediaPipe Solutions was introduced. This transition offered more capabilities for on-device machine learning.[23] MediaPipe is now under Google's subdivision, Google AI Edge.
MediaPipe is primarily written in the programming languageC++, although this is not the sole programing language used in its creation. The other notable programming languages used within its source code include Python, Starlark, and Java.[21]
The ability for MediaPipe to separate itself into a system of components allows for customization. Pre-built solutions are also available and it may help to start with these and slightly optimize them for an ideal output.[24]
How MediaPipe Works
MediaPipe contains a multitude of different components that all work together to create a general purpose computer vision framework. Each component works in its own unique way with different architectures.
Hand Tracking
MediaPipe includes a hand tracking system that has been designed to run efficiently on devices with limited computational resources. This works by estimating a set of 3D landmarks for each detected hand and is intended to remain stable across a wide range of environments including different poses, lightning conditions, and motions.[25]
MediaPipe works off of a pre-trained deep learning model that is trained to detect the palm area on human hands, which is done through a detector model named BlazePalm.[24] Starting with the identification of the palm, MediaPipe is able to use the positioning of the palm as an input to a second model that predicts the positions of key landmarks that will represent the hand's structure.[25]
Hands before MediaPipe hand detectionHands after MediaPipe hand detection
MediaPipe continuously monitors the confidence of its predictions and re-runs detection when needed to maintain its accuracy, while temporal smoothing helps reduce the jitter between frames. For scenes with more than one hand, the process is repeated independently for each detected region.[25][24]
Human Pose Estimation
Another area that MediaPipe specializes in is recognizing changes in the human body specifically posture. Mediapipe can support the creation of body posture analysis systems. This can aid in many fields such as ergonomic industry, the arts, sports, and entertainment. [24]
↑Sreenath, Sreehari; Daniels, D. Ivan; Ganesh, Apparaju S. D.; Kuruganti, Yashaswi S.; Chittawadigi, Rajeevlochana G. (2021-09-30). "Monocular Tracking of Human Hand on a Smart Phone Camera using MediaPipe and its Application in Robotics". 2021 IEEE 9th Region 10 Humanitarian Technology Conference (R10-HTC). IEEE. pp.1–6. doi:10.1109/R10-HTC53172.2021.9641542. ISBN978-1-6654-3240-5.
↑Nunes, João; Nascimento, Thamer Horbylon; Felix, Juliana; Soares, Fabrizzio (2025-07-08). "Real-Time Hand Gesture Recognition for Touchless Video Control Using MediaPipe and Random Forest". 2025 IEEE 49th Annual Computers, Software, and Applications Conference (COMPSAC). IEEE. pp.1776–1781. doi:10.1109/COMPSAC65507.2025.00242. ISBN979-8-3315-7434-5.
↑Patel, Meenu; Rao, Saksham; Chauhan, Shweta; Kumar, Bibek (2024-12-16). "Real-time Hand Gesture Recognition Using Python and Web Application". 2024 1st International Conference on Advances in Computing, Communication and Networking (ICAC2N). IEEE. pp.564–570. doi:10.1109/ICAC2N63387.2024.10895151. ISBN979-8-3503-5681-6.
↑Kulkarni, Pavan Kumar V; M S, Sudha; K R, Divya; S, Vignesh; M, Sindhu (2024-05-22). "Real-Time Gesture Recognition For Arduino-Based LED and Servometer Manipulation Using OpenCV and MediaPipe". 2024 1st International Conference on Communications and Computer Science (InCCCS). IEEE. pp.1–5. doi:10.1109/InCCCS60947.2024.10593524. ISBN979-8-3503-5885-8.
↑Williams-Linera, Eric; Ramírez-Cortés, Juan Manuel (2024-09-18). "Stereo Vision System based on the NVIDIA Jetson Nano for Real-time Evaluation of Yoga Poses". 2024 IEEE International Symposium on Technology and Society (ISTAS). IEEE. pp.1–7. doi:10.1109/ISTAS61960.2024.10732331. ISBN979-8-3315-4070-8.
This page is based on this Wikipedia article Text is available under the CC BY-SA 4.0 license; additional terms may apply. Images, videos and audio are available under their respective licenses.