2.Attitude detection

2.1. Introduction

MediaPipe is an open source data stream processing machine learning application development framework developed by Google. It is a graph-based data processing pipeline for building and using multiple forms of data sources, such as video, audio, sensor data, and any time series data. MediaPipe is cross-platform and can run on embedded platforms (Raspberry Pi, etc.), mobile devices (iOS and Android), workstations and servers, and supports mobile GPU acceleration. MediaPipe provides cross-platform, customizable ML solutions for real-time and streaming media.

The core framework of MediaPipe is implemented in C++ and provides support for languages such as Java and Objective C. The main concepts of MediaPipe include Packet, Stream, Calculator, Graph and Subgraph.

MediaPipe Features:

2.2, MediaPipe Pose

MediaPipe Pose is an ML solution for high-fidelity body pose tracking,using BlazePose research, 33 3D coordinates and full background segmentation masks are inferred from RGB video frames. This research also provides motivation for the ML Kit pose detection API.

The landmark model in MediaPipe poses predicts the positions of 33 pose coordinates (see figure below).

image-20231016095802066

2.3. Attitude detection

2.3.1.Startup

Note: Before running this case, please make sure that the [/dev/video0] device has been successfully mounted into the docker container, otherwise the camera will not be opened.

Enter the docker container and execute:

image-20231016100146613

 

To view the point cloud data and node topic communication diagram, please refer to the [Hand Detection] case course.

 

2.3.2. Source code

Source code location: