Ausnahme gefangen: SSL certificate problem: certificate is not yet valid ๐Ÿ“Œ MediaPipe 3D Face Transform

๐Ÿ  Team IT Security News

TSecurity.de ist eine Online-Plattform, die sich auf die Bereitstellung von Informationen,alle 15 Minuten neuste Nachrichten, Bildungsressourcen und Dienstleistungen rund um das Thema IT-Sicherheit spezialisiert hat.
Ob es sich um aktuelle Nachrichten, Fachartikel, Blogbeitrรคge, Webinare, Tutorials, oder Tipps & Tricks handelt, TSecurity.de bietet seinen Nutzern einen umfassenden รœberblick รผber die wichtigsten Aspekte der IT-Sicherheit in einer sich stรคndig verรคndernden digitalen Welt.

16.12.2023 - TIP: Wer den Cookie Consent Banner akzeptiert, kann z.B. von Englisch nach Deutsch รผbersetzen, erst Englisch auswรคhlen dann wieder Deutsch!

Google Android Playstore Download Button fรผr Team IT Security



๐Ÿ“š MediaPipe 3D Face Transform


๐Ÿ’ก Newskategorie: Programmierung
๐Ÿ”— Quelle: feedproxy.google.com

Posted by Kanstantsin Sokal, Software Engineer, MediaPipe team

Earlier this year, the MediaPipe Team released the Face Mesh solution, which estimates the approximate 3D face shape via 468 landmarks in real-time on mobile devices. In this blog, we introduce a new face transform estimation module that establishes a researcher- and developer-friendly semantic API useful for determining the 3D face pose and attaching virtual objects (like glasses, hats or masks) to a face.

The new module establishes a metric 3D space and uses the landmark screen positions to estimate common 3D face primitives, including a face pose transformation matrix and a triangular face mesh. Under the hood, a lightweight statistical analysis method called Procrustes Analysis is employed to drive a robust, performant and portable logic. The analysis runs on CPU and has a minimal speed/memory footprint on top of the original Face Mesh solution.

MediaPipe image

Figure 1: An example of virtual mask and glasses effects, based on the MediaPipe Face Mesh solution.

Introduction

The MediaPipe Face Landmark Model performs a single-camera face landmark detection in the screen coordinate space: the X- and Y- coordinates are normalized screen coordinates, while the Z coordinate is relative and is scaled as the X coordinate under the weak perspective projection camera model. While this format is well-suited for some applications, it does not directly enable crucial features like aligning a virtual 3D object with a detected face.

The newly introduced module moves away from the screen coordinate space towards a metric 3D space and provides the necessary primitives to handle a detected face as a regular 3D object. By design, you'll be able to use a perspective camera to project the final 3D scene back into the screen coordinate space with a guarantee that the face landmark positions are not changed.

Metric 3D Space

The Metric 3D space established within the new module is a right-handed orthonormal metric 3D coordinate space. Within the space, there is a virtual perspective camera located at the space origin and pointed in the negative direction of the Z-axis. It is assumed that the input camera frames are observed by exactly this virtual camera and therefore its parameters are later used to convert the screen landmark coordinates back into the Metric 3D space. The virtual camera parameters can be set freely, however for better results it is advised to set them as close to the real physical camera parameters as possible.

MediaPipe image

Figure 2: A visualization of multiple key elements in the metric 3D space. Created in Cinema 4D

Canonical Face Model

The Canonical Face Model is a static 3D model of a human face, which follows the 3D face landmark topology of the MediaPipe Face Landmark Model. The model bears two important functions:

  • Defines metric units: the scale of the canonical face model defines the metric units of the Metric 3D space. A metric unit used by the default canonical face model is a centimeter;
  • Bridges static and runtime spaces: the face pose transformation matrix is - in fact - a linear map from the canonical face model into the runtime face landmark set estimated on each frame. This way, virtual 3D assets modeled around the canonical face model can be aligned with a tracked face by applying the face pose transformation matrix to them.

Face Transform Estimation

The face transform estimation pipeline is a key component, responsible for estimating face transform data within the Metric 3D space. On each frame, the following steps are executed in the given order:

  • Face landmark screen coordinates are converted into the Metric 3D space coordinates;
  • Face pose transformation matrix is estimated as a rigid linear mapping from the canonical face metric landmark set into the runtime face metric landmark set in a way that minimizes a difference between the two;
  • A face mesh is created using the runtime face metric landmarks as the vertex positions (XYZ), while both the vertex texture coordinates (UV) and the triangular topology are inherited from the canonical face model.

Effect Renderer

The Effect Renderer is a component, which serves as a working example of a face effect renderer. It targets the OpenGL ES 2.0 API to enable a real-time performance on mobile devices and supports the following rendering modes:

  • 3D object rendering mode: a virtual object is aligned with a detected face to emulate an object attached to the face (example: glasses);
  • Face mesh rendering mode: a texture is stretched on top of the face mesh surface to emulate a face painting technique.

In both rendering modes, the face mesh is first rendered as an occluder straight into the depth buffer. This step helps to create a more believable effect via hiding invisible elements behind the face surface.

MediaPipe image

Figure 3: An example of face effects rendered by the Face Effect Renderer.

Using Face Transform Module

The face transform estimation module is available as a part of the MediaPipe Face Mesh solution. It comes with face effect application examples, available as graphs and mobile apps on Android or iOS. If you wish to go beyond examples, the module contains generic calculators and subgraphs - those can be flexibly applied to solve specific use cases in any MediaPipe graph. For more information, please visit our documentation.

Follow MediaPipe

We look forward to publishing more blog posts related to new MediaPipe pipeline examples and features. Please follow the MediaPipe label on Google Developers Blog and Google Developers twitter account (@googledevs).

Acknowledgements

We would like to thank Chuo-Ling Chang, Ming Guang Yong, Jiuqiang Tang, Gregory Karpiak, Siarhei Kazakou, Matsvei Zhdanovich and Matthias Grundman for contributing to this blog post.

...



๐Ÿ“Œ MediaPipe 3D Face Transform


๐Ÿ“ˆ 43.51 Punkte

๐Ÿ“Œ Training a new model with MediaPipe Model Maker - ML on Android with MediaPipe


๐Ÿ“ˆ 40.96 Punkte

๐Ÿ“Œ Face and hand tracking in the browser with MediaPipe and TensorFlow.js


๐Ÿ“ˆ 29.62 Punkte

๐Ÿ“Œ Implementing face stylization - ML on Android with MediaPipe


๐Ÿ“ˆ 29.62 Punkte

๐Ÿ“Œ Detecting face landmarks in MediaPipe for Raspberry Pi


๐Ÿ“ˆ 29.62 Punkte

๐Ÿ“Œ Transform decorator (@transform) not working using web-pack in NestJs


๐Ÿ“ˆ 27.78 Punkte

๐Ÿ“Œ How face IDs works? Oneplus face unlock v iPhone Face ID v Galaxy iris scanner: Which is more secure?


๐Ÿ“ˆ 27.42 Punkte

๐Ÿ“Œ Object Detection and Tracking using MediaPipe


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Alfred Camera: Smart camera features using MediaPipe


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ MediaPipe KNIFT: Template-based Feature Matching


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Instant Motion Tracking with MediaPipe


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Instant Motion Tracking with MediaPipe, MySQL 8, ML Kit Pose Detection, and more!


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Iris landmark tracking in the browser with MediaPipe and TensorFlow.js


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ SignAll SDK: Sign language interface using MediaPipe is now available for developers


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ High Fidelity Pose Tracking with MediaPipe BlazePose and TensorFlow.js


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Control your Mirru prosthesis with MediaPipe hand tracking


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ I updated my webcam background replacement tool to use Google's Mediapipe as image segmentation backend. It is now so much faster!


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Body Segmentation with MediaPipe and TensorFlow.js


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Introduction to On-device Machine Learning Solutions with MediaPipe


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Advanced on-device ML made easy with MediaPipe


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Get inspired in 2023 with new machine learning solutions for web developers with MediaPipe


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Introducing MediaPipe Solutions for On-Device Machine Learning


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ 5 things to know before customizing your first machine learning model with MediaPipe Model Maker


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ MediaPipe On-Device Text-to-Image Generation Solution Now Available for Android Developers


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ 7 dos and don'ts of using ML on the web with MediaPipe


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ MediaPipe for Raspberry Pi and iOS


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ HealthPulse AI Leverages MediaPipe to Increase Health Equity


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ MediaPipe on the Web


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Introduction to ML on Android with MediaPipe


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Introduction to computer vision - ML on Android with MediaPipe


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Applying computer vision tasks to Android apps - ML on Android with MediaPipe


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Getting started with Natural Language Processing - ML on Android with MediaPipe


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Large Language Models On-Device with MediaPipe and TensorFlow Lite


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ Human Pose Tracking with MediaPipe in 2D and 3D: Rerun Showcase


๐Ÿ“ˆ 20.48 Punkte

๐Ÿ“Œ The MediaPipe LLM Inference API lets you run LLMs on Android and iOS


๐Ÿ“ˆ 20.48 Punkte











matomo