HTML demos for face, hand, gesture, and posture tracking using MediaPipe. Includes Python CLI tools for processing video files. Co-Authored-By: Claude Opus 4.5 <noreply@anthropic.com>
299 lines
9.5 KiB
HTML
299 lines
9.5 KiB
HTML
<!DOCTYPE html>
|
|
<html lang="en">
|
|
<head>
|
|
<meta charset="utf-8">
|
|
<meta http-equiv="Cache-control" content="no-cache, no-store, must-revalidate">
|
|
<meta http-equiv="Pragma" content="no-cache">
|
|
<meta name="viewport" content="width=device-width, initial-scale=1, user-scalable=no">
|
|
<title>Pose Landmarker — Single File Demo</title>
|
|
|
|
<!-- Material Components (for the button styling) -->
|
|
<link href="https://unpkg.com/material-components-web@latest/dist/material-components-web.min.css" rel="stylesheet">
|
|
<script src="https://unpkg.com/material-components-web@latest/dist/material-components-web.min.js"></script>
|
|
|
|
<style>
|
|
/* Copyright 2023 The MediaPipe Authors.
|
|
Licensed under the Apache License, Version 2.0 (the "License");
|
|
you may not use this file except in compliance with the License.
|
|
You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0
|
|
Unless required by applicable law or agreed to in writing, software
|
|
distributed under the License is distributed on an "AS IS" BASIS,
|
|
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
See the License for the specific language governing permissions and
|
|
limitations under the License. */
|
|
|
|
/* NOTE: The original CSS used `@use "@material";` which is a Sass directive.
|
|
That's not valid in plain CSS, so it's removed here. */
|
|
|
|
body {
|
|
font-family: Roboto, system-ui, -apple-system, Segoe UI, Arial, sans-serif;
|
|
margin: 2em;
|
|
color: #3d3d3d;
|
|
--mdc-theme-primary: #007f8b;
|
|
--mdc-theme-on-primary: #f1f3f4;
|
|
}
|
|
|
|
h1 { color: #007f8b; }
|
|
h2 { clear: both; }
|
|
|
|
em { font-weight: bold; }
|
|
|
|
video {
|
|
clear: both;
|
|
display: block;
|
|
transform: rotateY(180deg);
|
|
-webkit-transform: rotateY(180deg);
|
|
-moz-transform: rotateY(180deg);
|
|
}
|
|
|
|
section {
|
|
opacity: 1;
|
|
transition: opacity 500ms ease-in-out;
|
|
}
|
|
|
|
header, footer { clear: both; }
|
|
|
|
.removed { display: none; }
|
|
.invisible { opacity: 0.2; }
|
|
|
|
.note {
|
|
font-style: italic;
|
|
font-size: 130%;
|
|
}
|
|
|
|
.videoView, .detectOnClick {
|
|
position: relative;
|
|
float: left;
|
|
width: 48%;
|
|
margin: 2% 1%;
|
|
cursor: pointer;
|
|
}
|
|
|
|
.videoView p, .detectOnClick p {
|
|
position: absolute;
|
|
padding: 5px;
|
|
background-color: #007f8b;
|
|
color: #fff;
|
|
border: 1px dashed rgba(255, 255, 255, 0.7);
|
|
z-index: 2;
|
|
font-size: 12px;
|
|
margin: 0;
|
|
}
|
|
|
|
.highlighter {
|
|
background: rgba(0, 255, 0, 0.25);
|
|
border: 1px dashed #fff;
|
|
z-index: 1;
|
|
position: absolute;
|
|
}
|
|
|
|
.canvas {
|
|
z-index: 1;
|
|
position: absolute;
|
|
pointer-events: none;
|
|
}
|
|
|
|
.output_canvas {
|
|
transform: rotateY(180deg);
|
|
-webkit-transform: rotateY(180deg);
|
|
-moz-transform: rotateY(180deg);
|
|
}
|
|
|
|
.detectOnClick { z-index: 0; }
|
|
.detectOnClick img { width: 100%; }
|
|
|
|
/* Simple layout fix for the video/canvas wrapper */
|
|
.video-wrapper {
|
|
position: relative;
|
|
width: 1280px;
|
|
max-width: 100%;
|
|
aspect-ratio: 16 / 9;
|
|
}
|
|
.video-wrapper video,
|
|
.video-wrapper canvas {
|
|
position: absolute;
|
|
top: 0;
|
|
left: 0;
|
|
width: 100%;
|
|
height: 100%;
|
|
}
|
|
</style>
|
|
</head>
|
|
|
|
<body>
|
|
<h1>Pose detection using the MediaPipe PoseLandmarker task</h1>
|
|
|
|
<section id="demos" class="invisible">
|
|
<h2>Demo: Webcam continuous pose landmarks detection</h2>
|
|
<p>Stand in front of your webcam to get real-time pose landmarker detection.<br>Click <b>enable webcam</b> below and grant access to the webcam if prompted.</p>
|
|
|
|
<div id="liveView" class="videoView">
|
|
<button id="webcamButton" class="mdc-button mdc-button--raised">
|
|
<span class="mdc-button__ripple"></span>
|
|
<span class="mdc-button__label">ENABLE WEBCAM</span>
|
|
</button>
|
|
<div class="video-wrapper">
|
|
<video id="webcam" autoplay playsinline></video>
|
|
<canvas class="output_canvas" id="output_canvas" width="1280" height="720"></canvas>
|
|
</div>
|
|
</div>
|
|
</section>
|
|
|
|
<script type="module">
|
|
// Copyright 2023 The MediaPipe Authors.
|
|
// Licensed under the Apache License, Version 2.0 (the "License");
|
|
|
|
import {
|
|
PoseLandmarker,
|
|
FilesetResolver,
|
|
DrawingUtils
|
|
} from "https://cdn.skypack.dev/@mediapipe/tasks-vision@0.10.0";
|
|
|
|
const demosSection = document.getElementById("demos");
|
|
|
|
let poseLandmarker = undefined;
|
|
let runningMode = "IMAGE";
|
|
let enableWebcamButton;
|
|
let webcamRunning = false;
|
|
const videoHeight = "360px";
|
|
const videoWidth = "480px";
|
|
|
|
// Load the Vision WASM and the Pose Landmarker model
|
|
const createPoseLandmarker = async () => {
|
|
const vision = await FilesetResolver.forVisionTasks(
|
|
"https://cdn.jsdelivr.net/npm/@mediapipe/tasks-vision@0.10.0/wasm"
|
|
);
|
|
poseLandmarker = await PoseLandmarker.createFromOptions(vision, {
|
|
baseOptions: {
|
|
modelAssetPath: "https://storage.googleapis.com/mediapipe-models/pose_landmarker/pose_landmarker_lite/float16/1/pose_landmarker_lite.task",
|
|
delegate: "GPU"
|
|
},
|
|
runningMode: runningMode,
|
|
numPoses: 2
|
|
});
|
|
demosSection.classList.remove("invisible");
|
|
};
|
|
createPoseLandmarker();
|
|
|
|
/********************************************************************
|
|
// Demo 1: Click an image to detect pose and draw landmarks.
|
|
********************************************************************/
|
|
const imageContainers = document.getElementsByClassName("detectOnClick");
|
|
for (let i = 0; i < imageContainers.length; i++) {
|
|
imageContainers[i].children[0].addEventListener("click", handleClick);
|
|
}
|
|
|
|
async function handleClick(event) {
|
|
if (!poseLandmarker) {
|
|
console.log("Wait for poseLandmarker to load before clicking!");
|
|
return;
|
|
}
|
|
|
|
if (runningMode === "VIDEO") {
|
|
runningMode = "IMAGE";
|
|
await poseLandmarker.setOptions({ runningMode: "IMAGE" });
|
|
}
|
|
|
|
// Remove old overlays
|
|
const allCanvas = event.target.parentNode.getElementsByClassName("canvas");
|
|
for (let i = allCanvas.length - 1; i >= 0; i--) {
|
|
const n = allCanvas[i];
|
|
n.parentNode.removeChild(n);
|
|
}
|
|
|
|
poseLandmarker.detect(event.target, (result) => {
|
|
const canvas = document.createElement("canvas");
|
|
canvas.setAttribute("class", "canvas");
|
|
canvas.setAttribute("width", event.target.naturalWidth + "px");
|
|
canvas.setAttribute("height", event.target.naturalHeight + "px");
|
|
canvas.style =
|
|
"left: 0px; top: 0px; width: " + event.target.width + "px; height: " + event.target.height + "px;";
|
|
|
|
event.target.parentNode.appendChild(canvas);
|
|
const canvasCtx = canvas.getContext("2d");
|
|
const drawingUtils = new DrawingUtils(canvasCtx);
|
|
|
|
for (const landmark of result.landmarks) {
|
|
drawingUtils.drawLandmarks(landmark, {
|
|
radius: (data) => DrawingUtils.lerp((data.from && data.from.z) ?? 0, -0.15, 0.1, 5, 1)
|
|
});
|
|
drawingUtils.drawConnectors(landmark, PoseLandmarker.POSE_CONNECTIONS);
|
|
}
|
|
});
|
|
}
|
|
|
|
/********************************************************************
|
|
// Demo 2: Live webcam pose detection.
|
|
********************************************************************/
|
|
const video = document.getElementById("webcam");
|
|
const canvasElement = document.getElementById("output_canvas");
|
|
const canvasCtx = canvasElement.getContext("2d");
|
|
const drawingUtils = new DrawingUtils(canvasCtx);
|
|
|
|
const hasGetUserMedia = () => !!(navigator.mediaDevices && navigator.mediaDevices.getUserMedia);
|
|
|
|
if (hasGetUserMedia()) {
|
|
enableWebcamButton = document.getElementById("webcamButton");
|
|
enableWebcamButton.addEventListener("click", enableCam);
|
|
} else {
|
|
console.warn("getUserMedia() is not supported by your browser");
|
|
}
|
|
|
|
function enableCam() {
|
|
if (!poseLandmarker) {
|
|
console.log("Wait! poseLandmarker not loaded yet.");
|
|
return;
|
|
}
|
|
|
|
if (webcamRunning === true) {
|
|
webcamRunning = false;
|
|
enableWebcamButton.innerText = "ENABLE PREDICTIONS";
|
|
} else {
|
|
webcamRunning = true;
|
|
enableWebcamButton.innerText = "DISABLE PREDICTIONS";
|
|
}
|
|
|
|
const constraints = { video: true };
|
|
|
|
navigator.mediaDevices.getUserMedia(constraints).then((stream) => {
|
|
video.srcObject = stream;
|
|
video.addEventListener("loadeddata", predictWebcam);
|
|
});
|
|
}
|
|
|
|
let lastVideoTime = -1;
|
|
async function predictWebcam() {
|
|
canvasElement.style.height = videoHeight;
|
|
video.style.height = videoHeight;
|
|
canvasElement.style.width = videoWidth;
|
|
video.style.width = videoWidth;
|
|
|
|
if (runningMode === "IMAGE") {
|
|
runningMode = "VIDEO";
|
|
await poseLandmarker.setOptions({ runningMode: "VIDEO" });
|
|
}
|
|
|
|
const startTimeMs = performance.now();
|
|
if (lastVideoTime !== video.currentTime) {
|
|
lastVideoTime = video.currentTime;
|
|
poseLandmarker.detectForVideo(video, startTimeMs, (result) => {
|
|
canvasCtx.save();
|
|
canvasCtx.clearRect(0, 0, canvasElement.width, canvasElement.height);
|
|
for (const landmark of result.landmarks) {
|
|
drawingUtils.drawLandmarks(landmark, {
|
|
radius: (data) => DrawingUtils.lerp((data.from && data.from.z) ?? 0, -0.15, 0.1, 5, 1)
|
|
});
|
|
drawingUtils.drawConnectors(landmark, PoseLandmarker.POSE_CONNECTIONS);
|
|
}
|
|
canvasCtx.restore();
|
|
});
|
|
}
|
|
|
|
if (webcamRunning === true) {
|
|
window.requestAnimationFrame(predictWebcam);
|
|
}
|
|
}
|
|
</script>
|
|
</body>
|
|
</html>
|