How do I detect the position of hands and fingers in a vision app?
Generated on 7/28/2024
1 search

This post is brought to you by Tailscan,
the best devtool for Tailwind CSS.
To detect the position of hands and fingers in a vision app, you can use the Vision framework's new Swift API. Here's a step-by-step guide based on the information from the WWDC sessions:
-
Create a Detect Human Body Pose Request:
- Set the
detectsHands
property on the request totrue
. This will enable the detection of hands and fingers.
- Set the
-
Process the Human Body Pose Observation:
- The request will produce a human body pose observation, which includes properties for both the right hand and left hand observations.
Here's a brief recap from the session Discover Swift enhancements in the Vision framework:
Create a detect human body pose request and set
detectsHands
on the request to true. This request produces a human body pose observation, which now has two additional properties, one for the right hand observation and one for the left.
For more detailed implementation, you can refer to the chapter "Get started with Vision" in the session Discover Swift enhancements in the Vision framework.
Additionally, if you are working with RealityKit and VisionOS, you can set up spatial tracking of hand anchors to understand hand poses as described in the session Build a spatial drawing app with RealityKit.
For game development, you can use Unity's hand tracking package to access information about the player's joints, as mentioned in the session Explore game input in visionOS.
Relevant Sessions:
- Discover Swift enhancements in the Vision framework
- Build a spatial drawing app with RealityKit
- Explore game input in visionOS
These sessions provide comprehensive guidance on how to implement hand and finger position detection in your vision app.

Explore game input in visionOS
Discover how to design and implement great input for your game in visionOS. Learn how system gestures let you provide frictionless ways for players to interact with your games. And explore best practices for supporting custom gestures and game controllers.

Build immersive web experiences with WebXR
Discover how WebXR empowers you to add fully immersive experiences to your website in visionOS. Find out how to build WebXR experiences that take full advantage of the input capabilities of visionOS, and learn how you can use Simulator to test WebXR experiences on macOS.

Platforms State of the Union
Discover the newest advancements on Apple platforms.

Discover RealityKit APIs for iOS, macOS and visionOS
Learn how new cross-platform APIs in RealityKit can help you build immersive apps for iOS, macOS, and visionOS. Check out the new hover effects, lights and shadows, and portal crossing features, and view them in action through real examples.

Discover Swift enhancements in the Vision framework
The Vision Framework API has been redesigned to leverage modern Swift features like concurrency, making it easier and faster to integrate a wide array of Vision algorithms into your app. We’ll tour the updated API and share sample code, along with best practices, to help you get the benefits of this framework with less coding effort. We’ll also demonstrate two new features: image aesthetics and holistic body pose.

Build a spatial drawing app with RealityKit
Harness the power of RealityKit through the process of building a spatial drawing app. As you create an eye-catching spatial experience that integrates RealityKit with ARKit and SwiftUI, you’ll explore how resources work in RealityKit and how to use features like low-level mesh and texture APIs to achieve fast updates of the users’ brush strokes.