VisionOS new features

Asked on 07/30/2024

1 search

Apple introduced several new features for VisionOS at WWDC 2024. Here are some of the highlights:

  1. Scene Understanding Enhancements:

    • Planes Detection: Now supports all orientations, allowing for anchoring objects on various surfaces.
    • Room Anchors: These consider the user's surroundings on a per-room basis and can detect movement across rooms.
    • Object Tracking API: This new API allows attaching content to individual objects around the user, enabling new dimensions of interactivity. For example, you can attach virtual instructions to physical objects.
  2. Hand Interaction:

    • Developers can now decide if the user's hands appear in front of or behind the content, providing more creative control over app experiences.
  3. HealthKit Integration:

    • VisionOS supports immersive spaces where users can reflect on their activities and save their experiences to HealthKit. This feature includes a "Guest User" mode, allowing others to try out Apple Vision Pro while preserving the owner's data and privacy.
  4. Enterprise APIs:

    • Enhanced known object tracking through parameter adjustment, allowing for tuning and optimizing object tracking to suit specific use cases.
  5. ARKit Updates:

    • Room Tracking: Customizes experiences based on the room.
    • Plane Detection: Improved to support more complex environments.
    • Hand Tracking: Used in apps like Blackbox and Super Fruit Ninja for interactive experiences.

For more detailed information, you can refer to the following sessions:

Relevant Sessions

  1. Platforms State of the Union
  2. Get started with HealthKit in visionOS
  3. Introducing enterprise APIs for visionOS
  4. Create enhanced spatial computing experiences with ARKit
VisionOS new features | Ask WWDC