VisionOS new features

Asked on 2024-07-30

1 search

Apple introduced several new features for VisionOS at WWDC 2024. Here are some of the highlights:

  1. Scene Understanding Enhancements:

    • Planes Detection: Now supports all orientations, allowing for anchoring objects on various surfaces.
    • Room Anchors: These consider the user's surroundings on a per-room basis and can detect movement across rooms.
    • Object Tracking API: This new API allows attaching content to individual objects around the user, enabling new dimensions of interactivity. For example, you can attach virtual instructions to physical objects.
  2. Hand Interaction:

    • Developers can now decide if the user's hands appear in front of or behind the content, providing more creative control over app experiences.
  3. HealthKit Integration:

    • VisionOS supports immersive spaces where users can reflect on their activities and save their experiences to HealthKit. This feature includes a "Guest User" mode, allowing others to try out Apple Vision Pro while preserving the owner's data and privacy.
  4. Enterprise APIs:

    • Enhanced known object tracking through parameter adjustment, allowing for tuning and optimizing object tracking to suit specific use cases.
  5. ARKit Updates:

    • Room Tracking: Customizes experiences based on the room.
    • Plane Detection: Improved to support more complex environments.
    • Hand Tracking: Used in apps like Blackbox and Super Fruit Ninja for interactive experiences.

For more detailed information, you can refer to the following sessions:

Relevant Sessions

  1. Platforms State of the Union
  2. Get started with HealthKit in visionOS
  3. Introducing enterprise APIs for visionOS
  4. Create enhanced spatial computing experiences with ARKit