[Bug Report] Why my Apple vision pro cannot properly render Isaac Lab environment?
Describe the bug
I'm trying to teleoperate an Isaac Lab robot with apple vision pro. While in AR mode, the teleoperation data can be properly received (Robotic arms and hands in Isaac Lab move according to tracking data published by apple vision pro), but the virtual graphical environment cannot be properly rendered on apple vision pro (only displaying colorful mosaics/pixels). Steps to reproduce
I followed Option 1 in isaac-sim.github.io/IsaacLab/main/source/how-to/cloudxr_teleoperation.html#known-issues both Isaac Lab and CloudXR containers together using Docker Compose, and built and ran the Isaac XR Teleop Sample Client on Apple Vision Pro.
-
./docker/container.py start \ --files docker-compose.cloudxr-runtime.patch.yaml \ --env-file .env.cloudxr-runtime -
./docker/container.py enter baseThen,I run an example teleop task with: 3../isaaclab.sh -p scripts/environments/teleoperation/teleop_se3_agent.py \ --task Isaac-PickPlace-GR1T2-Abs-v0 \ --teleop_device handtracking \ --enable_pinocchio
After clicking "Start AR" in the Isaac Sim UI, I received the following error:
`2025-10-11T02:42:23Z__ [875,296ms] [Error] [omni.usd] TF_PYTHON_EXCEPTION (secondary thread): in TfPyConvertPythonExceptionToTfErrors at line 114 of /builds/omniverse/usd-ci/USD/pxr/base/tf/pyError.cpp -- Tf Python Exception
sys:1: RuntimeWarning: coroutine 'ScriptManager._pending_sync_task_handler' was never awaited RuntimeWarning: Enable tracemalloc to get the object allocation traceback 2025-10-11T02:42:23Z [875,303ms] [Error] [omni.usd] TF_PYTHON_EXCEPTION (secondary thread): in TfPyConvertPythonExceptionToTfErrors at line 114 of /builds/omniverse/usd-ci/USD/pxr/base/tf/pyError.cpp -- Tf Python Exception
2025-10-11T02:42:23Z [875,306ms] [Error] [omni.usd] TF_PYTHON_EXCEPTION (secondary thread): in TfPyConvertPythonExceptionToTfErrors at line 114 of /builds/omniverse/usd-ci/USD/pxr/base/tf/pyError.cpp -- Tf Python Exception
2025-10-11T02:42:23Z [875,309ms] [Error] [omni.usd] TF_PYTHON_EXCEPTION (secondary thread): in TfPyConvertPythonExceptionToTfErrors at line 114 of /builds/omniverse/usd-ci/USD/pxr/base/tf/pyError.cpp -- Tf Python`
System Info
Isaac Sim Version: [5.0] OS: [Ubuntu 22.04] GPU: [ 2 x RTX 3090] CUDA: [12.9] GPU Driver: [575.64.03 ]
After I rebuilt the Docker container, although the terminal no longer shows errors, the Apple Vision Pro still displays a mosaic pattern.
The terminal shows `XR_ERROR_VALIDATION_FAILURE: xrWaitFrame(frameState->type == 0) LOG in xrCreateInstance: Instance created createInfo->applicationInfo.applicationName: Isaac-Sim createInfo->applicationInfo.applicationVersion: 1 createInfo->applicationInfo.engineName: Omniverse Kit XR createInfo->applicationInfo.engineVersion: 107039000 createInfo->applicationInfo.apiVersion: 1.1.45 appinfo.detected.engine.name: (null) appinfo.detected.engine.version: 0.0.0 quirks.disable_vulkan_format_depth: false quirks.disable_vulkan_format_depth_stencil: false quirks.no_validation_error_in_create_ref_space: true quirks.skip_end_session: false quirks.parallel_views: false
LOG in xrGetSystem: Selected devices
Head: 'Head Device'
Eyes: '
Thank you for posting this, the team will investigate this further.
Hi!
Is there any news on this?
I encountered the same issue using a Meta Quest 3, although the Meta Quest 3 is not officially supported by IsaacLab. However, cloudXR has provided an Android sample that allows it to detect both the head and controllers via openXR. Nevertheless, when connecting to the server, IsaacLab also receives the same error: XR_ERROR_FORM_FACTOR_UNSUPPORTED.
Thanks!