Unity play mode streaming on device

I need a way to delevop a project with Magic Leap 2 + OpenXR + MRTK3, testing and prototyping the app using the Unity “play mode”, remote rendering on the visor the Unity scene, but using the visor hand inputs + cameras streaming as input for the host.

The project should do extensive use of CV, using world cameras (or RGB) as input to scan a surface + native SLAM for tracking.

How can I test the app after you deprecated every “remote rendering” solution? (without every time compiling and deploying it to the device).

I previously used Hololens 2 and its MRTK3 remote rendering solution to fast prototyping.

I miss something on your docs or there is no way ?

Unity Editor version: 6000.0.37f1
ML2 OS version: 1.12
Unity SDK version: 2.6.0
Host OS: Windows

Remote Rendering never officially supported Unity applications - that said we do not have an alternative at this time. You can try to use the XRI or MRTK simulator from the editor and interact using mouse and keyboard.

thanks for the reply.

ok, to test the UI the simulator maybe the way.
I already use that.

But how to prototype the camera input and related algorithms?
Should have I to build and deploy every time?
Any ideas?
There’s no way to replicate what Microsoft Hololens did 6 years ago?

You are correct. You will need to build and deploy your application to test device inputs.