I'm trying to develop an app to Oculus Quest
yet i encounter many issues on the way.
the app basically consist of a room and teleportation.
the issue i encountered happens both on unity 2018.3.12 and 2019.1.8.
I created an app that did work as expected before on Quest using Oculus integration v1.35.
however, when re-exported and installed to a brand new Quest device:
teleportation that worked before refuse to work
pressing a button on 1 controller hids the other one from view.
hands movement is limited
even though target devices is set to Quest, I still (on v1.35) see the controller of oculus go
it only happens when exporting to quest, on rift it works just fine.
from the posts in Oculus Forum
https://forums.oculusvr.com/developer/discussion/comment/702108#Comment_702108
https://forums.oculusvr.com/developer/discussion/79144/hands-not-showing-up-with-localavator-unity#latest
it looks like there is a firmware issue (but then you should be encountering that too no?)
things I tried:
Start a new project from scratch:
followed the standard tutorial and documentation like those
https://www.youtube.com/watch?v=qiJpjnzW-mw&t=1s
https://developer.oculus.com/documentation/quest/latest/concepts/book-unity-gsg/
could not see the hands at all, and no teleportation was implemented, it is said that there is a bug in current v1.39
https://forums.oculusvr.com/developer/discussion/79144/hands-not-showing-up-with-localavator-unity
I tried to use both the unity OVR assets and the following plugins
https://assetstore.unity.com/packages/tools/input-management/vr-movement-system-for-oculus-47292
https://assetstore.unity.com/packages/tools/input-management/vr-arc-teleporter-61561
and in all 3 encountered the same issue.
did anyone encounter any issue similar to what's described?
as said, i expect to see the ands and controllers, and the code attached to the trigger press executed. none of this happened.
as some test, (since i do not have another quest)
i exported a build and sharing here.
please comment if you have tried it and it workd fine on your quest or not
https://www.dropbox.com/s/uvcmhyar2qljb19/k14.apk?dl=0
From what you wrote it's seems like an update issue with your device.
Check references with the new device .
For me when i publish i use the 2018.4.1f LTE
Since i encountered a lot of issues with android build with the 2019 version.
Go to the company and ask what changed with the new versions.
thanks to all the helpers.
update 1:
if hands don't appear at all, its because Oculus integration 1.39 is buggy
revert to 1.38
https://developer.oculus.com/downloads/package/unity-integration-archive/
second:
the problem is that the latest firmware update added GO controller support and defaulted it.
so Quest controllers begun appearing as GO
the trick was to copy the AndroidManifest to Assets/Plugins and edit it properly to specifically define the correct settings.
you can see the recommended changes here
https://developer.oculus.com/documentation/quest/latest/concepts/mobile-native-manifest/?fbclid=IwAR3AgasGPJFVGsz7lyzfJNfuTB8R1FOg88Quq8YZz67eQlwEFvgEMDGjSdo
Related
So this question I think is the same as this but the solution for this problem was downgrading to Unity 2017 LTS, which is incompatible with my current project.
Basically, whenever I build, there's maybe a 10% chance that the app runs in the Hololens. Most of the time however, no floating balls animation pops up, no Unity splash screen appears, and I don't see any of my app content. Strangely enough, the app does ask for microphone permissions, but that's all. It should ask for eye tracking too, but it doesn't. (Not sure if that's related but I'm out of ideas.) On a proper run, I get the floating balls, Unity splash, and all permissions asked before the content. The most frustrating part is that it sometimes works, and there has been no trend at all towards what helps or what doesn't.
I've tried rebuilding in Unity, redeploying in VS, reinstalling the appx, updating VS, updating Windows, and checked every forum post I could find but I can't figure this out. Does anyone have any ideas?
I thought it might be an OpenXR / Holographic remoting problem, but those have both been dead ends. Really not sure what to do at this point. Thanks in advance.
Unity version: 2020.3.36f1,
VS 2022, MRTKv2, HoloLens 2
Thanks for the help everyone! For some reason, my solution is just as confusing as the problem, but it seems to reliably work.
Problem has something to do with the Holographic Remoting for OpenXR, so simply enabling and then disabling that feature in the editor right before building seems to fix a buggy build. Tested this a few times now and it works pretty consistently. Hope this can help someone and eventually get patched.
EDIT: Previous answer didn't work. Now I'm thinking the problem was actually caused by a toggle in the Project Settings. For a successful build it seems as though XR Plug-in Management > Windows Mixed Reality > UWP > Use Primary Window should be checked on. Seems weird since I'm using OpenXR anyways, but this might help someone so thought I'd update. Works most of the time now - needs more testing.
There are two solutions:
Use "Windows Mixed Reality" plug-in provider.
Use "OpenXR plugin". For this case the order of installation of packages plays important role:
When you create new project, manually install OpenXR plugin using Package Manager of Unity.
Using "Mixed Reality Feature Tool", install other features of MRTK2 including "Mixed Reality Open XR plugin" in "Platform Support".
XR Plug-in Management
OpenXR features
I use the G2 to develop VR software in UE4 version 4.26.
we noticed that the facebuttons on either controller stopped working on Monday. at first we thought it was something I did in Unreal, but I can't get to the root of the problem, and all the other input is registering correctly (grab, trigger).
in some older versions that we know worked, we built again to test, now aren't working either.
beginning to think it's a bug in a controller update from Monday.
are there any known bugs?
Edit: 3/15
#Hernando-MSFT
After significant time spent on this I found out what the problem was.
The Windows Mixed Reality plugin in UE4 was preventing the G2's buttons from registering input. I disabled the plugin and then when opening the UE4 app it would run in OpenXR instead of WMR and now the buttons register.
Why that only started happening last week; and why the G2's buttons would not register on its native platform (WMR) is very mysterious.
Unity Version: Unity 2019.3.9f1
Vuforia Version: 9.0.12
The app works fine in unity editor, it builds and runs successfully without error but no camera feed (black screen). Can't seem to find up to date solutions online.
Updated the Vuforia version to 9.2.7 does not seem to solve the problem.
I don't think to check the "Vuforia Augmented Reality Supported" under the "XR Settings" is an option, because it's able to deprecate. Also Checking the "Windows Mixed Reality" plug-in under XR Plug-in Management didn't solve the problem.
I already checked the "InternetClient" and "WebCam" in the publish setting for UWP.
Do I need to change anything in ARCamera's "Open Vuforia Engine configuration"? Some solutions mention disabling the Vuforia play mode will help, but I can't seem to find the option anywhere.
Anyone encountered the same issue and find a solution? Thanks.
I don't have a current solution, but would like to share my experience working with vuforia, UWP.
I found that there are varying breaking versions of unity and vuforia's SDK.
The latest version may not necessarily work I had to trial and error before I found a compatible version, so you might want to try that out. perhaps a older version of either unity/vuforia in reference with https://library.vuforia.com/platform-support/supported-versions.html
And yeah, definitely run your "hello world" tests on the sample projects first.
Best of luck, it's a great SDK ,once you get it to work.
I am currently working on developing an application for the Hololens 1 using Unity and MRTK.
I have been unable to get the air tap or any other input working in my application. The ring pointer for airtap does not appear in the application even though it works in the Unity Play Mode Input Simulation and other applications on the device. I tried it with the MRTK examples and even those did not work on the device.
Also I had to add the Tracked Pose Driver from the Player Settings to get the camera working properly but have not figured out how to get the application to accept gesture input.
Also tried the solution listed here: Why is 'air tap' gesture not working on HoloLens1 in my Unity/MRTK app?
but that did not work.
Will appreciate any guidance to solve this problem.
I finally got my project working after a lot of trying.
It turned out that initially, I had unknowingly configured the project to use XR SDK in Unity 2019 and thus had to configure MRTK accordingly. Thank you to Kevleigh for helping me with the issue here: https://github.com/microsoft/MixedRealityToolkit-Unity/issues/7850
While this did work for my example project, I couldn't get the main project to work with the same settings even after installing all the plugins.
Finally, I had to switch to Unity 2018 and I got the project working with MRTK 2.3 and the default configuration. So, while Unity 2019 did not work for me you can get it working with XR SDK as suggested in the above link.
I have recently purchased a new computer :
https://pcandparts.com/hp-gaming-laptop-power-15-cb005ne-acid-green/
And I have installed Unity on it. The problem is that every time I try to move the layouts unity screen becomes black and freezes.
https://gyazo.com/fdbbba09cbee7005b1b55637cea19855
After trying to search online, post on multiple forums, contacting Unity Team and Nvidia support I have found that the only solution is running unity from the command using -force-glcore
In fact, Nvidia support have been unable to find the solution and Unity Team was stressing that I update my graphic drivers and Directx version however both of them are latest versions and so is Windows 10.
https://gyazo.com/ee810b7d18be7580d3703e42c7b46526
https://gyazo.com/ea751dcf39ebfc791f2dce59cd219a0c
https://gyazo.com/d43718cc693a7c9cac44f4924bf968da
This is the latest email I have received from Unity Team :
https://gyazo.com/df2d7e960484cd8c682e5b733860e935
as well as the link for unity forum that also includes the error and editor logs:
https://answers.unity.com/questions/1511848/unity-editor-crash-2.html?childToView=1512288#comment-1512288
I also tried multiple versions of Unity from 2017 and 2018.
I will provide any additional information needed.
How can I solve the problem so I can run Unity smoothly and without having to go to force-glcore ? And what is truly happening?
Thank you in advance for taking the time to read
How can I solve the problem so I can run Unity smoothly and without
having to go to force-glcore ?
There are other things you can try.
1.Use DirectX9 instead of DirectX11. The default on the new Unity verions is DirectX11.
Go to File ---> Build Settings... ---> Player Settings ---> Other Settings then un-check the Auto Graphics API for Windows API checkbox. From there drag and move Direct3D9 above Direct3D11 click Apply and restart the Unity Editor.
If Direct3D9 is already above Direct3D11 then move Direct3D11 above Direct3D9.
2.If #1 doesn't work, try other graphics API. You can try other 3 graphics API such as Direct3D9, Vulkan and OpenGL Core.
These are hidden and you must click the + sign then manually add them to the menu then drag them to the top.
Then you can chose which one to add and then drag them to the top to force Unity use them.
Once of these should work for you. Chose the one that works the most. You don't need -force-glcore when opening Unity to use this. Your issue is mostly a driver issue sometimes, a workaround like this solves it until the driver is fixed.