Multiple Cameras in SteamVR HTC vive - unity3d

I need to have multiple cameras in a Unity3d scene. However every time i create a camera, SteamVR automatically overrides it so that it is moving with the helmet. I simply need a normal camera without steamVR taking over it. How can that be done?

Note, assuming Unity here:
Simply, don't make it the main camera. You can also disable VR support in project player settings and use the SteamVR plugin's CameraRig.

Related

Unity 3D Shader Doesn't appear on VR mode

We have two scene, in one scene we create level design and mechanics for VR in runtime, and other scene we use SteamVR for VR mode. Shader show up in editor scene but doesn't work in VR scene.
Thank you for your answers,
Actually it was a save load problem. We use 3th party asset from unity asset store (Runtime Save & Load) and SceneAssetLibrary.asset that hold scene asset references was overwritten when worked on scene. So we revert that file thus it worked again.

How to change the AR Foundation Camera Rotation and position using joystick in Unity3d

I am creating a project like AR Portal in the AR foundation.
there are two modes in this:
AR Mode
Joystick Mode
Everything is going well with AR Camera and in AR mode.
In joystick mode, I am unable to rotate the camera and I am stuck on that problem and didn't find any solution.
If I disabled the Tracked pose Driver of AR Camera, then it will work but if I enable the Tracked Pose Driver, then it will not work.
Is it possible to rotate the camera when Tracked Pose Driver is enabled?
I'm not sure what you're trying to do there. AR camera is usually a physical camera, changing its position and rotation by code would not make a lot of sense. If you are using the joystick for a different scene/ part of the scene, try using a regular Unity camera and write a simple controller to move it around.

Vuforia with Video

I'd like to use Vive front camera for traking. but I'd don't know how to apply vive texture to ARcontroller in Vuforia.
So, I though that if there is a method to match video texture to ARcontroller, Vive texture could be matched also.
I use Unity program.
Do you know how to do this?
Front facing camera support was added to SteamVR - you can also see:
https://github.com/dariol/ViveTrackedCameraUnity and related blog for info on getting to the texture which you can pass into Vuforia.

Unity UI canvas not working with VR

I have been trying to get a very simple demo of a native Unity UI canvas working with VR.
I have read the oculus blog post here: https://developer3.oculus.com/blog/unitys-ui-system-in-vr/ but i need to use the native unity UI as i want to redistribute the code without license worries.I followed this tutorial https://unity3d.com/learn/tutorials/topics/virtual-reality/interaction-vr?playlist=22946 and downloaded the unity vr samples project from the asset store. In this they provide some scripts to place on the camera (VRInput and VREyeRaycaster) and some scripts to place on the target object (VRInteractiveItem and ExampleInteractiveItem).
When i apply the target scripts to a regular GameObject in the scene (e.g. a cube) the raycast works fine and the appropriate calls are made when fire1 is activated. When i try and do this for a canvas object (e.g. a button) - no hit is detected. I have tried placing the two target scripts (VRInteractiveItem and ExampleInteractiveItem) on the canvas, the image containing the button and the button itself and none work. What am i doing wrong? Why would it work on a regular gameobject and not on a UI canvas? I have made sure all my canvas elements have their raycast target boolean property ticked
EDIT:
It seems to work when i attach a box collider to the UI element, is this required? i thought it should just work with a GraphicsRaycaster attached. but the configuration below doesn't work (when box collider is disabled and graphics raycaster is enabled)
This is what is on my players camera:
I dont have a problem using box colliders if i have to but i wanted to take advantage of the UI buttons changes in highlighted and pressed color properties
In Unity raycasting works only with game objects having colliders. Raycast returns true when it hits a collider. Without colliders there is nothing the ray can hit.
Unity Physics.Raycast documentation
I believe, for anyone just seeing this for the first time, a potential reason it is not working is because the canvas from the above picture is using a "Graphics Raycaster" element and not an "OVR Raycaster" element. The OVR Raycaster is meant to replace the graphics raycaster to connect Oculus to Unity UI.
If you want to use the unity's UI in VR you might want to take a look at this asset: VRTK
There are some examples of VR UI using controllers or camera targeting.
Go to your canvas, you should have an option that is "Plane Distance" it's set to 100 , I change it to 0.5 and it works quite well.

Handling GearVR Touchpad input?

When I add an OVRPlayerController into a Unity3d scene and build and run the scene for the GearVR the built-in touchpad spins the camera around the vertical axis, which is redundant with head tracking. What do I need to change so that the touchpad instead allows the camera to move forward and backward, as if walking? Is there a thorough tutorial?
The Oculus SDK 0.4.3 comes with the support for the GearVR Samsung GamePad.
All you need to do:
import the SDK.
overwrite the projectsettings folder of your project with the one that comes with the SDK.
add the overplayercontroller to your scene.
add a gameobject below the overplayercontroller e.g. plane, quad... this will act like the ground (keep the player from falling)
add a collider to the gameobject e.g. mesh collider
then once you run it you will see that you can move around using the gamepad as well as turning the came around the vertical axis...
basically use any first person shooter tutorial for Unity3D and because the Oculus SDK comes with support for the gamepad you can quickly do this...
this link might help
https://www.youtube.com/watch?v=mbm9lPB5GPw