trying to integrate vuforia 4.2 and cardboard in Unity 5.1. Everything works except for field of view. Can not modify vuforia camera, it breaks matching between 3d object and videobackground, so I need to scale it in cardboard somewhere. Any ideas?
You need to adjust the field of view of the AR cameras, the problem if you match the cardboard size, it will zoom in too much for your brain to adapt, so you get dizzy and headaches. You need to play with those fields in order to get the best of all.
Related
I'm trying to get an object to show up with Vuforia in Unity by using a multi target object.
So far this works, but only on my Android phone, but not on the AR glasses.
On the AR glasses, it shows up as this:
It's flickering a lot.
I tried changing camera clipping plane values, different shaders, etc. - the issue goes away if I use ZTest Always in my shader (but ZTest Always causes other issues so I can't use that).
So what might be the difference in rendering here which causes the flickering on the glasses, but not on the phone, when both use Android?
Maybe you are using a renderer that isnt native to your glasses, you should check what renderer your glasses use and use that instead
I am playing video over image target using vuforia plugin in unity3d. It is simple green screen video. I am using shader to remove it.
Sometimes when video played over image target it becomes too shaky and jittery and thus reducing the AR experience. How can I avoid or reduce it.
I tried multiple ways to get rid of it but no success. Here is what I tried:
Previously I was embedding the video in a Plane then I used Quad
but no success.
I tried to change AR camera (World Center Mode) to different
values like FIRST_TARGET,CAMERA and SPECIFIC_TARGET but still same problem.
Also my vuforia target image has 5 rating in vuforia database.
What could be the solution to this problem. Any help would be highly appreciated. Thanks!
I'd like to use Vive front camera for traking. but I'd don't know how to apply vive texture to ARcontroller in Vuforia.
So, I though that if there is a method to match video texture to ARcontroller, Vive texture could be matched also.
I use Unity program.
Do you know how to do this?
Front facing camera support was added to SteamVR - you can also see:
https://github.com/dariol/ViveTrackedCameraUnity and related blog for info on getting to the texture which you can pass into Vuforia.
I'm developing a VR app in Unity for the Samsung Gear VR and I'm trying to implement a pointer so the user can interact with the objects in the scene. When you look at distant objects it looks fine, but when you focus on close objects (which is highly needed for the app mechanics) the pointer appears to be duplicated, so you need to center the desired object in the middle of the points :P
What I've tried
-Using the GvrReticlePointer that comes with the GoogleVR package for cardboard
-Creating my own pointer by adding a canvas to the main camera with an image in the center
-Changing some of the Camera settings like field of view, stereo separation, etc.
-Configure my phone via a QR code http://imgur.com/fVrNrQk
Steps to reproduce (With canvas added to camera)
1.- Create a simple scene with a few objects to look at in Unity
2.- Set build settings for android
3.- Configure player settings to enable "Virtiual Reality Supported"
4.- Add Oculus as Virtual Reality SDK
5.- Set package name and minimum API level
6.- Add a canvas to the camera
7.- Add an image to the canvas, a cross will do the job
Observations
I'm using Unity 5.6.0b10 since google cardboard's site recommends using this version for the GoogleVR package. And I'm using the Samsung Gear VR with a Samsung Galaxy S6 edge + phone.
Solved
Apparently this is a well documented issue called voluntary Diplopia, and it's a human bug not a software one (read here, Unity's documentation, section The Reticle Interaction in VR).
The problem is trying to put the reticle at a fixed point in the user interface, like traditional 3D games. When looking at closer objects in VR this is going to cause this seeing double problem.
The solution is to position the reticle at the point in the 3D space the user is looking at. If he's looking closer, the reticle is drawn closer. Of course now you also have to scale the reticle accordingly, so the users can see it the same size no matter where they're looking at.
Unity also provides some example scripts about this, you can find them in the assets store, is called VR Samples.
Now I have performance issues (I'm working on mobile platforms): sometimes, when you turn your head fast you can see the reticle where it was drawn before. But looks way better than the double reticle version.
I am trying to wrap my head around Vuforia's capabilities. I want to make an app which lets me place a 3D object into a camera view and have that 3D object stick to the world. I've been learning how to use Vuforia in Unity3D, and Vuforia seems to be slightly capable of this, but is severely limited by its craving for "Targets". It doesn't seem to be able to do much if I don't give it some sort of target.
One workaround I've found is to set the ARCamera's World Center Mode to DEVICE_TRACKING. This seems to let me place a 3D object into the world (in Unity) and have this object overlay into the camera feed, almost making it seem like it's anhcored to the real world. This doesn't work perfectly though: it tracks properly when I angle the device up/down/left/right (rotation), but it does not seem to track the device's translational motion; that is, when I move the device forward/back/left/right, the overlaid object doesn't get closer/farther nor does it rotate as I move around it.
Is it possible to get this sort of tracking out of Vuforia, or am I better off switching to something like Google Tango?
The difficulty with setting World Center Mode to CAMERA in Vuforia is that apparently 3D objects rotate around the camera based on its accelerometer/gyroscope changes. This doesn't allow for objects to be anchored to the environment. Instead they follow with the camera.
Kudan is a good markerless tracking option.