I am not familiar with Unity , I need to know if I can import my real environment pictures for my apartment to Unity and using these Pictures as Virtual environment (Scene).
I am looking forward for your answers.
Unity, AR, VR
Related
I have a question about Unity VR development for Oculus Quest 2.
Is it possible to have a GUI to import a local (stored in Quest2) modelfile and then assign it to an new gameobject during runtime?
Thanks in advance.
Depends on what you mean by "local". If you're shipping the model with your Unity project, you could use Resources to load and instantiate a model from the Resources folder (and its sub-directories).
If you're adding a model outside of the Unity application, you can use the Application.persistentDataPath property to access a file directory in the local storage of your Quest 2 device.
After many months, I made a multiplayer game in Unity 3d using Photon 2. How can I run the game on a second computer? If, for example, I want to play with a friend
In Unity, go into file -> build and this should create an executable file that you can then run on the other computer.
My multiplayer VR project is based on Unity3D, Oculus Integration v1.39 and PUN2 running on Oculus Quest. I'm using the standard teleport script provided in the Oculus library. PhotonAvatarView is what I use to keep avatar position/rotation in sync across clients.
Unfortunately, when a player teleports to a new location, the other one doesn't detect any change in the remote avatar. It seems like PhotonAvatarView doesn't see the change in location of the user, which is really strange. How can I fix it?
PhotonAvatarView code is available at this URL: https://doc.photonengine.com/en-us/pun/current/demos-and-tutorials/oculusavatarsdk
i'm developing VR using google cardboard SDK..
i want to move on virtual environment when i walk on real world, like this : https://www.youtube.com/watch?v=sZG5__Z9pzs&feature=youtu.be&t=48
is it possible to make VR application like that for android...? maybe using accelerometer sensor ? how can i implement this using unity...?
i try to record accelerometer sensor while i walk with smartphone, here are the result : https://www.youtube.com/watch?v=ltPwS7-3nOI [i think the accelerometer value is so random -___- ]
Actually it is not possible with only mobile:
You're up against a fundamental limitation of the humble IMU (the primary motion sensor in a smartphone).
I won't go into detail, but basically you need an external reference frame when trying to extract positional data from acceleration data. This is the topic of a lot of research right now, and it's why VR headsets that track position like the Oculus Rift have external tracking cameras.
Unfortunately, what you're trying to do is impossible without using the camera on your phone to track visual features in the scene and use those as the external reference point, which is a hell of a task better suited to a lab full of computer vision experts.
One another possible but difficult way is:
This may be possible if you connect device to internet then watch it's position from satelite(google maps or something like that)but that is a very hard thing to do.
I would like to know if It is possible to make a 3D model of an entire building on my college campus. If I am able to make the 3D model of each room, and then somehow combine all the rooms to make a full 3D building it would be a great project for my senior internship. Please direct me to the correct information. Or please give me instructions on how to use the Project Tango Device to create a full 3D building. Ultimately, I want to use the Project Tango Device to conduct indoor mapping using augmented reality.
Project Tango can export your scanned meshes to .obj files. Programs like 3DSMax allows you to import several .obj-files.
To create a mesh of a room with project tango and export the files you can use the constructor app.