Hololens - create app on desktop - unity3d

I can't figure out the wording to research the following idea.
With the Holograms app, I can set the hologram and see the other apps front window or use web browser at the same time.
How can I create an app that does not occupy the whole system but only run on the desktop with browser and others?
EDIT: I am trying to run an hologram within the shell.
https://developer.microsoft.com/en-us/windows/holographic/hololens_shell_overview
Mostly, this is for 2D apps, but the Holograms app runs 3D holograms, so is it possible to duplicate this?

3D Hololens apps do not currently support running side by side with other applications. As of the May release you are able to run multiple "flat apps" (UWP apps) side by side:
https://developer.microsoft.com/en-us/windows/holographic/release_notes_-_may_2016
Currently there has not been any announcement about running multiple 3d apps side by side. I optimistically hope that this is coming in a future OS release.

If I understand you correctly, you are interested in creating an application for the Hololens that rather than having it "appear in the world around you", would run in a Window (Desktop) that would appear in the spatial world (like Edge). If this is correct, any UWP application can run Windowed and be placed in you virtual space. You don't need a specific Hololens project for that. If however you want to pull in spatial data, you would need to wire-up the required events so that spatial data callbacks handle the incoming data. I'm not 100% sure Unity is a best fit for this problem (at least currently) since they are focusing attention on Spatial hologram applications rather than desktop UWP applications that consume spatial data.

Just to note, it is worth stating that a single 3D app (ie Unity developed) takes over the whole HoloLens device and experience. If you bloom (windows key in emulator) you can change to the UWP app view but you can't see other apps at the same time.

Related

Is it possible to stream in Unity?

I am working in a vr game for Android in unity and it has a quiet large map and many objects to render and my phone doesn't has enough power to run my game so i was thinking that is it possible to stream my game made for desktop pc to my android device inside unity without using any other external software. And also i need to use my phone's gyroscope to control the game.
What you are talking about is called remote rendering. As you said you don't want to use any other external software you will have to decide and implement everything by yourself. The trickiest part in this will be the streaming of the game view to the phone: there are implementations that send every single frame over a TCP connection. If you want to use something external you could investigate into WebRTC implementations like this.
By the way, what you are trying to achieve is already available as a product, e.g. here

Is it possible to develop a virtual world that is accessible via both a PC and a virtual reality headset?

For example, if you have a virtual reality headset, you can interact with this virtual world in VR (i.e. WebVR); however, if you don't have a VR headset and/or WebVR compatibility can still access and explore this virtual world (i.e. like Runescape) and interact with characters, whether they are VR or web in the same virtual world?
The A-Frame framework handles that for you automatically, or you can roll it yourself, if you're using another framework. Either way, the different control schemes require a fair amount of thought.
You could also take a look at React360 (https://facebook.github.io/react-360/), it is a WebVR based framework from Facebook and can handle most 3D media out of the box. It performs quite well and has the advantage of progressive enhancement, i.e. if you view it from a desktop/tablet you get a 2D experience, if you are on a 3D capable device you get a full VR experience.
It's also cross platform so will run on android/iOS/Windows/MacOS/Oculus/Vive. Samples are included with it which should be enough to give you an idea of its capabilities.
Depending on the complexity of the game you are trying to develop and the graphical control required A-Frame is another option to look at.

HoloLens sharing with UWP application for PC

The idea is to create a HoloLens application that displays a hologram which can then be manipulated through the UWP application run on the desktop. The desktop application would contain various UI elements that manipulate the hologram(ie. a rotation button to turn it 45 degrees) and of course, see the same object as in the HoloLens. Naturally, I arrived at the 240 academy tutorial, but that seems a bit outdated compared to the current version of the Holotoolkit. It also doesn’t really fit my scenario, since I am not sharing between two HoloLens devices, but a desktop and a HoloLens. I figured that that shouldn’t really matter since you are still targeting for the UWP, but I wasn’t sure.
So what I tried so far was trying to edit the example scene “SharingSpawnTest” and target it for the PC to see what would happen, but I thought this wasn’t the way to do it since the project settings are set for Mixed Reality and not a regular UWP application for the desktop.
My question is basically if this is even possible and if so, how do I achieve this? Do I have to create two separate projects, one for the desktop and one specifically for the HoloLens and communicate that way?
Add a NetworkManagerHUD component to your NetworkManager object. Once you add it, type in the HoloLens IP address and connect to it. Currently, this only works for HoloLens to be the host and the desktop to be the client.

Oculus Rift - Multi Desktop Application

I am currently thinking about building a Multi Desktop Application for Oculus Rift. The idea is, that the user can use multiple screens like in a regular computer, but can see three windows (left, center, right) when moving their head. (So far so good)
But this is where it gets tricky: On the three windows, I would like to use three different browser tabs or applications that the user can watch and use simultaneously with mouse and keyboard.
Can anybody please suggest how to start the whole thing or is there a framework that I can use? UE4 or Enity3d will be able to give me multiple screens, but bringing the content/apps/browsers to them is what I cant figure out...
Thanks for any help!
fj
You can install VNC Server in the machine and render 3 VNC Clients. This is doable using, VNC Java client and render its output into a texture.
Using Open Wonderland it is possible since it has Oculus support
Regards

Google Cardboard controlling PC

When I saw the Google Cardboard for Unity, I assumed this meant that you would be able to make a Unity PC game and use your phone as a screen/controller. All I can see is it wanting me to make an android app which is all well and good, but it doesn't allow for input from the keyboard.
Is there a way to stream the Unity PC project to the device and retrieve input (i.e. Headtracking, NFC magnet)?
The problem with such a solution is latency. In VR latency is a big deal. The overall latency from input to photons reaching your eyes should be 20ms or lower. Regular games have 30-60 ms latency by themselves. Add to that the gyro latency, the phone display latency... If you want to add another 25ms or more ping to your VR experience, that's gonna be painful and may even make you sick. If you want to read more on why latency is such a big deal in VR, Michael Abrash wrote an excellend blogpost about it: post on latency
If you want to necessairly use a keyboard for navigation, consider using a bluetooth keyboard that can be used with android devices. Also keep in mind that with the current technology, especially without a dedicated headset, really dynamic vr experiences probably won't work very well and can make some people uncomfortable or sick. For a good read on designign virtual reality experiences, please refer to this guide from the Oculus Rift: http://static.oculus.com/sdk-downloads/documents/Oculus_Best_Practices_Guide.pdf
There's nothing in the Cardboard SDK for talking with a PC-hosted Unity game. You could adapt the code from the Unity Remote 4 project:
https://www.assetstore.unity3d.com/en/#!/content/18106
We are developing the app what you want except it uses GearVR instead of Cardboard. Please check the link below.
http://challengepost.com/software/airvr
Streaming from your PC to your phone's Cardboard is possible using third-party apps, such as Trinus VR (the client app on your phone) and Vireio (the streaming app on your computer). The two apps will then communicate via your home network (Wi-Fi or other) to stream the images.