Unity Read USB HidDevice Data - unity3d

Excuse me, can I read the Raw data of USB HID Sensor device on Unity?
I have seen relevant documents from Microsoft, but it seems to be only applicable to UWP projects.
https://learn.microsoft.com/en-us/uwp/api/windows.devices.humaninterfacedevice.hiddevice?view=winrt-22621

Related

Hololens 2 audio stream from Desktop

I'm currently developing an app for the HoloLens 2 that needs to stream audio from a desktop PC.
The idea is to send control information (position, orientation, etc.) to a Cycling'74 Max/Msp application running on a Windows 10 computer to process audio for 3D audio playback. I now need to somehow stream the resulting sound to the Unity app running on the HoloLens. Both devices run on the same network.
At the moment I've achieved something using mrtk webrtc for unity in combination with a virtual cable as input. My issue is that this seems to be optimized for microphone use as it applies some options like noise reduction and smaller bandwidth. I can't find a way to set the options for webrtc to stream what I need (music) with better quality.
Does anyone know how to change that on mrtk webrtc or has a better solution for the audio streaming to the hololens?
WebRTC project for Mixed Reality is deprecated and it is designed for real-time communication. If your requirement is media consumption, you need other workaround solutions.
For dedicated media streaming, you can set up a DLNA server on your PC for media access.
You may also set up Samba or NFS on your PC if you need to access files in other formats.

How to transimit video streaming from HoloLens (C#) to PC (python) by socket communication?

I can realize sending data from HoloLens (using Unity coding by C#) to PC (coding by C#) by socket communication. But how to sending video steaming in real-time (the video starts to be recorded when I open the application in HoloLens) from HoloLens to PC by my original socket frame. In my view, maybe I should add some sentences to recognize the HoloLens camera, record video and encode the video to data, then transmit the data by my previous socket. Is it right and how to realize it?
By the way, I hope that the PC can receive the video by python so that I can process the video in the following steps.
To send video steaming in real-time between HoloLens and PC client, WebRTC should can meets your needs. Please check out this MixedReality-WebRTC project, it can help you to integrate peer-to-peer real-time audio and video communication into your application. It also implements local video capture you need and encapsulation it as a Unity3D component for rapid prototyping and integration.
You can read its official documentation via this link: MixedReality-WebRTC 1.0.0 documentation.
Moreover, this project can be used in desktop applications or even other non-mixed reality applications, which can save your development costs.

Alexa Raspberry Pi

I have a Raspberry PI model B+ and I was thinking of integrating it to Alexa Voice Service. So I was able to manage my Raspberry PI and Alexa Voice Service until the part that Alexa says hello. In order to achieve this I used also PC108 media USB external sound card. So I’m getting both input and output from my plug-in microphone or my mini jack audio output to speaker. The thing is that something is missing in order to work .What do I have to do in order to make Alexa listen ?
Thank you in advance.
At re:Invent 2016 they had a workshop on doing this. Take a look at the slides from the session and the workshop instructions. We used a simple USB microphone and sound is built into the Pi. The sample app is still being updated so it should be good to go.
This was with a Pi3 but the basics should still be the same.
You can also use PiCroft that is an image of Mycroft a open source assistant it's just burn it on a sdcard and use
https://mycroft.ai/mycroft-now-available-raspberry-pi-image/
if you want to create skills https://docs.mycroft.ai/skill.creation

Does VR headsets using smartphones have its own sensors?

As far as i know, Samsung Gear VR is the only VR headset that has its own head-tracking sensors and sends sensor data to its mounted smartphone(galaxy series).
Am i getting it right?
From technical specifications of ZEISS VR One, it says
Tracking sensors : Internal tracking by smartphone sensors
Does this mean it has no in-built sensors?
Just like Google Cardboard, all the other low-cost VR headsets are dependent FULLY on smartphone sensors, right?
Cardboard-like headsets are just a way for your eyes to look at your phone in stereo mode, they don't provide any sensors (and very often not even an interaction button like the original google cardboard), these devices typically use your phone's gyroscope to track your orientation.

Take pictures/video with my goPro from the command line

I'm looking for a way to capture pictures or video from my entry-level GoPro using the command line and with the camera connected via USB. There're lots tutorials of how to do it with the goPros that have Wi-Fi. Is there any way to do it like we do with webcams?
None of the GoPros support live streaming via USB. USB is primarily used for charging camera and access SD card. You simply cannot live stream from entry-level GoPro.