Kinect not being detected on other computers after deployment of UWP - deployment

I recently finished to develop a UWP based on the SDK example CameraFrames. On the second screen of this example (and also on my app) you can choose to preview the frames taken by the Kinect v2.0 as shown below.
On the top of the image on the right of "Grupo de fontes" (Source group) I am able to choose between the different cameras that are connected to my PC. The problem comes when I deploy my app to another PC, I am unable to see "Kinect V2 Video Sensor". Thus rendering my app obsolete as it needs to be portable between PCs. I have checked that inside "Package.appxmanifest->Capabilities->Web Cam" checkbox is ticked.
I am out of ideas as I don't have a clue why my app works flawlesly on my computer but not on the rest. If you need any extra info to help me with my problem please let me know.
It is worth noting that on the other PCs that I've tried my app can read frames via Kinect Studio or MatLab.
Inside "Camera privacy settings" my app has privileges to use the camera.
Thank you in advance.
Update due to Nico Zhu's comment
My OS is Windows 10 and my target version is "10.0.16299.0". I have downloading and deploying CameraFrames on the second computer that I'm working with but it doesn't recognize the Kinect as in input source. Even though CameraFrames doesn't work doesn't read anything, I can properly make use of the kinect using Kinect Studio.
It seems that my problem is focused on my second computer not being able to make use of the Kinect on any deployed UWPs. What should I installed in order to make sure that I have all that's needed to properly read from the Kinect?

At the end of a lot of headaches I just needed to upgrade my drivers.

Related

Unreal Engine 4 Viewport is Black/Blue

enter image description here
enter image description here
Please see the pictures in the link above.
I'm new to unreal engine. I've been trying to get Unreal Engine 4 to work on my linux machine ( 5GB RAM & 500GB Harddisk) but I couldn't get the viewport to work. It always shows blue or black screen.
Please I need assistance. Thanks in advance
I am also a UE4 user / game dev and use *nix. Usually the behavior you describe happens when there is a video display issue. In Window this is usually the Nvidia Game Ready driver, sometimes new updates break things, there is a common fix involving a registry edit called mpo_disable.reg - anyway, you are not using Windows, just for anyone else who lands here via Google-Fu.
In your case I would say it could be related to the video drivers. Which distro / kernel and video driver are you running? Nvidia card?

How to support OTA video in custom web app (for LG TV)?

I've searched considerably for this answer (including all WebOS, Enyo, LG specific documentation, and countless Google searches), but have turned up nothing.
In short, the goal is to develop a WebOS application that can resize an existing video stream (coming from coax or HDMI), unfortunately ...
Update: This is a web-based app that resides on a server.
After specifying the IP address (via hidden config menu), the TV saves and exits, without any channel scanning, or additional setup. Upon reboot, it displays the index.html page at the specified IP address.
So, unless there's a way to explicitly configure OTA sources within the web application, I'm assuming that any OTA video will have to be provided via streaming.
WebOS offers a way to specify media sources from files -- so, is there any way to point to a video file that's perpetually streaming an OTA source?
Is Plex, or Sling capable of this?
Any direction that can be offered is greatly appreciated.
original answer: I'm not sure this is going to be possible. I believe only system apps are going to have access to the video pipeline.
new answer:
If you can configure your OTA stream to go over HLS, then you can use an m3u8 playlist for your <source> and it should "just work" on a webOS TV.
It's been a while since I posted this, and I've learned a great deal about this topic since then. So, for anyone who may run across this, with the same question -- here's the answer ...
When it comes to LG Smart-TVs their are two configuration options specific to commercial TVs (Lodging and Hospital models) ...
"Pro:Centric Wizard" and "EZ-Manager"
The Pro:Centric Wizard is intended for hospitals and won't automatically import TV configurations from a server, but it can scan for channels.
The EZ-Manager option is intended for lodging/hotels and provides a great deal more control, but the TV configuration and channel listing must all be obtained from a server (or usb drive).
Modification of the TV settings and channel listing is performed by using the FTG File Manager.
Downloadable Stand-alone Application:
http://www.lg.com/us/business/display-solutions/ftg
Online Version:
http://cct.procentric.tv/ftg_manager
Channels must be either manually programmed within the FTG File Manager, or imported from a TV clone file (created after the TV has performed a channel scan).
In order to create a clone file of scanned channels ...
Initiate the "Pro:Centric Wizard" setup process
Scan the channels
Complete Setup
Once the TV restarts, hit "Menu"
Highlight the "General" or "Settings" icon (should be "Settings") and hit "7" seven times (or until clone menu appears).
The option to export to USB will only appear after pressing "9876" on the remote.
Once the file is exported to USB, it can then be opened within the FTG File Manager and edited for server delivery through the "EZ-Manager" setup.
Hopefully this will help someone. :)

Disabling Microsoft PixelSense (aka Surface) Surface Shell Timeouts

I am writing a game for the Microsoft PixelSense written in Unity, communicating with the table through the SurfaceToTUIO Bridge and the unity3d-tuio Unity Plugin.
I am currently trying to get the game to play nicely with the Microsoft PixelSense launcher. I have been able to get my application to appear in the launcher by mimicking the Surface Bing Application - duplicating a link to an XML file in C:\ProgramData\Microsoft\Surface\v2.0\Programs and creating the corresponding XML in the proper format.
When I go into Surface Mode - either through a dedicated Surface User Account, or through the Surface Shell on the Administrator's Profile, the game appears correctly on the launcher bar with the custom icon I set in the XML. The game launches correctly from the bar, and I can play it without any errors for about two minutes.
At that point, the Launcher closes my game. With a little research, I learned that its the Application's Responsibility to dismiss the Launcher.
Being that this is part of the Microsoft PixelSense SDK and not accessible to Unity, I've tried various methods to get around this:
I tried running the game in Single Application Mode. It turns out there is a different timeout that still waits for the SignalApplicationLoadComplete call.
I tried using the CriticalProcessMonitoring Disable and ApplicationProcessMonitoring Disable keys in the Registry.
I tried setting the LoadingScreenTimeout and SingleAppLoadingScreenTimeout Registry Keys to 0 - below their valid threshold.
What I want is to provide my users with the smoothest experience getting into and out of the Unity game. Currently we have to launch our game from the Windows Desktop, which can frustrate users due to the fact that Windows can't differientiate between a finger touching the screen, and a palm hovering above the screen.
Does anyone have a workaround, an good understanding of how I could spoof a SignalApplicationLoadingCall from Unity, or a suggestion of something to try?
Thanks!
Update: I've found some more things that didn't work:
I found the Microsoft.Surface DLL at C:\Windows\Microsoft.NET\assembly\GAC_MSIL\Microsoft.Surface\v4.0_2.0.0.0__31bf3856ad364e35. I imported it into my Unity project, but recieved a System.TypeLoadException that appears to be that the DLL is compiled with .NET 4.0, which Unity does not currently support.
I have been unable to find any COM objects that would allow me to communicate with the launcher without needing the DLL.
I cannot use a XNA wrapper program, as System.Diagnostics.Process.Start doesn't work in Surface Mode according to this post.

iPad/iPhone Screen Mirror

I'm trying to figure out a way to mirror an iPad screen to other iPads. This doesn't seem to be supported on the platform though.
Basically, a teacher would have an iPad, then the students would have iPads and see everything that is happening on the teachers screen, but on their screens.
Thoughts?
I have been attempting to find a solution to this problem myself. I have not found any apps that can mirror exactly what is happening on another IPAD, but some come close.
RabbleBrowser and Ideaflight both had potential. Ideaflight appears to be more for business. RabbleBrowser appears to allow the mirroring, except it only works as a browser and a file/picture mirroring.
Both iPads are linked to the same wifi and when you join a session, they will mirror the iPad that started session. Also allows chat (controlled by session starter).
It does NOT continue to mirror if you move out of browser and into another app however. I had dreams of leading a class through a a lesson on google earth, but no go .:(
Another option is attaching a laptop to a projector. Then you download Airserver on the laptop. Go to the menu bar at bottom of iPad and turn on AirPlay. The laptop will mirror the iPad perfectly and project it! It's wireless and works well. I tried the HDMI connector to laptop but it gives a poor quality, shaky image.
Hope they allow mirroring in future updates. The capability is there, don't know why they don't! Guess trying to sell more appletv!
A similar question was asked on the Apple forum (https://discussions.apple.com/thread/3118281?start=0&tstart=0), and the following app seemed to help them or answer their question.
Have a look at Replicate Pro on the app store:
http://itunes.apple.com/us/app/replicate-pro-for-ipad/id363286515?mt=8
One feature listed in the notes:
Share files between two iPads/iPhones that are running this app. (Pro
version only)
I'm not sure if this will cover multiple devices or simply between two, but it may be worth a look. Sadly, the only way to try would be to spend $5.99.
You'll need to create an application for the student iPad that emulates the screen of the teachers iPad. I would suggest that, although i dont know if its possible, the teacher somehow starts up and app that emulates their entire iPad. Meaning, from within the app named "teacher share" (or whatever it is), they can access the music, settings, notes and other apps found on their ipad. Then that information could be sent over a network to the students.
Nearpod is an app that will allow you to mirror a presentation on several iPads. I have had up to 9 at one time. Through the Nearpod program you can make a presentation similar to PowerPoint, and also incorporate interactive questions, which can be multiple choice, short answer, and even drawings. The only drawback is the full version costs $10/month. The free version is still good, you are just limited on the size of the presentation.
After doing lots of research, I found one app which shares iPhone device into another iPhone device. Really great logic they have applied for screen mirroring.
No idea about detailed how they have implemented but after installing and checking the app I came to know that I think they have used iPhone Screen Recording and broadcasting it on to their server and then on another device they are syncing from the same URL.
OliOli a free and simple screen sharing app for iOS.
iOS App: https://itunes.apple.com/us/app/olioli-screen-sharing/id1382253993?mt=8
WebSite: https://olioli.io/

Profiling a running iOS app without a computer attached

I would like to profile a running app without a computer attached. Pretty much what the sample command does on Mac OS X. Is it possible ?
I would like to do this without any computer attached because my app gets into an endless loop only at certain GPS locations. I tried reproducing the problem by faking GPS fixes but I couldn't. Yet the issue is 100% reproductible in the field…
You can try logging suspected method and loop entry/exits to a file, and recover the file later using iTunes Document sharing or Xcode.
No - I'm pretty sure it isn't, as Apple does not allow developers to use any iOS-development tools on the platform itself. You might be able to build analytical stuff into your own app (RAM monitor etc.), but not by using another app.
Anyway, your computer is much better suited to the task than a device - so sick with it.
Maybe some day Apple will let us test/write iPhone apps from iPads...