Camera bt656 without i2c. How capture? - linux-device-driver

I have camera (camera + videoprocessor TW8834) sending video data over bt656 protocol, but the camera interface have not i2c on cable, so, it isn't recognizing by linux video driver.Is there some way to capture video from my camera in linux without modifieng the video driver?
I got advice to make a fake probe in the existing video driver (sun6i), but i don't think that it is the optimal way. May be i can write some driver or something simmilar that will snap up i2c messaging with my camera? Can i use it with default video driver by this way?
May be, are there some other way?
What should i learn to solve my problem?

I solve my problem by driver and DeviceTree modification. OS Armbian, easy to build, easy to make and apply patches. Big community. Some course about kernel hacking, which was usefull for me: https://courses.linuxchix.org/kernel-hacking-2002.html
For create solution we with my collegue spend about 4 days, we both was non-experienced in kernel modification earlier. I hope, this information will help someone with similar problem.

Related

How to do High End video encoding on BeagleboneBlak

As we know BeagleBone Black dont have a DSP on SoC specific for the Video processing but is there any way we can achieve that by adding some extra DSP board.
I mean like Raspberry got Video Processing, so anyone tried to integrate both to get, so we have both the things to make that work.
I know its not the optimal way and these both are different but i have only one BBB and one Raspberry and I am trying to achieve some 1080p video streaming with better quality.
There is no DSP on BeagleBoneBlack, you need to use DSP functions.
If your input is audio, you can use ALSA.
When you say "dont have a DSP on SoC specific for the Video processing" - I think you mean what is usually called a VPU (Video Processing Unit), and indeed Beaglebone Black's AM3358 processor doesn't have it (source: http://www.ti.com/lit/gpn/am3358)
x264 has ARM NEON optimizations, so it can encode video reasonably well in software, 640x480#30fps should be fine, but 1920x1080#30fps is likely out of reach (you may get 8-10fps).
On Raspberry Pi, you can use gstreamer with omxh264enc to take advantage of the onboard VPU to encode video. I think it is a bit rough (not as solid as raspivid etc) but this should get you started: https://blankstechblog.wordpress.com/2015/01/25/hardware-video-encoding-progess-with-the-raspberry-pi/

How to access Iphone camera using MATLAB Mobile

I have been searching for how to access the iphone camera using MATLAB and I have found it can be accessed by app called IP CAM with the use of a local network. Yet the solution of IP Cam program existed on apple store isn't working so well for my application because I'm trying to build a real time image capturing program using Iphone's camera and Matlab mobile with later processing (and this method keeps MATLAB busy as long as it display the scene and I still want MATLAB to run in the foreground instead of IP Cam).
So far I've downloaded MATLAB Mobile and the connector and connected the Iphone to MATLAB on my laptop, so is there any one who knows how to access the Iphone's camera on MATLAB Mobile and capture the image so that it can be stored on MATLAB workspace for later processing ? or is there any one who can suggest tutorials any piece of material helping me through this.I'd appreciate your answer very much and thank you in advance.
P.S: if there is a solution on android's devices it's also work for me.
I am not aware of iPhone based solutions but here is an opportunity for android based system. I will suggest you to used Sensor EX. I have used it for sometime to acquire Accelerometer, Gyroscope data set along with live images. This tool has bindings available for MATLAB beside other programming environments. Feel free to ask questions if you cannot figure out that how this system works.

Stream screen from Mac to iPhone

I want to create a program that stream the screen of my Mac to my iPhone. Kind of like it is done in Liveview. I'm still relatively new to Objective-C, so I don't know where to start to make such an application.
It seems you have to have something installed both on your Mac and on your iPhone, but how would you actually stream the screen of your Mac to your iPhone?
Hope someone can point me in the right direction.
Update of question
Thanks for the answers. Still seems a bit vague to me and I'm not sure I really need full video streaming. Implementing also seems to be a pain, since there aren't any real good resources for it.
Taking a screenshot every second or so and streaming it to my iPhone as an image, would actually be ok. I've figured out how to stream an image with Bonjour from my Mac to my iPhone.
The screenshot I need to send to my iPhone is of the design that I'm currently working on in photoshop. I've figured out how to take a screenshot and how to get a list of all open windows. But how to make a snapshot of an open PSD-file, I don't know.
Any suggestions on that?
It's a very big subject, so not really something that can be tackled with a simple response. However, I would suggest that one approach would be to write a VNC client for the iPhone. Indeed, this open source exists that's probably worth a look:
http://code.google.com/p/vnsea/
Tim
I would go with the frequent screenshot approach. You would prepare a screenshot of the item you want to transmit and then use some easy library like my DTBonjour to transmit these objects via WiFi to iOS clients.
https://www.cocoanetics.com/2012/11/and-bonjour-to-you-too/
If you were using layer-backing then you could also use the renderLayer... methods which would also include sub-layers.
The most fidelity you'd get from encoding the individual screen shots in a streaming video format, though this is way more work.
This is called RFB (or RDP), and most remote-screen applications use RFB/RDP protocol and libraries which implement it.

What Webcam is Compatible with MATLAB

I am working on a project of mine which requires a webcam and MATLAB. I have a Logitech Webcam, and I dont know if I could talk to it through MATLAB because Im trying to work with images and image processing, so I just want to know if there is a way to find out if the webcam is combatible with matlab or if I need to get some other type of webcam to get the job done, if I do need to get another one, it would be helpful if you suggest a certain cam that is cheap and available around.
Thank you.
I've used several Logitech webcams with the Image Acquisition Toolbox on Windows. You'll find a list of supported hardware here.

Getting live data from the iPhone on the Mac

I'm experimenting with the sensors of the iPhone. At the moment I visualize the data of the accelerometer by logging it to a file, downloading the file via Xcode and loading the file into my Ruby-Processing app. As I'm playing around a lot this process is getting annoying. It would be cool to have a live stream of the data on the Mac and see what happens in real time.
So what would be the easiest / fastest way to build a prototype for playing around with the data in real time?
I imagine, I could just send the data over the network and have a Cocoa / OpenGL app read an visualize it. What do you think? Could you give me some clues on how to do that? Thanks!
Take a look at NSNetServices and bonjour. There's some good examples and using bonjour greatly simplifies the discovery and communication process. The hardest part is the socket fu you use to set up your "server," but the examples provide you with all the code you need.