It's difficult to tell what is being asked here. This question is ambiguous, vague, incomplete, overly broad, or rhetorical and cannot be reasonably answered in its current form. For help clarifying this question so that it can be reopened, visit the help center.
Closed 11 years ago.
I am interested in creating an auto tune app and did a bit of research and apparently thats now possible by using accelerometer that apple provides but I have no idea how to do that. can someone help out by giving me a link to a tutorial or give me an example code to change the voice inserted into the function.
Thanks in advance.
The accelerometer reports back physical acceleration of the device. This has absolutely nothing to do with digital signal processing of a recorded waveform, and as others said, there will be no easy way to do this. You can browse the audacity source (plugins specifically) to see algorithms having to do with time-stretching and will see that this is very complex. Not only is the science of the signal processing complicated, the art of getting it to sound good (at all) is incredible.
Related
It's difficult to tell what is being asked here. This question is ambiguous, vague, incomplete, overly broad, or rhetorical and cannot be reasonably answered in its current form. For help clarifying this question so that it can be reopened, visit the help center.
Closed 10 years ago.
Using CIFaceFeature we can detect face feature. But I want to detect hand and leg too.
How can we detect these? Is there any other classes or logic using which we can detect these.
OpenCV is 'just beginning'. Recognition is hard problem and still remain unsolved at most part. There is no hand/leg detection in openCV. You have to train the data yourself using machine learning algorithm provide by openCV. You have to collect positive data (human body image for example) and negative data (non human body image) in order to train.
I suggest to look at this link to speed you up:
http://docs.opencv.org/doc/tutorials/objdetect/table_of_content_objdetect/table_of_content_objdetect.html
OpenCV haar training for static image
http://kang.blog.com/2009/08/12/how-to-use-haartraining-in-opencv/
http://www.technolabsz.com/2011/08/how-to-do-opencv-haar-training.html
It's difficult to tell what is being asked here. This question is ambiguous, vague, incomplete, overly broad, or rhetorical and cannot be reasonably answered in its current form. For help clarifying this question so that it can be reopened, visit the help center.
Closed 10 years ago.
I am intending to write a simple OCR engine for iPhone like "Auto Cheat" application which scans game board of "words with friends". How should I go about doing it? Are there any source codes available for me to use and alter it according to my need? I am a newbie to this OCR thingy. Pls help.
Thanks
You need to check out OpenCV for iPhone. This question: Tutorial for iPhone OpenCV on shape recognising, has some links that will help you out. Keep in mind if you have never worked with OpenCV before this is not a easy project to start with. Good Luck!
It's difficult to tell what is being asked here. This question is ambiguous, vague, incomplete, overly broad, or rhetorical and cannot be reasonably answered in its current form. For help clarifying this question so that it can be reopened, visit the help center.
Closed 11 years ago.
Can you auto take multiple pictures and then edit them in real time for an iphone app? Like say u want to make a video of a man and add a beard to him can you do it while streaming? the mustach moves according to where his face is detected and say take about 5-15 frames per second?
I guess you can.
It would involve tracking some facial features or markers added for the purpose at the least. However this is such a vast and complex field, you'll hardly get a single advice here that will get you going.
If you really mean it, I'd suggest looking for Augmented Reality libraries, there's a few out there. Most of them work by tracking a special pattern and not arbitrary features though, so be prepared for a big load of work.
Check this SO question for a first few hints, you'll find more information on the topic easily through the search engine of your choice.
It's difficult to tell what is being asked here. This question is ambiguous, vague, incomplete, overly broad, or rhetorical and cannot be reasonably answered in its current form. For help clarifying this question so that it can be reopened, visit the help center.
Closed 11 years ago.
Skype added an anti-shake feature in it's video conference app on iPhone. How can that be done?
This is quite a complicated thing to pull off, but it's probably a combination of some powerful blur detection/removal algorithms, and the gyroscope. I would start by looking into how to detect motion with the iPhone, and see what kind of results you can get with that. If it's not enough, start looking into shift/blur direction detection algorithms. This is not a trivial problem, but is something that you could probably accomplish given enough time. Hope that Helps!
It's difficult to tell what is being asked here. This question is ambiguous, vague, incomplete, overly broad, or rhetorical and cannot be reasonably answered in its current form. For help clarifying this question so that it can be reopened, visit the help center.
Closed 9 years ago.
I want to implement a kind of video conference on the iPhone. I know that the Faceti e API comes available sometime I can't wait for that..
Does somebody know where to start? Or some sample code would be very welcome!
Thanks in advance!
This website listed some source codes of real apps, including a "Two-way video conferencing for iPhone"
I would start reading this guide it contains brief introduction to the video recording.
Briefl just use UIImagePickerController to record a movie.
Encrypt it and transfer it over the network.