i know in iphone simulator to input using multi-touch, hold down the OPT key while using the mouse. will see two virtual fingertips.and move the mouse to pinch in and out.
my problem is in simulator,these two fingertips will only be overlap at the center all the time, if i jus want move one finger and another finger fix at a point,rather move two fingertips together how to achieve this in iphone simulator?in other word i want this two virtual finger can be overlap anywhere on the screen but not just center
If you press Shift-Option they will also move together with a static distance.
Related
In MapBoxGL JS the pitch of a map can be changed by holding down the CTRL key, holding down the mouse and then moving it. Is there an equivalent way to accomplish this on a touch screen with no keyboard (e.g. smartphones). The context of this question - I want to use mapInstance.getPitch to record the last pitch selected by the user.
Yes, using two fingers on one's smartphone.
For my app I would like to have it so the user can swipe in every direction using the UISwipeGestureRecgoniser. Would it be possible to instead of having just up, down, left and right, but one for ever degree and therefore I can use that to move my sprite in that direction?
UISwipeGestureRecgonizer doesn't have such possibility, it only contains 4 directions
If you want just to move object along with user finger you can try dragging event (UIControlEventTouchDragInside).
Here is nice example how to achieve this: Basic Drag and Drop in iOS
When you double tap on MKMapView : zoom in.
But how to zoom out??
There's always using two fingers to zoom in and out. On the emulator you need to hold down an option key to get 'two fingers' appear on the emulated screen (I think it's the Alt key but I use a PC keyboard on my mac so may differ on a real mac keyboard)
As for doing it procedurally you need to set the region on the mapkit.
via the setRegion method of your MKMapKit object.
Make sure you're within bounds though as it will cause a crash if not. (i.e. long -180
I am trying to build an app that detects if a user puts their lips on the screen. How could I get an image of their lips when they put it on the screen?
I know how to detect finger touches, would it be similar to that?
Edit: Also, wondering about doing this for Android.
I believe the iphone hardware supports somewhere between 10 and 20 touch points, but the SDK only gives you access to 5. To get a good outline of the lips you'd need a much higher number of touch points.
Could you use the front facing camera and avoid the obvious hygiene issues?!
You can't get the form of the touches, so you wouldn't be able to detect lips on the screens.
As others have posted here, it is not possible to detect all of the contact points of the lips and the screen. Yes, it would be similar, but more difficult. The iPhone screen is capacitive, so it should detect lips, which are similar to fingertips. (When outside in the cold and I need to scroll with my gloves on, I've used my chin.)
As far as the shape, you may be able to map multiple touches, but touches don't take on any particular shape, so you won't be able to draw the lips based on that. You may want to consider preloading generic pictures of lips and then resizing it to match the size of their lips, depending on where the touch inputs are.
I have a graphic that is 3 times the width of an iphone landscape view.
I am trying to auto scroll it so that it appears that it is moving sideways, without using the touchscreen scrolling method.
My aim is to maybe have a button you can press and it moves it left or right across the screen like an animation.
I can deal with everything else but am having trouble finding a solution.
Any example code would be appreciated or even any info on whether it is possible or not.
Thanks. Dave
You can wrap a UIView in an animation block. The animation sweeps the origin value in its frame property from one point to another, over a set period of time.