Mall Map Integration into iPhone Application - iphone

Hello everyone i am stuck into one problem. I am working on Application in which Mall Map required to integrate.
I need to display Mall map like this:
I have done proper r&d but not found any solution yet.
Can it be possible if yes then how can i implement this ?
Any hint or suggestion will be appreciated.
Thanks.

You could display the map in a UIImage and place this image in a UIScrollView. After that you need to divide the image into region rectangles, each identifying one room.
When the user taps the UIScrollView you need to determine which region he tapped on.
Using your room grid you can detect which room he tapped on. Of course this is easier for not zoomed images.
In a zoomed state you need to convert the touched point to the "unzoomed touched point" and check this point for the grid.

Related

Tile Overlays in iPhone App

There is a webpage I am looking at that displays a Google Map (via javascript), but certain areas of the map have an overlay (textures appear different than the rest of the map, etc.). From what I can gather, this is done on the webpage via a getTileURL function, which depends on the zoom level and current coordinate.
I am interested in portraying this same overlay in my iPhone application, which currently has a mapView set up. Is there anyway to implement the same sort of getTile function in objective-c for a mapView?
If someone could please provide some sample code or point me in the right direction I would greatly appreciate it. Thank you.

UIImageView interactions

I am working on an app in which I want similar kind of functionality as that of WebMD body image.
How can I identify which part of image is touched in an optimal way? Do I have to slice the image according to requirements?
How can I add some tags into the image? Similar to the facebook photo upload functionality in iphone.
You need some way to figure out what the user touched, or tried to touch.
You might use a list of annotation-like objects, where each object has a location. When the user touches the image, you'll need to find the annotation in the list that's closest to the touch location and react appropriately. The "optimal" way to do that is probably to use a quad tree. For an iPhone app, though, the number of touchable points is probably pretty small (several dozen?), and a brute force search through the list will probably be more than fast enough.
Another option would be to overlay a transparent view on top of your image for each region that you want the user to be able to touch. Doing this would also make it simple to draw a "tag" at each of those locations.

Putting GPS current location and pin annotations on a UIImageView

I am making an iPhone app for a local college. I am currently trying to take a drawn map of the campus and put pin annotations on it. I also want to be able to show the users current location on this map image. Right now I just have a UIImageView inside of a UIScrollView. It can currently scroll and zoom in/out. How would I go about accomplishing this? I have searched everywhere for an answer and can't seem to find a solution (or have overlooked it). Am I going in the right direction? Thanks for your help.
You would have to come up with a function to translate latitude and longitude into pixel offsets in the image. Assuming the campus is relatively small compared to the Earth (so you can ignore the planet's curvature), this should be a simple matter of determining the coordinates of the four corners of the image and interpolating.

How to get a 3D picture in an iphone app?

In an application, i saw that they used to display pictures of vehicles. But what was amazing was when we touch and swipe in that picture, it rotates in 3d way left and right. And from the front view we can rotate and get to see its back view also. It is a very good feature and i was trying to replicate it. But couldnt get an idea of how and where to start. My doubts are
Whats the actual format of the thing, it surely isn't a picture.
How do they get to rotate it?
Could someone give me an idea where i should start or where I should look upon?
Just like the KennyTM told you, OpenGL-ES is the weapon of choice. Take pictures of that object from all the sides you need to show, then use those as textures for the faces of the cube. Got the idea?

Buttons and Cartesian Coordinates

I have this scrollview "eager" to be filled up with buttons (4 per row), and resembling iphone apps icons, I want the user to be able to rearrange them has she/he pleases. If it were one row of buttons only, Apple Autoscroll tutorial app would be an excellent example to follow. Not the case, this is 2D and I cannot figure out how to do it. Couple of ideas tried: An array keeping buttons coordinates as reference; at this moment I'm trying to find out a way to pick up the buttons that have "lower/lesser" coordinates than the one being dragged. I've been searching and researching with no success. Any mathematical algorithm to do this? I suspect there are some other solutions out there seriously tested and robust. Can you pin point some?
Thank you very much.
Why don't you use the TTLauncherView?