Any possible way to develop 3d earth map application for iphone? - iphone

I need to develop an iPhone 3d map application similar to virtual earth or google earth. The application will have images overlay above the 3d map, like clouds or location pin. Anyone has any ideas on that?
Regards
Edit:
Try to make the phrase not vague this time:
As far as I know, google earth and microsoft virtual earth api (3d) are not supported for any iOS devices.
In stead of redoing everything from ground up using OpenGL ES, which is the only way to do 3d in iOS devices with hardware acceleration, I want to develop a map application with established map services, such as google map. However, the map will be in 3d.
Of course, I can make a simple 3d earth using openGL ES with hardcore geo location similar to living earth HD, but I try to avoid that.

Have a look at WhirlyGlobe library : http://code.google.com/p/whirlyglobe/

You can also check out eeGeo's iOS and Android SDK, which offers vector-based 3D maps for all of the USA, Canada and UK.

Related

ARCore in unity vs Sceneform features/use cases?

The way I understand it is that there are several environments that support ARCore and Unity and Sceneform SDK are some of the options.
I was wondering how are they different from each other besides one being in Java and the other being in C#? Why would someone choose one over the other aside from language preference?
Thank you
Sceneform empowers Android developers to work with ARCore without learning 3D graphics and OpenGL. It includes a high-level scene graph API, realistic physically based renderer, an Android Studio plugin for importing, viewing, and building 3D assets, and easy integration into ARCore that makes it straightforward to build AR apps. Visit this video link of Google I/O '18.
Whereas ARCore in Unity uses three key capabilities to integrate virtual content with the real world as seen through your phone's camera:
Motion tracking
Environmental understanding allows the phone to detect the size
and location of all type of surfaces: horizontal, vertical and
angled surfaces like the ground, a coffee table or walls.
Light estimation allows the phone to estimate the environment's
current lighting conditions.
ARCore is Google’s platform for building augmented reality experiences. Using different APIs, ARCore enables your phone to sense its environment, understand the world and interact with information. Some of the APIs are available across Android and iOS to enable shared AR experiences.

Fixing object when camera open Unity AR

Im trying to create a AR Game in Unity for educational project.
I want to create something like pokemon go: when the camera open the object will be fixed somewhere on the real world and you will have to search for it with the camera.
My problem is that ARCore and vuforia groundDetection (I dont want to use targets) are only limited for few types of phone and i tried to use kudan sdk but it didnt work.
Any one can give me a tool or a tutorial on how to do this? I just need ideas or someone to tell me where to start?
Thanks in advance.
The reason why plane detection is limited to only some phones at this time is partially because older/less powerful phones cannot handle the required computing power.
If you want to make an app that has the largest reach, Vuforia is probably the way to go. Personally, I am not a fan of Vuforia, and I would suggest you use ARCore (and/or ARKit for iOS).
Since this is an educational tool and not a game, are you sure Unity is the way to go? I am sure you may be able to do it in Unity, but choosing the right platform for a project is important - just keep that in mind. You could make a native app instead.
If you want to work with ARCore and Unity (which is a great choice in general), here is the first in a series of tutorials that can get you started as a total beginner.
Let me know if you have other questions :)
You can use GPS data from phone to display object when the user arrived specific place you can show the object. You can search GPS based Augmented Reality on google. You can check this video : https://www.youtube.com/watch?v=X6djed8e4n0

Unity. Move player when mobile moves (android VR)

i'm developing VR using google cardboard SDK..
i want to move on virtual environment when i walk on real world, like this : https://www.youtube.com/watch?v=sZG5__Z9pzs&feature=youtu.be&t=48
is it possible to make VR application like that for android...? maybe using accelerometer sensor ? how can i implement this using unity...?
i try to record accelerometer sensor while i walk with smartphone, here are the result : https://www.youtube.com/watch?v=ltPwS7-3nOI [i think the accelerometer value is so random -___- ]
Actually it is not possible with only mobile:
You're up against a fundamental limitation of the humble IMU (the primary motion sensor in a smartphone).
I won't go into detail, but basically you need an external reference frame when trying to extract positional data from acceleration data. This is the topic of a lot of research right now, and it's why VR headsets that track position like the Oculus Rift have external tracking cameras.
Unfortunately, what you're trying to do is impossible without using the camera on your phone to track visual features in the scene and use those as the external reference point, which is a hell of a task better suited to a lab full of computer vision experts.
One another possible but difficult way is:
This may be possible if you connect device to internet then watch it's position from satelite(google maps or something like that)but that is a very hard thing to do.

3D Object Joints Accessing in iPhone

I have 3D object in Maya which have joints for its movements, how can I access the object's joints in iPhone, Kindly guide me so that I can make it possible.
There is no iPhone API for this. You'll have to find a game engine that runs on iPhone and integrates with Maya.
Unity and SIO2 both run on iPhone and cite Maya support, and there are probably others. Whether said support includes inverse joint data, I don't know.
Actually, try asking on https://gamedev.stackexchange.com/. I only just found out about it.

iPhone Dev: technologies being used for 3d games?

I have an idea for a 3d game for the iPhone and I'm wanting to get a feel for what is used mostly for iPhone 3d game programming. Are 3d games being developed in opengl es only or are 3d models being used as well?
I'm just trying to get an idea of some of the technologies being used for iPhone 3d game programming.
For instance, I think the following game will be really cool and would like to know what is being used to create it.
http://vodpod.com/watch/2623062-yes-plastic-for-iphone-ipod-touch
Also, are people using any 3d modeling tools like maya, 3ds max, or blender with this 3d game engines?
There's a good overview on this blog, but it may be a bit out of date.
Unity is a popular 3d engine that supports iPhone
SIO2 is an open source 3D engine for iPhone development.
There are lots of different ways to make 3D games. As the two above answers, you can use things like Unity and SI02. However, the vast, vast majority of games are developed with OpenGL, mainly because there's already so much existing material on game development with OpenGL. If the game's not too graphically intense, you can even try using core animation -- but don't expect the same performance as other things.