I'm Android developer, my app use kotlin and opengl es 3.0, you can think of it as a simple 2D photo edit app, I want make it run on windows,web,ios and macos, so I learn flutter, Because in consideration phase, I would like to ask experienced person, if Flutter supports the use of OpengL ES on multiple platforms? If not, is it still necessary to use platform-specific graphical interface language for each platform while using flutter?
At the moment this is what we have.
Can I build 3D (OpenGL) apps with Flutter? Today we don’t support for
3D via OpenGL ES or similar. We have long-term plans to expose an
optimized 3D API, but right now we’re focused on 2D.
Related
I have to make an app that uses virtual reality, so should I drop the idea of using Flutter?
yes as much as I have seen flutter does support AR,I have been following a flutter developer on twitter he posts some cool AR stuff built with flutter here's a plugin ARCore he has built for flutter.
here are some of sample AR videos from the developer himself
https://twitter.com/i/status/1123893412279791616
https://twitter.com/i/status/1129117305303175168
Can I build 3D (OpenGL) apps with Flutter?
Today we don’t support for 3D via OpenGL ES or similar.We have long-term plans to expose an optimized 3D API, but right now we’re focused on 2D.
https://flutter.dev/docs/resources/faq#can-i-build-3d-opengl-apps-with-flutter
there aren't any OpenGL bindings supported by flutter. Flutter is only a 2d only application.
https://flutter.io/faq/#can-i-build-3d-opengl-apps-with-flutter
https://github.com/flutter/flutter/issues/14591
https://github.com/flutter/flutter/issues/7053
https://github.com/flutter/flutter/issues/179
I am not sure how VR would work at all on flutter.
You can use google's ar core with flutter. Check out the arcore_flutter_plugin to work with ar in a flutter.
As of now, there aren't any packages that specifically target VR. But you can use ARKIT arkit_flutter_plugin.
NOTE: ARCORE only works with android. And ARKIT only works with iOS.
I recently created a Flutter plugin for AR that supports both Android and iOS by wrapping around ARCore and ARKit: https://pub.dev/packages/ar_flutter_plugin
The plugin is a work in progress, but it already supports collaborative AR and sharing content through Google's Cloud Anchor Service and a lot of other useful features
guys i am working on a project which uses unity engien and kinect as input source ..now according to my knowledge there is not much support between unity and kinect sdk ..i have heard about zigfu framework but it is not giving me all functionalities i need..so what are options for me? im thinking to take some functionalities from zigfu and some from a background application build in .net 4.0 and using kinect official sdk ? can i connect to kinect via two interfaces at the same time? i.e zigfu and kinect sdk ....my background app will connect to unity via pipes ..is that agood option?
I've already done something similar. I'd like to use Unity 3D engine and do some interactions to animate the model using Kinect (Kinect SDK). Some functionality in Kinect SDK are not available in Zigfu, such as Hand Gripping detection.
Because Kinect SDK is suitable for WPF application, here is my solution :
Build your Unity into Unity Standalone (PC, Mac, Linux).
Create WPF application with Kinect stuff inside.
Add WindowsFormsHost inside your XAML of WPF application.
Embed your Unity Standalone into WPF using WindowsFormsHost.
To do a communication between WPF and Unity, you can use Raknet. It will work as socket does.
in my experience, its usually not a good idea to use "two of" something, when they both do the same thing. I've never heard of zigfu before, but it seems relatively easy to learn. Since its available as a unity plug in, it may be best to use that over kinect. The reason being that Unity isn't to "friendly" with third party applications.
If your aiming for XNA, its possible to convert easily if the plug-in doesn't already do it for you.
I Highly recommend looking over the unity forums, and the ZDK documentation.
http://forum.unity3d.com/threads/127924-Zigfu-dev-kit-for-Unity3D-Product-Launch
as you know monotouch is the c# face of ios native api. is there any opportunity to use monotouch in ios games? i examined unity3d and unreal development kit. there is no ephesized feature like this. is there any way to use c# in an ios game engine?
For example: i want to start a phone call in a game
In Unity iPhone you can access native features of iOS by calling them through plugins. Usually those plugin methods are written in ObjC/C/C++.
http://unity3d.com/support/documentation/Manual/Plugins.html
It might be possible to write plugins using MonoTouch, and access them from Unity after they've been compiled to native code, but I seriously doubt anyone has attempted that.
I have decided to use the game engine Unity to develop my game for iPhone. But I need to use some functions built into the iOS SDK.
If I choose to develop using Unity, can I still use functions from the standard iOS SDK? Like functions to access a url, etc...
When you compile a Unity project for iOS it takes your C# or JavaScript code and AOT compiles it to a native dll. It then creates an xcode project which loads that dll. You can add native objectiveC, or C/C++ functions to this project and expose them for calling from the Unity engine.
http://unity3d.com/support/documentation/Manual/Plugins.html
Unity also has built-in methods for accessing URLs.
I don't know how the Unity SDK is built, but if it's a static Objective-C library then you can access iOS SDK functions without any problems. If its a script engine where you develop your scripts outside the IDE then you may not access iOS function.
Have you checked other engines/sdk's for your game development? I know that Cocos2D is well documented and I know for sure that you can call iOS methods.
Good luck!
I used both unity3d and corona which are very good at their field, and yes you can use all standart SDK over them, sometimes you need to do tricks though (which is quite a pain). If you are developing a 3d game and familiar with NVIDIA PhysX engine go for unity, it gives you nice features. But if you are clueless about collasions/rigidbodies/ragdolls etc you would have a hard time for sure.
As for 2d application development since I don't like Objective-C I prefer corona which you code in lua (easy to learn and use). You can use most of the IOS SDK in it without any problem.
I really reccomend you starting with Unity, it is the best engine I have ever used (I have used a ton of bad engines).
Unity is the best place to start and to stay. Coding, graphics and everything is simple to manage in this engine.
Unity is specially 3d but you can also build 2d games, some examples are Zombieville and OMG Pirates (very succesfull games on the appstore).
What options are there for game developers in regards to working with the MonoTouch framework?
Is OpenGL fully supported? Or just a subset?
Is there any additional tools, frameworks or engines that would be compatible with MonoTouch? Would these libraries also have to be written in .NET?
Do you have any other tips?
Thanks all.
OpenGL is fully supported in MonoTouch, we distribute OpenTK with support for OpenGL ES 1 and ES 2, so it has a nice binding that you can use from managed languages like C#.
As the previous poster hinted, there is XnaTouch available today, and there is some talk about bringing Cocos2D on the Monotouch-o-sphere.
An alternative to using MonoTouch with OpenGL or XnaTouch is to use another Mono-based framework that is entirely focused on gaming, the Unity3D platform. This is a professional tool for building 3D and 2D games and goes beyond what Xna does, it is an entire gaming IDE with a full game engine.
XNATouch: Free implementation of the XNA Framework for MonoTouch
unity3d: C# game library for iPhone. It's based on Mono but don't think it is related to MonoTouch.