Is there any way we could add video filters in flutter? - flutter

Can we create a video streaming app in flutter.I am trying to make a video streaming app but can find filters in flutter.

As far as I know, there is what we call Deep AR that could provide filter for your video just like in Snapchat:
DeepAR is a web platform that provides an SDK for Snapchat kind of 3D
face filters, face lenses, and effects for any iOS, Android, Unity, or
HTML5 app. The platform makes use of a technology called Augmented
Reality (AR).
So you need to setup an account here. For a more detailed process on adding filter in your video, you can visit this blog.
For creating livestream alone, check this out: Live Streaming With Mux, Stream, and Flutter

Related

How do we integrate flutter with mux data?

I am trying to create a video streaming app in Flutter, using Mux. I also want to track user data to help recommend more specific videos per user. However, I do not know how exactly to integrate mux/data with the video_player plugin in flutter. Have I misunderstood how exactly these plugins/services work? I am using flutter for my frontend and python for the backend.
Links:
video_player: https://pub.dev/packages/video_player
mux/data: https://docs.mux.com/guides/data/track-your-video-performance
You could use this player instead which is functional: https://pub.dev/packages/mux_player
However, I'm struggling to make it autoplay and not go fullscreen when you play it as it feels kind of weird. It does track all data as you would expect.

Does Agora.io for Unity provide these features?

I'm a bit lost looking through all the various Agora.io modules (and not sure what it means that only some of them have Unity-specific downloads).
I want to make a Unity app where two remote phones exchange data as follows:
Streaming voice in both directions
Streaming video in one direction (recorded from device camera)
Streaming a small amount of continuously-changing custom data in the other direction (specifically, a position + orientation in a virtual world; probably encoded as 7 floats)
The custom data needs to have low latency but does not need reliability (it's fine if some updates get lost; app only cares about the most recent update). Updates basically every frame.
Ideally I want to support both Android and iOS.
I started looking at Agora video (successfully built a test project) and it seems like it will cover the voice and video, but I'm struggling to find a good way to send the custom data (position + orientation). It's probably theoretically possible to encode it as a custom video feed but that sounds complex and inefficient. Is there some out-of-band signalling mechanism I could use to send some extra data alongside/instead of a video?
Agora real-time messaging sounds like it would probably work for this, but I can't seem to find any info about integrating it with Unity (either on Agora's web site or in a general web search). Can I roll this in somehow?
Agora interactive gaming could maybe also be relevant? The overview doesn't seem real clear about how it's different from regular Agora video. I suspect it's overkill but that might be fine if there isn't a large performance cost.
Could anyone point me in the right direction?
I would also consider alternatives to Agora if there's a better plugin for implementing this feature set in Unity.
Agora's Video SDK for Unity supports exporting projects to Android, iOS, MacOS, and Windows (non-UWP).
Regarding your data streaming needs, Agora's RTM SDK is in the process of being ported to work within Unity. At the moment the best way to send data using the Agora SDK is to use CreateDataStream to leverage Agora's ability to open a data stream that is sent along with the frames. Data stream messages are limited to 1kb per frame and 30kb/s so I would be cautious about running it on every frame if you are using a frame-rate above 30fps.

Android: How to achieve ARCore and Video call simultanously using same camera?

I'm developing an Android app where the user is in an ARCore Session and at the same time should be able to share the (raw, not augmented) camera video stream in a video call.
The used technologies are:
Unity 2018.1.6
Google ARCore Unity SDK package
WebRTC Video Chat Unity package
The required ARCore functionality and the WebRTC Video chat work great - but only individually. Regarding having both simultaneously I'm aware it cannot work immediately; in fact having built a test app I saw that starting the video call would stop ARCore preview and reverse, dependent on what is started later.
So my questions are:
Is there a way to extract the video stream from ARCore?
How do I use this video stream as input for the WebRTC video call?

Using video chat apis for IOS without managing the serverside

I am looking for some ready-to-go APIs provided by platforms to deal with video chatting etc. There are some examples like Twilio or OpenTok but you have to create the server side by yourself.
Are there any other examples where I don't have to deal with the server?
I recommend to try ConnectyCube, they have a Swift code sample for video calling
You do not need to bother with any server-side stuff, can concentrate only on client-side app development
Code samples https://developers.connectycube.com/ios/code-samples
Video Chat swift code sample https://github.com/ConnectyCube/connectycube-ios-samples/tree/master/sample-videochat-swift
Getting Started guide https://developers.connectycube.com/ios/
Video Chat API documentation https://developers.connectycube.com/ios/videocalling
The following features are supported:
1-1 video chat
Group video chat
Cross-platform
Screen sharing
WebRTC based
End-to-end Encryption by default
VP8/H264 video codecs supported
Mute/Unmute audio/video stream
Switch video input devices (cameras)
Video recording
Vidyo.io
You can use this its free for 2000 minutes but it will work without a server,
https://vidyo.io/platform/mobile/

MPEG Dash streaming on multiple screen using EXOplayer

I have a requirement to play the MPEG Dash stream on multiple screen. So I want to play same stream on different devices using EXO player.
If anyone have any idea, how to play same MPEG DASH stream on multiple screen please get back to me.
Thank You,
If you are new to MPEG-DASH I suggest to read some introducing articles, like http://www.streamingmedia.com/Articles/Editorial/What-Is-.../What-is-MPEG-DASH-79041.aspx, and https://www.bitcodin.com/blog/2015/04/mpeg-dash/.
Next step could be to define which platforms will be supported and if you are going for a native App, like ExoPlayer, browser based playout, like dash.js, or bitmovin, or make use of the native MPEG-DASH support of some devices, like some Smart TVs.
Once you share that knowledge and what is the goal of your efforts, people on this platform can help you with more details.