I am using xmpp framework in my application for chat functionality. I have sent the Image but how can i send the Audio file using this. Any idea.
Related
I want develop an video call application for mobile. I search a lot but I dont' find a library in flutter, dart programming language for capture audio and video from webcam and microphone. I want get a single data flow/stream thus I can send it via socket from client to client.
Can you tell me if there is a way to capture video with audio and send it via socket with dart?
I am using StreamYard services to do a live stream to multiple destinations like Facebook and youtube. I want to create a mobile app using flutter that can receive that stream and use it (i mean to display the live stream). Streamyard uses only RMTP for a custom destination.
My question is: is there a way to create and host a custom RMTP online that can be shared between the StreamYard platform and my mobile app. I want it to work whenever I do live stream from the StreamYard it will be shared to Facebook, youtube, and my mobile app as well. I have done a lot of research but i find out the only way is to use windows or linux as a host, which i want it online.
Also, alternative solutions are welcome like using webRTC.
Because you use StreamYard, I think you need to use the INVITE feature to start a Video Chat then convert to live streaming, it works like bellow:
UserA --WebRTC--->---+
+--->- StreamYeard -->-RTMP-->- YouTube/Twitch.tv
UserB --WebRTC---->--+
You need to buy a non-free plan to support Custom RTMP destinations to publish the RTMP to your media server like SRS or Nginx, then you could broadcast to multiple destination, like this:
+->-- YouTube/Twitch.tv
|
StreamYeard ->-RTMP-+->- Custom RTMP destinations --+--RTMP-> YouTube/Twitch.tv
(SRS/Nginx media server) |
+--HLS/FLV--> Flutter App
Note: Once streaming to your RTMP server or video cloud platform, you could covert to HLS/HTTP-FLV for your FlutterAPP to play it. About player and protocol, please read here. It depends on which part you want to build by yourself, and it's possible to build by open-source projects.
Note: Note that you could use StreamYeard to streaming to YouTube and Custom RTMP server, or use FFmpeg to pull stream from your custom server then publish to any other live streaming platform.
For this solution, the StreamYeard actually plays as Video Chat or video conference platform, like ZOOM. It will transcode each WebRTC stream and mix all the audio and videos to one RTMP stream.
So you could use WebRTC server to build your StreamYeard, then use FFmpeg to transcode and mix the streams, because it is off topic so let me stop here.
Is there any way to send recorded voice in message programmatically using objective-c and iPhone.
I am sending text message using messageUI framework but now I want to send voice message too.
There's nothing straightforward built for voice messages built into iOS, unlike "MFMessageComposeViewController". You'll have to write your own voice recording code or make use of something open source or third party/commercial. Here's a related question that talks about how to record audio through iOS.
Plus this depends on how you want the recipient to receive the recorded voice message? A mp3 or some other audio file sent via e-mail (which you can do using the MFMailComposeViewCOntroller) or some other way?
I am trying existing to stream music/video on an iphone using HTTP Live Streaming. I read the apple docs on HTTP live streaming (http://developer.apple.com/library/mac/#documentation/NetworkingInternet/Conceptual/StreamingMediaGuide/Introduction/Introduction.html), and I get how it works.
What it doesn't say is how one would use iphone as a server? Do I have to add the tools to my ios app(mediastreamsegmenter, variantplaylistcreator) and run it as a NSTask or is there some kind of native support to stream media files.
If you really want to stream from an iPhone app you can't do this with the iPhone acting as a server. You need a separate server where you can send data from the iPhone app. So you can use the camera or the microphone in the app to get live content and then you can send asynchronously data to the server, which using mediastreamsegmenter and variantplaylistcreator will convert the data to ts segments and then will append them at the end of the m3u8 file and meanwhile another iPhone app can act as a client and watch the live content that you are streaming from the first app.
From my experience this is the only way to achieve that. Hope that helps.
I am using Red5 + RTMP in a client-server flash application.
There aren't audio/video streams in my application. RTMP is used to transfer messages from app to server and back.
Now I need to develop the application for iPhone:
is there any RTMP implementation on iPhone?
If not, how could I solve this problem? Is there any alternative to RTMP on iPhone?
And most important question: could it be solved without rewriting whole server part of application? (Red5 + RTMP)
If all you need is to pass "messages" and no a/v, might I suggest using a REST service or something similar such as doing GET or POST requests containing your "message" encoded in some form? Such as JSON or BASE64?