I am looking for a source to explain how to use FFmpeg with wowza to transcode live rtmp to http. Does anyone know anything about it or know where to point me to get some info?
Thank you all,
Wowza does not currently support live transcoding of video, however, transcoding is only necessary when you need to create new video data. As long as you are outputting the bitrate/resolutions that you need wowza can re-package to most protocals you may need for internet delivery.
Example:
3 H.264 streams in an RTMP container --> WOWZA ---> 3 RTMP streams
---> 3 Apple HLS streams (http)
---> 3 Silverlight smooth streams (http)
Quickstart Guide Here
Wowza V3 (released in October 2011) has transcoding capabilities. It sounds like you're really just trying to rewrap the H.264 RTMP steam into Apple HLS, which doesn't require any transcoding, just segmenting into HTTP packets.
Related
I set up a nginx server and I can broadcast using rtmp. I do this by compressing it with h264, but I also need to compress it with h265. rtmp and flv do not support h265.
as a result can you offer me a server and protocol to use h265? get open source if possible.
edit:
okey should not explain the problem a little more. I can send a video to my nginx server with a client and watch this video in hls and dash.
But what I have to do is compress this video with a h265 applet with a converter like ffmpeg and then watch it or send it to another client. I couldn't find anything other than rtmp to send and receive videos.
MP4 and MKV both support H.265 video streams.
You can use DASH or HLS to stream your MP4 segments.
How can I get RTSP stream on server from WebRTC and then use it for my own purposes, for example, to retranslate it somewhere. Maybe there is some solution or library?
Any help is much appreciated!
Denis
If you are willing to hook into the webrtc code then this can be done easily by doing the following:
Build a native webrtc C++ client on server
If the codec used is H.264, tap the output (encoded frames) from H264 RTP receiver.
Use a third party library such as (http://www.live555.com/openRTSP/) to build a RTSP server out of it.
If vp8, you might as well tap the output of vp8 decoder and then use ffmpeg to create rtsp stream out of it. Or consider webm may be?
most browsers can directly display a HTTP MJPEG Stream... while they can´t display h.264 via RTSP without the help of plugins...
I have a security cam that can only stream h.264 via RTSP so I can´t view the live video on my browser (nor iPhone etc.) and I do not want to install any APPs or Plugins...
But I have a Linux Server... I would like to retrieve the h.264 on my server and restream it in MJPEG, so I can browse to my server and see the MJPEG stream via HTTP...
After three days googling around and trying a lot with live555, ffmpeg, VLC and other tools I still did not get it running...
What is the right way to achieve my goal (with free tools like ffmpeg, live555 or whatsever needed...)?
thanks for any help.
I found the answer by myself... using vlc it was important to use quotation marks :sout'....' which I did not know... now it works as expected..
I'm interested in setting up a streaming video server (perhaps on a cloudfront server) with videojs. I understand that flash video can be streamed, however, is it possible to stream video using videojs and a different codec? (like h246). I tried looking through the videojs documentation and forums but did not find any additional info.
Video.js has limited support for RTMP streaming in Flash, but hopefully more in the next few months.
HTTP Live Streaming (HLS) is the most supported streaming format for HTML5 (iOS, Safari, latest Android). Video.js can support that on the devices that support HLS natively.
I think you would have to transcode the h264 file on the fly to get the effect you want. Subsonic is a program which will read your file structure, display your videos and music in a webui, transcode the audio/video and stream it--but it uses jwplayer, not videojs.
However, it is opensource, so if you want to try to modify that, I'm sure it would be possible.
Will the ALAC format support live streaming in iPhone ? the ALAC audio recording format is streamed to Server machine? so will i be able to play the audio chunk data, does ALAC format support?
Thank You.
Assuming you mean "Apple Lossless" audio...
I don't see why it wouldn't, but I don't know the details. You'll probably need to embed it in a transport stream instead of a MPEG 4 container (but then, I don't know how the HTTP live streaming works either).
I don't think streaming lossless audio is sensible, though.
Streaming lossless audio is possible, we have flac streaming using icecast and it works beautifully. However, we are not using HTTP Live Stream (HLS) to do it. We stream flac from the source generator to a number of servers and they create HLS's from there.
It is technically possible to mux alac into mpegts (ffmpeg can do this) as well as play it back (using ffmpeg), but there isn't a format identifier for other clients. Adding this feature to HLS will be as easy as calling/writing Apple and asking them to add ALAC to this list:
http://www.smpte-ra.org/mpegreg/mpegreg.html
and update their products accordingly. If you've purchased an Apple product less than 90 days ago, or you have AppleCare: give them a call. They have to work on the issue for you if you are covered. The more requests that get elevated to their engineers, the more likely they are to add support for alac in HLS.