I am playing a video on my Custom Receiver via the Media Player Library. In my receiver, I want to update the video Metadata to set the Title of the currently playing media. What is the best way to do this?
Currently, I am updating the Media Information object in the Media Manager when the video starts to play, and this seems to work okay, but I am wondering if there is a better way to do this? Possibly inside of the Media Manager On Load method I have overridden.
Thanks!
Related
Please help me, I have a problem when I play audio from local files, if I play one file at a time, I can. and it's working. but I want like applications in general to be able to play, next, stop, etc. using a background service. the data file has been successfully displayed in the application, the data of the audio file is in the form of a LIST, while the data requested by the audioservice is in the form of a MEDIAITEM
Suggest you to have a look on audio_service package. This is meant for playing audio in background, the use case you are looking for(play,next,pause).
My question is if there is any built-in interpretation of metadata by the video player in iOS. I know one can add meta-data to a video and interpret it within a custom application as shown here.
In iOS on ipod or iphone, an HTML video is opened within the native player. I would like to display a message above or below the video for a short duration at the beginning. Since I cannot control the native player I thought there might be some built in metadata interpretation that could be used to perform this. I have not been able to find any information on this.
Any help is appreciated.
The blog you've posted includes details on using the native player MPMoviePlayerController to display meta data, which is pretty cool actually. You learn something new every day! If you're making a Phonegap App I suppose you could write a plugin to do this?
Or alternatively, have a look at this other OS question which appears to suggest that it is possible - though not seemingly with metadata embedded in the actual video. Apparently this works on iOS.
Reading metadata from the <track> of an HTML5 <video> using Captionator
Is the custom player only for soundcloud mp3's, or is it possible to use the player for any project, like playing your own mp3 files from your server.
Thanks
Pascal
In order to play audio using the Soundcloud player, you have to create a Soundcloud account & upload your music to your "own" account.
But you can actually customize your player if this so happens to be what you mean. By connecting to a Soundcloud API which then returns data from Soundcloud, which then. You use CSS/ to style your output results.
Hope I could help.
It can only work with audio on SoundCloud.
I am trying to create a streaming video DVR like functionality in an app I am developing. I have an HTTP Live Stream that I have successfully gotten to play on the iPad. I want the user to be able to push the "Record" button, and begin recording the video that is currently playing from that point. This video file will be accessible from the app or from the camera roll. Currently, I am using the MPMoviePlayerController object to play the video stream. I do not see any methods of accessing the data from the object in Apple's documentation. Here are some thoughts I had on ways of going about this.
1) Somehow access the video data from MPMoviePlayerController, and write this to a file. Or use another type of player object that will allow me to play the video and access the currently playing data.
2) Implement some sort of screen capture recording that gets a video capture of the iPad's screen. This would allow me to record the video in a "screenshot" sort of way.
3) Locate the HTTP Live Streaming video segments where they are stored by MPMoviePlayerController. Presumably they need to be stored somewhere on the iPad for playback. Is there a way of accessing these files?
4) Manually download the stream video segments over http while streaming the file. This seems like its not ideal since the stream would have to be downloaded twice.
5) This could work. Periodically download the video segments to the iPhone. Set up a local http server on the iPhone and server the videos to the MPMoviePlayerController. This way the video segments could be marked for recording and assembled into a video.
6) I do have control of the streaming server. I could write some server side code to record the video on the server end, then send the video to the iPad after the fact. I would rather not do this.
Has anyone done any of these things? Ideally the iPhone would just be able to access the video data somehow and easily record it. I would rather not get into options 4, 5, or 6 (above) if I don't have to.
Thanks in advance.
DVR on the device is somewhat not encouraged, due to the limited space available and other factors like battery life, processing power, cleanup procedures after the user stops the dvr, etc.
If you want to achieve DVR playback on iOS devices (or other devices using HLS), I suggest you keep the video server side. The live stream is already captured and segmented server side, all you would have to do is keep the segments a bit longer, instead of deleting them. By using the EXT-X-PLAYLIST-TYPE and EXT-X-MEDIA-SEQUENCE tags, you can suggest to the player that he's opening a live stream which has DVR (earlier) video available.
Alternatively, you can use a server that does that out of the box, for example Wowza. Here's an article on how to achieve this with Wowza
I have a number of music tracks which I would like the user to be able to preview a small clip of each.
These tracks are on a server.
How is media streamed into the app and which player is used? Can a custom player be created to play the clips within the view, without e.g. quicktime player opening?
Thanks
If you don't want to use QuickTime, the matter is rather complex, as far as I know. Fortunately, a lot of work already has been done for you by Matt Gallagher. See this excelent post for further information. The code, that he provides works perfectly in my application.