Play Transparent Video - swift

I was sent a video (made in after effects) with an alpha channel (has transparency). I need to add that video to an iOS app, like adding a transparent GIF.
There is this tutorial online on how to do just that, however he is using a single mp4 that displays alpha info on the bottom and the colour info on the top, I am using Quicktime 444 which is already transparent by nature. Is there anyway I can play a transparent video on Swift for iOS? I am using SwiftUI.

Related

play video on flutter with transparent background

I did some complex 3d animations on blender and I want to play them on flutter. The problem is that when I export the animation into a video and I put it on my flutter app the transparent background is gone. I'm using video_player to play videos on flutter and the video extension I use is .webm because is the only one I know that can be readed on flutter and have a transparent background but the application puts a black background in the video.
The conclusion I get is that video player plugin is not the best idea to do it. So i've been searching about how flutter manage gifs and it's manage alpha channel perfectly because it can manage alpha channel on image. But the next problem is how to control the gif, for that I use gifimage plugin which works really well to reproduce an animation. Finally the last problem is to export from blender to a gif, blender can't export on gif format so you should export into a quicktime format with Qt animation codec and then convert it to gif. If for some reason the first frame of the gif stay as background of the gif you should edit the gif with photoshop for example and change in the timeline the first frame to not disappear to disappear.
Current player plugin uses ExoPlayer on Android. I found this issue on the repo.

Adobe Premiere Pro H.264 video export - Youtube upload - Facebook embed - black bars

I have got an Video, which was shot with 1920x1080 and 25FPS. I imported the clips into Adobe Premiere Pro and exported them with the H.264 format and with the Youtube presettings (1920x1080 and 25FPS).
In Youtube the video looks okay, also there are small black bars on top and on the bottom of the video. If i share and play the video through Facebook, there are big black bars on top and on bottom of the video.
Before this i edited videos with the same setting with Adobe Premiere Elements and there was no such a problem.
I also checked the sequenz setting, but they also seem to be alright. I can't change anything here, but they look exactly like the source video settings!
Does someone knows where the problem might be?
If using the pre-settings remember you have two options SD and HD. When selecting the HD option your export won't appear to be letter-boxed (black bars).
Letterbox means that a 16:9 widescreen video, played on a 4:3 display, has the black bars top and bottom. From my understanding 1920x1080 will give you bigger bars than if 720. If uploading to youtube or FB the black bars are pronounced even though the bars to appear on the file when viewing for example on your computer.
I can't really determine if the black bars you're getting are abnormal without a screenshot. But here is some detail info from Adobe on exporting. http://helpx.adobe.com/media-encoder/using/export-settings-reference.html#video_exports_settings

Programmatically add filter for real time video for iphone

How to add filter effect for iphone camera, such as video blur or old 8mm effect?
It looks like the current method to do this involves capturing frames from the camera as images using AVCaptureSession, also see this technical Q&A from Apple. After starting an image stream, you'd have to modify the images before displaying them.

Applying Effect to iPhone Camera Preview "Video" Using OpenGL

My goal is to write a custom camera view controller that:
Can take photos in all four interface orientations with both the back and, when available, front camera.
Properly rotates and scales the preview "video" as well as the full resolution photo.
Allows a (simple) effect to be applied to BOTH the preview "video" and full resolution photo.
My previous effort is documented in this question. My latest attempt was to modify Apple's sample GLVideoFrame (from WWDC 2010). However, I have not been able to get the iPhone 4 to display the preview "video" properly when the session preset is AVCaptureSessionPresetPhoto.
Has anyone tried this or know why the example doesn't work with this preset?
Apple's example uses a preset with 640x480 video dimensions and a default texture size of 1280x720. The iPhone 4 back camera delivers only 852x640 when the preset is AVCaptureSessionPresetPhoto.
iOS device camera video/photo dimensions when preset is AVCaptureSessionPresetPhoto:
iPhone 4 back: video is 852x640 & photos are 2592x1936
iPhone 4 front: video & photos are 640x480
iPod Touch 4G back: video & photos are 960x720
iPod Touch 4G front: video & photos are 640x480
iPhone 3GS: video is 512x384 & photos are 2048x1536
Update
I got the same garbled video result when switching Brad Larson's ColorTracking example (blog post) to use the AVCaptureSessionPresetPhoto.
The issue is that AVCaptureSessionPresetPhoto is now context-aware and runs in different resolutions based on whether you are displaying video or still image captures.
The live preview is different for this mode because it pads the rows with extra bytes. I'm guessing this is some sort of hardware optimization.
In any case, you can see how I solved the problem here:
iOS CVImageBuffer distorted from AVCaptureSessionDataOutput with AVCaptureSessionPresetPhoto
The AVCaptureSessionPresetPhoto is for taking pictures, not capturing live feed. You can read about it here: http://developer.apple.com/library/ios/#documentation/AudioVideo/Conceptual/AVFoundationPG/Articles/03_MediaCapture.html
(My belief is that this is actually two different cams or sensors, as they behave very differently, and there's a couple of seconds delay just for switching between the Photo and, say, 640x480).
You can't even use both presets at the same time, and switching between them is a headache as well - How to get both the video output and full photo resolution image in AVFoundation Framework
HTH, although not what you wanted to hear...
Oded.

Customizing the screen of Movieplayer in iPhone

I've few doubts regarding movie player app
How can I customize the screen size. I want only video or movie to be played only to half the screen and I want to add custom controls buttons to the remaining half of movie player.
In movie player example overlay view is there but how to change the name of the button and label in that and I'm unable to do that.
The MPMovePlayer UI cannot be modified. It plays fullscreen video, and provides the default controls. There's currently no way to do what you need in the official iPhone SDK.
Use this link for video player customization:
https://developer.apple.com/library/ios/samplecode/MoviePlayer_iPhone/Introduction/Intro.html#//apple_ref/doc/uid/DTS40007798
download sample code and modify accordingly.