> On Jan 15, 2017, at 3:12 PM, Quincey Morris > <quinceymor...@rivergatesoftware.com> wrote: > > On Jan 15, 2017, at 09:22 , davel...@mac.com wrote: >> >> I have an iOS presentation app >> (https://itunes.apple.com/app/redraw/id1114820588?mt=8) that I currently >> make videos from by AirPlaying it to my Mac and using Screenflow on the Mac >> to show the iPad screen and record my audio from a microphone (and then >> edit). I'd like to build this functionality into my app directly > > AVFoundation doesn’t seem to have the ability of capturing screen video on > iOS — AVCaptureScreenInput is documented as Mac only. That would rule out > AVFoundation for the basic video capture within your app. You might be able > to capture a series of screen shots, but it has to be done in real time, and > that’s going to be tricky to get right on iOS where you’ll need to buffer the > captured images to storage that might not be fast enough. > > If you mean you want to write a companion Mac app, then I guess you can use > AVCaptureScreenInput to capture the raw video, and then you could use > AVAssetWriter to export your final, composed video. However, AVAssetWriter is > *not* a real-time function, so you couldn’t rely on it keeping up if you > tried to export as the user interleaves the still images with the raw video. > What you’d need to do is add a playback/edit phase, where you played the raw > video, captured the timing of the user edits (letting the playback skip > frames if the edits held up the playback), then export the “composition” when > the user is done. (Or, you could export in the background *during* editing, > which would mean it would be done soon after the user finishes, but this may > have adverse effects on playback on a lower-end Mac.) > > AVCaptureScreenInput does let you choose the screen, though. > > FWIW, since I’m not sure I properly understood exactly what solution you’re > looking for.
I'm talking about doing this on the iPad (not with a separate Mac app). I know the only option for recording the screen itself is using ReplayKit. I don't really need to record the screen. I want to write a video in real-time that consists of the audio from the microphone and an image that changes periodically (and that image happens to be shown on a UIImageView of the second screen of my app). So given that AVAssetWriter is not real time, I think my best option (if I want to do it all the work on the iPad without a separate Mac app) is to use ReplayKit. The other option would be to let the user navigate to a specific image on the screen, record audio for that image, navigate to the next image, record audio for that image, and so on and then I could probably use AVAssetWriter to write that audio and the fixe image and then the next audio segment and the next fixed image, etc. Thanks, Dave Reed _______________________________________________ Cocoa-dev mailing list (Cocoa-dev@lists.apple.com) Please do not post admin requests or moderator comments to the list. Contact the moderators at cocoa-dev-admins(at)lists.apple.com Help/Unsubscribe/Update your Subscription: https://lists.apple.com/mailman/options/cocoa-dev/archive%40mail-archive.com This email sent to arch...@mail-archive.com