Am 29.09.2020 um 14:58 schrieb Mark Filipak (ffmpeg):
On 09/29/2020 04:06 AM, Michael Koch wrote:
Am 29.09.2020 um 04:28 schrieb Mark Filipak (ffmpeg):

I just want to understand the frame structures that ffmpeg creates, and that ffmpeg uses in processing and filtering. Are Y, Cb, Cr separate buffers? That would be logical. Or are the Y, Cb, Cr values combined and organized similarly to macroblocks? I've found some code that supports that. Or are the Y, Cb, Cr values thrown together, pixel-by-pixel. That would be logical, too.

As far as I understood it, that depends on the pixel format.
For example there are "packed" pixel formats rgb24, bgr24, argb, rgba, abgr, bgra,rgb48be, rgb48le, bgr48be, bgr48le.
And there are "planar" pixel formats gbrp, bgrp16be, bgrp16le.

Hi Michael,

"Packed" and "planar", eh? What evidence do you have? ...Share the candy!

As far as I know, this is not described in the official documentation. You can find it for example here:
https://video.stackexchange.com/questions/16374/ffmpeg-pixel-format-definitions


Now, I'm not talking about streams. I'm talking about after decoding. I'm talking about the buffers. I would think that a single, consistent format would be used.


There is no single consistent format used internally. See Gyan's answer here:
http://ffmpeg.org/pipermail/ffmpeg-user/2020-September/050031.html

Michael


_______________________________________________
ffmpeg-user mailing list
ffmpeg-user@ffmpeg.org
https://ffmpeg.org/mailman/listinfo/ffmpeg-user

To unsubscribe, visit link above, or email
ffmpeg-user-requ...@ffmpeg.org with subject "unsubscribe".

Reply via email to