Small update, it turns out GL_viv_direct_texture can only be used to upload
textures to the GPU, not the other way around. I did try this out to be
sure, no success.
The recommendations I get from NXP is to use their "virtual framebuffer"
kernel module, that creates as many virtual framebuffers
Hi Gunnnar,
thanks a lot, this gives me a good starting point. I didn't know I could
use GL_viv_direct_texture "in reverse" like this, I'll try it out. Already
using it to stream decoded h264 video to a texture for rendering to the
screen.
I don't know which one will work best, but you want to
Hi Ola,
I think I would have preferred to skip the FBO -> QImage step as the
normal FBO readback is through glReadPixels which is quite costly, so
you'll lose a lot of performance that way. You might also benefit
from doing the RGB -> YUV conversion on the GPU if there is time
enough to do that.
Hi,
I need to render a Qt Quick scene and stream it as H264 data on an
i.MX6-device running Linux. The use case is to be used live with as little
latency as possible.
I'm thinking of using the QQuickRenderControl class. The easy way is
probably to render the scene to a QOpenGLFramebufferObject,