Hello,

I'm looking for pointers on how to properly integrate the following
SoC functions into
Linux (media).

The SoC in question has the following components:
- camera interface
- MPEG2/4 HW decoder
- yuv->rgb conversion and scaling engine

I have prototype code which captures 2 interlaced fields from the
camera interface,
constructs a full frame in system memory using the chips DMA engine and finally
uses the scaler to convert the yuv data to rgb and scale it up to full
display resulution;
however it is one monolithic driver and one has to unload it to use
the mpeg decoder
(which also uses the scaler).

The hardware blocks work independently from each other, the scaler is
usually used
to DMA its output into one of four framebuffer windows.

The goal is to have mplayer capture analog video from the camera
interface and/or
have it feed digital video (DVB) through the mpeg decoder and display
it. Additionally,
since the scaler is indepedent from the rest, mplayer should be able
to at least use
the colorspace converter and decode other video formats in software.

Are there any drivers in the kernel already which implement this kind of scheme?

Thanks!
        Manuel Lauss
--
To unsubscribe from this list: send the line "unsubscribe linux-media" in
the body of a message to majord...@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to