[st-ericsson] v4l2 vs omx for camera
felipe.contreras at gmail.com
Sat Feb 26 05:26:36 PST 2011
On Thu, Feb 24, 2011 at 3:04 PM, Hans Verkuil <hverkuil at xs4all.nl> wrote:
> On Thursday, February 24, 2011 13:29:56 Linus Walleij wrote:
>> 2011/2/23 Sachin Gupta <sachin.gupta at linaro.org>:
>> > The imaging coprocessor in today's platforms have a general purpose DSP
>> > attached to it I have seen some work being done to use this DSP for
>> > graphics/audio processing in case the camera use case is not being tried or
>> > also if the camera usecases does not consume the full bandwidth of this
>> > dsp.I am not sure how v4l2 would fit in such an architecture,
>> Earlier in this thread I discussed TI:s DSPbridge.
>> In drivers/staging/tidspbridge
>> you find the TI hackers happy at work with providing a DSP accelerator
>> Isn't it possible for a V4L2 component to use this interface (or something
>> more evolved, generic) as backend for assorted DSP offloading?
Yes it is, and it has been part of my to-do list for some time now.
>> So using one kernel framework does not exclude using another one
>> at the same time. Whereas something like DSPbridge will load firmware
>> into DSP accelerators and provide control/datapath for that, this can
>> in turn be used by some camera or codec which in turn presents a
>> V4L2 or ALSA interface.
> Yes, something along those lines can be done.
> While normally V4L2 talks to hardware it is perfectly fine to talk to a DSP
> The hardest part will be to identify the missing V4L2 API pieces and design
> and add them. I don't think the actual driver code will be particularly hard.
> It should be nothing more than a thin front-end for the DSP. Of course, that's
> just theory at the moment :-)
The pieces are known already. I started a project called gst-dsp,
which I plan to split into the gst part, and the part that
communicates with the DSP, this part can move to kernel side with a
It's easier to identify the code in the patches for FFmpeg:
> The problem is that someone has to do the actual work for the initial driver.
> And I expect that it will be a substantial amount of work. Future drivers should
> be *much* easier, though.
> A good argument for doing this work is that this API can hide which parts of
> the video subsystem are hardware and which are software. The application really
> doesn't care how it is organized. What is done in hardware on one SoC might be
> done on a DSP instead on another SoC. But the end result is pretty much the same.
More information about the gstreamer-devel