Message ID | 20180917173022.9338-1-ezequiel@collabora.com (mailing list archive) |
---|---|
Headers | show |
Series | Add Rockchip VPU JPEG encoder | expand |
On 09/17/2018 07:30 PM, Ezequiel Garcia wrote: > This series adds support for JPEG encoding via the VPU block > present in Rockchip platforms. Currently, support for RK3288 > and RK3399 is included. > > Please, see the previous versions of this cover letter for > more information. > > Compared to v5, the only change is in the V4L2_CID_JPEG_QUANTIZATION > control. We've decided to support only baseline profile, > and only add 8-bit luma and chroma tables. > > struct v4l2_ctrl_jpeg_quantization { > __u8 luma_quantization_matrix[64]; > __u8 chroma_quantization_matrix[64]; > }; > > By doing this, it's clear that we don't currently support anything > but baseline. > > This series should apply cleanly on top of > > git://linuxtv.org/hverkuil/media_tree.git br-cedrus tag. > > If everyone is happy with this series, I'd like to route the devicetree > changes through the rockchip tree, and the rest via the media subsystem. OK, so I have what is no doubt an annoying question: do we really need a JPEG_RAW format? The JPEG header is really easy to parse for a decoder and really easy to prepend to the compressed image for the encoder. The only reason I can see for a JPEG_RAW is if the image must start at some specific address alignment. Although even then you can just pad the JPEG header that you will add according to the alignment requirements. I know I am very late with this question, but I never looked all that closely at what a JPEG header looks like. But this helped: https://en.wikipedia.org/wiki/JPEG_File_Interchange_Format and it doesn't seem difficult at all to parse or create the header. I also think there are more drivers (solo6x10) that manipulate the JPEG header. Regards, Hans > > Compliance > ========== > > (Same results as v3) > > v4l2-compliance SHA: d0f4ea7ddab6d0244c4fe1e960bb2aaeefb911b9, 64 bits > > Compliance test for device /dev/video0: > > Driver Info: > Driver name : rockchip-vpu > Card type : rockchip,rk3399-vpu-enc > Bus info : platform: rockchip-vpu > Driver version : 4.18.0 > Capabilities : 0x84204000 > Video Memory-to-Memory Multiplanar > Streaming > Extended Pix Format > Device Capabilities > Device Caps : 0x04204000 > Video Memory-to-Memory Multiplanar > Streaming > Extended Pix Format > Media Driver Info: > Driver name : rockchip-vpu > Model : rockchip-vpu > Serial : > Bus info : > Media version : 4.18.0 > Hardware revision: 0x00000000 (0) > Driver version : 4.18.0 > Interface Info: > ID : 0x0300000c > Type : V4L Video > Entity Info: > ID : 0x00000001 (1) > Name : rockchip,rk3399-vpu-enc-source > Function : V4L2 I/O > Pad 0x01000002 : Source > Link 0x02000008: to remote pad 0x1000005 of entity 'rockchip,rk3399-vpu-enc-proc': Data, Enabled, Immutable > > Required ioctls: > test MC information (see 'Media Driver Info' above): OK > test VIDIOC_QUERYCAP: OK > > Allow for multiple opens: > test second /dev/video0 open: OK > test VIDIOC_QUERYCAP: OK > test VIDIOC_G/S_PRIORITY: OK > test for unlimited opens: OK > > Debug ioctls: > test VIDIOC_DBG_G/S_REGISTER: OK (Not Supported) > test VIDIOC_LOG_STATUS: OK (Not Supported) > > Input ioctls: > test VIDIOC_G/S_TUNER/ENUM_FREQ_BANDS: OK (Not Supported) > test VIDIOC_G/S_FREQUENCY: OK (Not Supported) > test VIDIOC_S_HW_FREQ_SEEK: OK (Not Supported) > test VIDIOC_ENUMAUDIO: OK (Not Supported) > test VIDIOC_G/S/ENUMINPUT: OK (Not Supported) > test VIDIOC_G/S_AUDIO: OK (Not Supported) > Inputs: 0 Audio Inputs: 0 Tuners: 0 > > Output ioctls: > test VIDIOC_G/S_MODULATOR: OK (Not Supported) > test VIDIOC_G/S_FREQUENCY: OK (Not Supported) > test VIDIOC_ENUMAUDOUT: OK (Not Supported) > test VIDIOC_G/S/ENUMOUTPUT: OK (Not Supported) > test VIDIOC_G/S_AUDOUT: OK (Not Supported) > Outputs: 0 Audio Outputs: 0 Modulators: 0 > > Input/Output configuration ioctls: > test VIDIOC_ENUM/G/S/QUERY_STD: OK (Not Supported) > test VIDIOC_ENUM/G/S/QUERY_DV_TIMINGS: OK (Not Supported) > test VIDIOC_DV_TIMINGS_CAP: OK (Not Supported) > test VIDIOC_G/S_EDID: OK (Not Supported) > > Control ioctls: > test VIDIOC_QUERY_EXT_CTRL/QUERYMENU: OK > test VIDIOC_QUERYCTRL: OK > test VIDIOC_G/S_CTRL: OK > test VIDIOC_G/S/TRY_EXT_CTRLS: OK > test VIDIOC_(UN)SUBSCRIBE_EVENT/DQEVENT: OK > test VIDIOC_G/S_JPEGCOMP: OK (Not Supported) > Standard Controls: 2 Private Controls: 0 > > Format ioctls: > test VIDIOC_ENUM_FMT/FRAMESIZES/FRAMEINTERVALS: OK > test VIDIOC_G/S_PARM: OK (Not Supported) > test VIDIOC_G_FBUF: OK (Not Supported) > test VIDIOC_G_FMT: OK > test VIDIOC_TRY_FMT: OK > test VIDIOC_S_FMT: OK > test VIDIOC_G_SLICED_VBI_CAP: OK (Not Supported) > test Cropping: OK (Not Supported) > test Composing: OK (Not Supported) > test Scaling: OK > > Codec ioctls: > test VIDIOC_(TRY_)ENCODER_CMD: OK (Not Supported) > test VIDIOC_G_ENC_INDEX: OK (Not Supported) > test VIDIOC_(TRY_)DECODER_CMD: OK (Not Supported) > > Buffer ioctls: > test VIDIOC_REQBUFS/CREATE_BUFS/QUERYBUF: OK > test VIDIOC_EXPBUF: OK > > Test input 0: > > Streaming ioctls: > test read/write: OK (Not Supported) > test blocking wait: OK > test MMAP: OK > test USERPTR: OK (Not Supported) > test DMABUF: Cannot test, specify --expbuf-device > > Total: 48, Succeeded: 48, Failed: 0, Warnings: 0 > > v6: > - As agreed with Hans change the quantization control > to support only strict baseline JPEGs, with 8-bit > coefficients. > v5: > - Make coefficients 2-byte wide, to support 8-bit and 16-bit > precision coefficient. Also, add a 'precision' field, to > allow the user to specifiy the coefficient precision. > - Minor style changes as requested by Hans. > - Get rid of all the unused leftover code. > - Move driver to staging. The driver will support video encoding, > via the request API. So let's use staging until the > controls are stabilized. > v4: > - Change luma and chroma array controls, with a compound > control, as suggested by Paul. > v3: > - Refactor driver to allow a more elegant integration with > other codec modes (h264 decoding, jpeg decoding, etc). > Each variant can now be encoders, decoders or both. > - Register driver in the media controller framework, > in preparation for the Request API. > - Set values for JPEG quantization controls in the core, as suggested > by Tomasz and Hans. > - Move pm_runtime_get/put to run/done, reducing power consumption. > This was possible thanks to Miouyouyou, who pointed out the power > domains missing [1]. > - Use bulk clock API for simpler code. > v2: > - Add devicetree binding documentation and devicetree changes > - Add documentation to added pixel format and controls > - Address Hans' review comments > - Get rid of unused running_ctx field > - Fix wrong planar pixel format depths > - Other minor changes for v4l2-compliance > - Drop .crop support, we will add support for the > selector API later, if needed. > > [1] https://github.com/Miouyouyou/RockMyy/blob/master/patches/kernel/v4.18/DTS/0026-ARM-DTSI-rk3288-Set-the-VPU-MMU-power-domains.patch > > Ezequiel Garcia (4): > dt-bindings: Document the Rockchip VPU bindings > ARM: dts: rockchip: add VPU device node for RK3288 > arm64: dts: rockchip: add VPU device node for RK3399 > media: add Rockchip VPU JPEG encoder driver > > Shunqian Zheng (2): > media: Add JPEG_RAW format > media: Add controls for JPEG quantization tables > > .../bindings/media/rockchip-vpu.txt | 30 + > .../media/uapi/v4l/extended-controls.rst | 25 + > .../media/uapi/v4l/pixfmt-compressed.rst | 9 + > .../media/videodev2.h.rst.exceptions | 1 + > MAINTAINERS | 7 + > arch/arm/boot/dts/rk3288.dtsi | 14 +- > arch/arm64/boot/dts/rockchip/rk3399.dtsi | 14 +- > drivers/media/v4l2-core/v4l2-ctrls.c | 8 + > drivers/media/v4l2-core/v4l2-ioctl.c | 1 + > drivers/staging/media/Kconfig | 2 + > drivers/staging/media/Makefile | 1 + > drivers/staging/media/rockchip/vpu/Makefile | 9 + > drivers/staging/media/rockchip/vpu/TODO | 5 + > .../media/rockchip/vpu/rk3288_vpu_hw.c | 123 ++++ > .../rockchip/vpu/rk3288_vpu_hw_jpeg_enc.c | 126 ++++ > .../media/rockchip/vpu/rk3288_vpu_regs.h | 442 +++++++++++++ > .../media/rockchip/vpu/rk3399_vpu_hw.c | 123 ++++ > .../rockchip/vpu/rk3399_vpu_hw_jpeg_enc.c | 154 +++++ > .../media/rockchip/vpu/rk3399_vpu_regs.h | 601 +++++++++++++++++ > .../staging/media/rockchip/vpu/rockchip_vpu.h | 292 +++++++++ > .../media/rockchip/vpu/rockchip_vpu_common.h | 31 + > .../media/rockchip/vpu/rockchip_vpu_drv.c | 528 +++++++++++++++ > .../media/rockchip/vpu/rockchip_vpu_enc.c | 604 ++++++++++++++++++ > .../media/rockchip/vpu/rockchip_vpu_hw.h | 65 ++ > include/uapi/linux/v4l2-controls.h | 10 + > include/uapi/linux/videodev2.h | 2 + > 26 files changed, 3225 insertions(+), 2 deletions(-) > create mode 100644 Documentation/devicetree/bindings/media/rockchip-vpu.txt > create mode 100644 drivers/staging/media/rockchip/vpu/Makefile > create mode 100644 drivers/staging/media/rockchip/vpu/TODO > create mode 100644 drivers/staging/media/rockchip/vpu/rk3288_vpu_hw.c > create mode 100644 drivers/staging/media/rockchip/vpu/rk3288_vpu_hw_jpeg_enc.c > create mode 100644 drivers/staging/media/rockchip/vpu/rk3288_vpu_regs.h > create mode 100644 drivers/staging/media/rockchip/vpu/rk3399_vpu_hw.c > create mode 100644 drivers/staging/media/rockchip/vpu/rk3399_vpu_hw_jpeg_enc.c > create mode 100644 drivers/staging/media/rockchip/vpu/rk3399_vpu_regs.h > create mode 100644 drivers/staging/media/rockchip/vpu/rockchip_vpu.h > create mode 100644 drivers/staging/media/rockchip/vpu/rockchip_vpu_common.h > create mode 100644 drivers/staging/media/rockchip/vpu/rockchip_vpu_drv.c > create mode 100644 drivers/staging/media/rockchip/vpu/rockchip_vpu_enc.c > create mode 100644 drivers/staging/media/rockchip/vpu/rockchip_vpu_hw.h >
On Fri, 2018-09-28 at 14:33 +0200, Hans Verkuil wrote: > On 09/17/2018 07:30 PM, Ezequiel Garcia wrote: > > This series adds support for JPEG encoding via the VPU block > > present in Rockchip platforms. Currently, support for RK3288 > > and RK3399 is included. > > > > Please, see the previous versions of this cover letter for > > more information. > > > > Compared to v5, the only change is in the V4L2_CID_JPEG_QUANTIZATION > > control. We've decided to support only baseline profile, > > and only add 8-bit luma and chroma tables. > > > > struct v4l2_ctrl_jpeg_quantization { > > __u8 luma_quantization_matrix[64]; > > __u8 chroma_quantization_matrix[64]; > > }; > > > > By doing this, it's clear that we don't currently support anything > > but baseline. > > > > This series should apply cleanly on top of > > > > git://linuxtv.org/hverkuil/media_tree.git br-cedrus tag. > > > > If everyone is happy with this series, I'd like to route the devicetree > > changes through the rockchip tree, and the rest via the media subsystem. > > OK, so I have what is no doubt an annoying question: do we really need > a JPEG_RAW format? > Not annoying, as it helps clarify a few things :-) I think we do need the JPEG_RAW format. The way I see it, using JPEG opens a can of worms... > The JPEG header is really easy to parse for a decoder and really easy to > prepend to the compressed image for the encoder. > > The only reason I can see for a JPEG_RAW is if the image must start at > some specific address alignment. Although even then you can just pad the > JPEG header that you will add according to the alignment requirements. > > I know I am very late with this question, but I never looked all that > closely at what a JPEG header looks like. But this helped: > > https://en.wikipedia.org/wiki/JPEG_File_Interchange_Format > > and it doesn't seem difficult at all to parse or create the header. > > ... I think that having JPEG_RAW as the compressed format is much more clear for userspace, as it explicitly specifies what is expected. This way, for a stateless encoder, applications are required to set quantization and/or entropy tables, and are then in charge of using the compressed JPEG_RAW payload in whatever form they want. Stupid simple. On the other side, if the stateless encoder driver supports JPEG (creating headers in-kernel), it means that: *) applications should pass a quality control, if supported, and the driver will use hardcoded tables or... *) applications pass quantization control and, if supported, entropy control. The kernel uses them to create the JPEG frame. But also, some drivers (e.g. Rockchip), use default entropy tables, which should now be in the kernel. So the application would have to query controls to find out what to do. Not exactly hard, but I think having the JPEG_RAW is much simpler and more clear. Now, for stateless decoders, supporting JPEG_RAW means the application has to pass quantization and entropy controls, probably using the Request API. Given the application has parsed the JPEG, it knows the width and height and can request buffers accordingly. The hardware is stateless, and so is the driver. On the other hand, supporting JPEG would mean that drivers will have to parse the image, extracting the quantization and entropy tables. Format negotiation is now more complex, either we follow the stateful spec, introducing a little state machine in the driver... or we use the Request API, but that means parsing on both sides kernel and application. Either way, using JPEG_RAW is just waaay simpler and puts things where they belong. > I also think there are more drivers (solo6x10) that > manipulate the JPEG header. Well, I've always thought this was kind of suboptimal. Thanks, Ezequiel
On Mon, 2018-10-01 at 14:54 -0300, Ezequiel Garcia wrote: > On Fri, 2018-09-28 at 14:33 +0200, Hans Verkuil wrote: > > On 09/17/2018 07:30 PM, Ezequiel Garcia wrote: > > > This series adds support for JPEG encoding via the VPU block > > > present in Rockchip platforms. Currently, support for RK3288 > > > and RK3399 is included. > > > > > > Please, see the previous versions of this cover letter for > > > more information. > > > > > > Compared to v5, the only change is in the V4L2_CID_JPEG_QUANTIZATION > > > control. We've decided to support only baseline profile, > > > and only add 8-bit luma and chroma tables. > > > > > > struct v4l2_ctrl_jpeg_quantization { > > > __u8 luma_quantization_matrix[64]; > > > __u8 chroma_quantization_matrix[64]; > > > }; > > > > > > By doing this, it's clear that we don't currently support anything > > > but baseline. > > > > > > This series should apply cleanly on top of > > > > > > git://linuxtv.org/hverkuil/media_tree.git br-cedrus tag. > > > > > > If everyone is happy with this series, I'd like to route the devicetree > > > changes through the rockchip tree, and the rest via the media subsystem. > > > > OK, so I have what is no doubt an annoying question: do we really need > > a JPEG_RAW format? > > > > Not annoying, as it helps clarify a few things :-) > I think we do need the JPEG_RAW format. The way I see it, using > JPEG opens a can of worms... > > > The JPEG header is really easy to parse for a decoder and really easy to > > prepend to the compressed image for the encoder. > > > > The only reason I can see for a JPEG_RAW is if the image must start at > > some specific address alignment. Although even then you can just pad the > > JPEG header that you will add according to the alignment requirements. > > > > I know I am very late with this question, but I never looked all that > > closely at what a JPEG header looks like. But this helped: > > > > https://en.wikipedia.org/wiki/JPEG_File_Interchange_Format > > > > and it doesn't seem difficult at all to parse or create the header. > > > > > > ... I think that having JPEG_RAW as the compressed format > is much more clear for userspace, as it explicitly specifies > what is expected. > > This way, for a stateless encoder, applications are required > to set quantization and/or entropy tables, and are then in > charge of using the compressed JPEG_RAW payload in whatever form > they want. Stupid simple. > > On the other side, if the stateless encoder driver supports > JPEG (creating headers in-kernel), it means that: > > *) applications should pass a quality control, if supported, > and the driver will use hardcoded tables or... > > *) applications pass quantization control and, if supported, > entropy control. The kernel uses them to create the JPEG frame. > But also, some drivers (e.g. Rockchip), use default entropy > tables, which should now be in the kernel. > > So the application would have to query controls to find out > what to do. Not exactly hard, but I think having the JPEG_RAW > is much simpler and more clear. > > Now, for stateless decoders, supporting JPEG_RAW means > the application has to pass quantization and entropy > controls, probably using the Request API. > Given the application has parsed the JPEG, > it knows the width and height and can request > buffers accordingly. > > The hardware is stateless, and so is the driver. > > On the other hand, supporting JPEG would mean that > drivers will have to parse the image, extracting > the quantization and entropy tables. > > Format negotiation is now more complex, > either we follow the stateful spec, introducing a little > state machine in the driver... or we use the Request API, > but that means parsing on both sides kernel and application. > > Either way, using JPEG_RAW is just waaay simpler and puts > things where they belong. > As discussed on IRC, I'm sending a v7 for this series, fixing only the checkpatch issues and the extra line in the binding document. We've agreed to move forward with JPEG_RAW, for the reasons stated above. Plus, we've agreed to keep this in staging until the userspace support for JPEG_RAW format is clear. Regards, Ezequiel
Em Thu, 04 Oct 2018 20:39:31 -0300 Ezequiel Garcia <ezequiel@collabora.com> escreveu: > On Mon, 2018-10-01 at 14:54 -0300, Ezequiel Garcia wrote: > > On Fri, 2018-09-28 at 14:33 +0200, Hans Verkuil wrote: > > > On 09/17/2018 07:30 PM, Ezequiel Garcia wrote: > > > > This series adds support for JPEG encoding via the VPU block > > > > present in Rockchip platforms. Currently, support for RK3288 > > > > and RK3399 is included. > > > > > > > > Please, see the previous versions of this cover letter for > > > > more information. > > > > > > > > Compared to v5, the only change is in the V4L2_CID_JPEG_QUANTIZATION > > > > control. We've decided to support only baseline profile, > > > > and only add 8-bit luma and chroma tables. > > > > > > > > struct v4l2_ctrl_jpeg_quantization { > > > > __u8 luma_quantization_matrix[64]; > > > > __u8 chroma_quantization_matrix[64]; > > > > }; > > > > > > > > By doing this, it's clear that we don't currently support anything > > > > but baseline. > > > > > > > > This series should apply cleanly on top of > > > > > > > > git://linuxtv.org/hverkuil/media_tree.git br-cedrus tag. > > > > > > > > If everyone is happy with this series, I'd like to route the devicetree > > > > changes through the rockchip tree, and the rest via the media subsystem. > > > > > > OK, so I have what is no doubt an annoying question: do we really need > > > a JPEG_RAW format? > > > > > > > Not annoying, as it helps clarify a few things :-) > > I think we do need the JPEG_RAW format. The way I see it, using > > JPEG opens a can of worms... > > > > > The JPEG header is really easy to parse for a decoder and really easy to > > > prepend to the compressed image for the encoder. > > > > > > The only reason I can see for a JPEG_RAW is if the image must start at > > > some specific address alignment. Although even then you can just pad the > > > JPEG header that you will add according to the alignment requirements. > > > > > > I know I am very late with this question, but I never looked all that > > > closely at what a JPEG header looks like. But this helped: > > > > > > https://en.wikipedia.org/wiki/JPEG_File_Interchange_Format > > > > > > and it doesn't seem difficult at all to parse or create the header. > > > > > > > > > > ... I think that having JPEG_RAW as the compressed format > > is much more clear for userspace, as it explicitly specifies > > what is expected. > > > > This way, for a stateless encoder, applications are required > > to set quantization and/or entropy tables, and are then in > > charge of using the compressed JPEG_RAW payload in whatever form > > they want. Stupid simple. > > > > On the other side, if the stateless encoder driver supports > > JPEG (creating headers in-kernel), it means that: > > > > *) applications should pass a quality control, if supported, > > and the driver will use hardcoded tables or... > > > > *) applications pass quantization control and, if supported, > > entropy control. The kernel uses them to create the JPEG frame. > > But also, some drivers (e.g. Rockchip), use default entropy > > tables, which should now be in the kernel. > > > > So the application would have to query controls to find out > > what to do. Not exactly hard, but I think having the JPEG_RAW > > is much simpler and more clear. > > > > Now, for stateless decoders, supporting JPEG_RAW means > > the application has to pass quantization and entropy > > controls, probably using the Request API. > > Given the application has parsed the JPEG, > > it knows the width and height and can request > > buffers accordingly. > > > > The hardware is stateless, and so is the driver. > > > > On the other hand, supporting JPEG would mean that > > drivers will have to parse the image, extracting > > the quantization and entropy tables. > > > > Format negotiation is now more complex, > > either we follow the stateful spec, introducing a little > > state machine in the driver... or we use the Request API, > > but that means parsing on both sides kernel and application. > > > > Either way, using JPEG_RAW is just waaay simpler and puts > > things where they belong. > > > > As discussed on IRC, I'm sending a v7 for this series, > fixing only the checkpatch issues and the extra line in the > binding document. > > We've agreed to move forward with JPEG_RAW, for the reasons > stated above. > > Plus, we've agreed to keep this in staging until the userspace > support for JPEG_RAW format is clear. The problem is that a new format is actually a V4L2 core change, and not something for staging. IMHO, it is prudent to wait for an userspace code before hushing merging this upstream. - I'm concerned about adding a JPEG_RAW format. We had this discussion in the past. On that time, it was opted to add a code inside the drivers that will be adding the jpeg headers, on the cases where the hardware was providing headerless frames. Among the reasons, different drivers may be requiring different logic to add the jpeg headers, depending on how they interpret the quality parameters and on the Huffmann tables used by an specific codec. I can't see why this couldn't be done on a stateless codec. I mean: in the specific case of JPEG, each frame is independently coded. It doesn't need to store anything from a previous frame in order to produce a JPEG header. So, from my side, at least for the first version of this driver, please make it produce a full JPEG frame with the headers, and use V4L2_PIX_FMT_JPEG format. That should allow us to add this driver on staging without adding a new JPEG_RAW format. If you still think we'll need a V4L2_PIX_FMT_JPEG_RAW, I'd like to merge it by the time we also have a decoder for libv4l. I would also like to see, at the Kernel patch series, the addition of such support for already existing non-staging camera drivers that internally receive headless JPEG headers from the hardware. This will allow us to test the userspace code with different models and proof that a generic JPEG_RAW decoder on userspace is doable. Thanks, Mauro
On Fri, 2018-10-05 at 09:10 -0300, Mauro Carvalho Chehab wrote: > Em Thu, 04 Oct 2018 20:39:31 -0300 > Ezequiel Garcia <ezequiel@collabora.com> escreveu: > > > On Mon, 2018-10-01 at 14:54 -0300, Ezequiel Garcia wrote: > > > On Fri, 2018-09-28 at 14:33 +0200, Hans Verkuil wrote: > > > > On 09/17/2018 07:30 PM, Ezequiel Garcia wrote: > > > > > This series adds support for JPEG encoding via the VPU block > > > > > present in Rockchip platforms. Currently, support for RK3288 > > > > > and RK3399 is included. > > > > > > > > > > Please, see the previous versions of this cover letter for > > > > > more information. > > > > > > > > > > Compared to v5, the only change is in the V4L2_CID_JPEG_QUANTIZATION > > > > > control. We've decided to support only baseline profile, > > > > > and only add 8-bit luma and chroma tables. > > > > > > > > > > struct v4l2_ctrl_jpeg_quantization { > > > > > __u8 luma_quantization_matrix[64]; > > > > > __u8 chroma_quantization_matrix[64]; > > > > > }; > > > > > > > > > > By doing this, it's clear that we don't currently support anything > > > > > but baseline. > > > > > > > > > > This series should apply cleanly on top of > > > > > > > > > > git://linuxtv.org/hverkuil/media_tree.git br-cedrus tag. > > > > > > > > > > If everyone is happy with this series, I'd like to route the devicetree > > > > > changes through the rockchip tree, and the rest via the media subsystem. > > > > > > > > OK, so I have what is no doubt an annoying question: do we really need > > > > a JPEG_RAW format? > > > > > > > > > > Not annoying, as it helps clarify a few things :-) > > > I think we do need the JPEG_RAW format. The way I see it, using > > > JPEG opens a can of worms... > > > > > > > The JPEG header is really easy to parse for a decoder and really easy to > > > > prepend to the compressed image for the encoder. > > > > > > > > The only reason I can see for a JPEG_RAW is if the image must start at > > > > some specific address alignment. Although even then you can just pad the > > > > JPEG header that you will add according to the alignment requirements. > > > > > > > > I know I am very late with this question, but I never looked all that > > > > closely at what a JPEG header looks like. But this helped: > > > > > > > > https://en.wikipedia.org/wiki/JPEG_File_Interchange_Format > > > > > > > > and it doesn't seem difficult at all to parse or create the header. > > > > > > > > > > > > > > ... I think that having JPEG_RAW as the compressed format > > > is much more clear for userspace, as it explicitly specifies > > > what is expected. > > > > > > This way, for a stateless encoder, applications are required > > > to set quantization and/or entropy tables, and are then in > > > charge of using the compressed JPEG_RAW payload in whatever form > > > they want. Stupid simple. > > > > > > On the other side, if the stateless encoder driver supports > > > JPEG (creating headers in-kernel), it means that: > > > > > > *) applications should pass a quality control, if supported, > > > and the driver will use hardcoded tables or... > > > > > > *) applications pass quantization control and, if supported, > > > entropy control. The kernel uses them to create the JPEG frame. > > > But also, some drivers (e.g. Rockchip), use default entropy > > > tables, which should now be in the kernel. > > > > > > So the application would have to query controls to find out > > > what to do. Not exactly hard, but I think having the JPEG_RAW > > > is much simpler and more clear. > > > > > > Now, for stateless decoders, supporting JPEG_RAW means > > > the application has to pass quantization and entropy > > > controls, probably using the Request API. > > > Given the application has parsed the JPEG, > > > it knows the width and height and can request > > > buffers accordingly. > > > > > > The hardware is stateless, and so is the driver. > > > > > > On the other hand, supporting JPEG would mean that > > > drivers will have to parse the image, extracting > > > the quantization and entropy tables. > > > > > > Format negotiation is now more complex, > > > either we follow the stateful spec, introducing a little > > > state machine in the driver... or we use the Request API, > > > but that means parsing on both sides kernel and application. > > > > > > Either way, using JPEG_RAW is just waaay simpler and puts > > > things where they belong. > > > > > > > As discussed on IRC, I'm sending a v7 for this series, > > fixing only the checkpatch issues and the extra line in the > > binding document. > > > > We've agreed to move forward with JPEG_RAW, for the reasons > > stated above. > > > > Plus, we've agreed to keep this in staging until the userspace > > support for JPEG_RAW format is clear. > > The problem is that a new format is actually a V4L2 core change, and > not something for staging. > > IMHO, it is prudent to wait for an userspace code before hushing > merging this upstream. > > - > > I'm concerned about adding a JPEG_RAW format. We had this discussion > in the past. On that time, it was opted to add a code inside the > drivers that will be adding the jpeg headers, on the cases where > the hardware was providing headerless frames. > I suspect such discussion was held on completely different hardware. What we are trying to support here has specific needs, and thus why we need new formats. > Among the reasons, different drivers may be requiring different > logic to add the jpeg headers, depending on how they interpret > the quality parameters and on the Huffmann tables used by an > specific codec. > > I can't see why this couldn't be done on a stateless codec. I mean: > in the specific case of JPEG, each frame is independently coded. > It doesn't need to store anything from a previous frame in order to > produce a JPEG header. > Well, this can be done. But since it's the wrong architecture, it will bring a more complex API and more complex implementation. In the case of a stateless decoder, it will mean a really cumbersome API. As I have explained above, using JPEG format on hardware than can only deal with JPEG_RAW format means: 1) Userspace sets JPEG format on the OUTPUT queue (sink) 2) Userspace needs to know the raw format, but it has to wait for the driver to parse it first. So, 3) it will QBUF a JPEG buffer, and 4) get the V4L2_EVENT_SOURCE_CHANGE event in order to move forward. This implies a stateful machinery in the driver. Like I said, more complex API, more complex implementation... and only to avoid a new pixelformat? Doesn't sound a good deal :-) In comparison, a proper stateless decoder driver will expose a simpler API, provide a 1:1 output/capture buffer relationship, and expose the full capabilities to userspace. > So, from my side, at least for the first version of this driver, > please make it produce a full JPEG frame with the headers, and > use V4L2_PIX_FMT_JPEG format. > This means using some form of JPEG quantization tables and huffman tables in the kernel, such as in drivers/media/usb/gspca/topro.c. Really doesn't sound like the best choice, violating the mechanism/policy separation, as now the kernel is deciding the tables for the user. The hardware can take any quantization table, and allowing userspace flexible usage, but we are going to restrict that to just a specific usage and a specific set of tables... again, only to avoid a new pixelformat? > That should allow us to add this driver on staging without adding > a new JPEG_RAW format. > AFAIK, Rockchip and Allwinner JPEG codecs handle so-called JPEG_RAW format, i.e. the payload of a SOF segment. I think having these two cases should be enough reason to add a new format. > If you still think we'll need a V4L2_PIX_FMT_JPEG_RAW, I'd like > to merge it by the time we also have a decoder for libv4l. I > would also like to see, at the Kernel patch series, the addition > of such support for already existing non-staging camera > drivers that internally receive headless JPEG headers from the > hardware. This will allow us to test the userspace code with > different models and proof that a generic JPEG_RAW decoder on > userspace is doable. > We can add a libv4l2 plugin for this decoder if that helps moving forward. Although I suspect it will just fall into oblivion, as chromeos and android will do their own thing. JPEG is a very well known format, and it's trivial for applications to add the headers. Thanks, Eze
Em Fri, 05 Oct 2018 12:37:43 -0300 Ezequiel Garcia <ezequiel@collabora.com> escreveu: > On Fri, 2018-10-05 at 09:10 -0300, Mauro Carvalho Chehab wrote: > > Em Thu, 04 Oct 2018 20:39:31 -0300 > > Ezequiel Garcia <ezequiel@collabora.com> escreveu: > > > > > On Mon, 2018-10-01 at 14:54 -0300, Ezequiel Garcia wrote: > > > > On Fri, 2018-09-28 at 14:33 +0200, Hans Verkuil wrote: > > > > > On 09/17/2018 07:30 PM, Ezequiel Garcia wrote: > > > > > > This series adds support for JPEG encoding via the VPU block > > > > > > present in Rockchip platforms. Currently, support for RK3288 > > > > > > and RK3399 is included. > > > > > > > > > > > > Please, see the previous versions of this cover letter for > > > > > > more information. > > > > > > > > > > > > Compared to v5, the only change is in the V4L2_CID_JPEG_QUANTIZATION > > > > > > control. We've decided to support only baseline profile, > > > > > > and only add 8-bit luma and chroma tables. > > > > > > > > > > > > struct v4l2_ctrl_jpeg_quantization { > > > > > > __u8 luma_quantization_matrix[64]; > > > > > > __u8 chroma_quantization_matrix[64]; > > > > > > }; > > > > > > > > > > > > By doing this, it's clear that we don't currently support anything > > > > > > but baseline. > > > > > > > > > > > > This series should apply cleanly on top of > > > > > > > > > > > > git://linuxtv.org/hverkuil/media_tree.git br-cedrus tag. > > > > > > > > > > > > If everyone is happy with this series, I'd like to route the devicetree > > > > > > changes through the rockchip tree, and the rest via the media subsystem. > > > > > > > > > > OK, so I have what is no doubt an annoying question: do we really need > > > > > a JPEG_RAW format? > > > > > > > > > > > > > Not annoying, as it helps clarify a few things :-) > > > > I think we do need the JPEG_RAW format. The way I see it, using > > > > JPEG opens a can of worms... > > > > > > > > > The JPEG header is really easy to parse for a decoder and really easy to > > > > > prepend to the compressed image for the encoder. > > > > > > > > > > The only reason I can see for a JPEG_RAW is if the image must start at > > > > > some specific address alignment. Although even then you can just pad the > > > > > JPEG header that you will add according to the alignment requirements. > > > > > > > > > > I know I am very late with this question, but I never looked all that > > > > > closely at what a JPEG header looks like. But this helped: > > > > > > > > > > https://en.wikipedia.org/wiki/JPEG_File_Interchange_Format > > > > > > > > > > and it doesn't seem difficult at all to parse or create the header. > > > > > > > > > > > > > > > > > > ... I think that having JPEG_RAW as the compressed format > > > > is much more clear for userspace, as it explicitly specifies > > > > what is expected. > > > > > > > > This way, for a stateless encoder, applications are required > > > > to set quantization and/or entropy tables, and are then in > > > > charge of using the compressed JPEG_RAW payload in whatever form > > > > they want. Stupid simple. > > > > > > > > On the other side, if the stateless encoder driver supports > > > > JPEG (creating headers in-kernel), it means that: > > > > > > > > *) applications should pass a quality control, if supported, > > > > and the driver will use hardcoded tables or... > > > > > > > > *) applications pass quantization control and, if supported, > > > > entropy control. The kernel uses them to create the JPEG frame. > > > > But also, some drivers (e.g. Rockchip), use default entropy > > > > tables, which should now be in the kernel. > > > > > > > > So the application would have to query controls to find out > > > > what to do. Not exactly hard, but I think having the JPEG_RAW > > > > is much simpler and more clear. > > > > > > > > Now, for stateless decoders, supporting JPEG_RAW means > > > > the application has to pass quantization and entropy > > > > controls, probably using the Request API. > > > > Given the application has parsed the JPEG, > > > > it knows the width and height and can request > > > > buffers accordingly. > > > > > > > > The hardware is stateless, and so is the driver. > > > > > > > > On the other hand, supporting JPEG would mean that > > > > drivers will have to parse the image, extracting > > > > the quantization and entropy tables. > > > > > > > > Format negotiation is now more complex, > > > > either we follow the stateful spec, introducing a little > > > > state machine in the driver... or we use the Request API, > > > > but that means parsing on both sides kernel and application. > > > > > > > > Either way, using JPEG_RAW is just waaay simpler and puts > > > > things where they belong. > > > > > > > > > > As discussed on IRC, I'm sending a v7 for this series, > > > fixing only the checkpatch issues and the extra line in the > > > binding document. > > > > > > We've agreed to move forward with JPEG_RAW, for the reasons > > > stated above. > > > > > > Plus, we've agreed to keep this in staging until the userspace > > > support for JPEG_RAW format is clear. > > > > The problem is that a new format is actually a V4L2 core change, and > > not something for staging. > > > > IMHO, it is prudent to wait for an userspace code before hushing > > merging this upstream. > > > > - > > > > I'm concerned about adding a JPEG_RAW format. We had this discussion > > in the past. On that time, it was opted to add a code inside the > > drivers that will be adding the jpeg headers, on the cases where > > the hardware was providing headerless frames. > > > > I suspect such discussion was held on completely different hardware. > What we are trying to support here has specific needs, and thus > why we need new formats. > > > Among the reasons, different drivers may be requiring different > > logic to add the jpeg headers, depending on how they interpret > > the quality parameters and on the Huffmann tables used by an > > specific codec. > > > > I can't see why this couldn't be done on a stateless codec. I mean: > > in the specific case of JPEG, each frame is independently coded. > > It doesn't need to store anything from a previous frame in order to > > produce a JPEG header. > > > > Well, this can be done. But since it's the wrong architecture, > it will bring a more complex API and more complex implementation. I'm not saying that this is a final solution. We're talking about a driver that it is intended to be merged (for now) at staging. "Polluting" V4L2 core with a new almost identical format that has a potential of bringing problems due to a staging driver is something that doesn't sound right. Before changing the core, we need to be able to test your proposal on more drivers, on both directions (capture and output). > In the case of a stateless decoder, it will mean a really cumbersome > API. As I have explained above, using JPEG format on hardware > than can only deal with JPEG_RAW format means: > > 1) Userspace sets JPEG format on the OUTPUT queue (sink) I'm assuming that you're actually talking about a M2M device whose output stream receives a JPEG frame, and its capture stream outputs something like RGB, right? Also, I'm assuming the description you gave for the "raw JPEG" as a JPEG image without JPEG headers. > 2) Userspace needs to know the raw format, but it has > to wait for the driver to parse it first. Why would userspace need to know the raw format? It doesn't really matter if the output is raw JPEG or full JPEG. For the output stream scenario: 1) for a headerless-JPEG (raw JPEG is really a crappy name): Userspace will use the headers to extract some information useful for the decoder to work, passing them via controls. Then, it will queue a buffer. Request API is needed, as the parameters will change frame by frame. 2) for "true" JPEG (e. g. a jpeg frame that follows the standard and has headers): Kernelspace will parse the jpeg headers, extracting the parameters they need. No need to use request API, as each frame will already contain the needed metadata. No matter if (1) or (2) is used, a M2M device will receive one JPEG frame, decode it and produce one RGB frame. No need to store any state. The opposite would equally work, and it is identical to what we have already in kernel, for the "true" JPEG output. A M2M device whose output stream is RGB and capture is MPEG: 1) for "true" JPEG: The driver will use a Huffmann table based on a quality parameter (V4L2_CID_JPEG_COMPRESSION_QUALITY). Such table would either be hardcoded at the driver or at the codec's firmware. The output stream will contain the headers. 1) for a headerless-JPEG: The driver will use a Huffmann table based on a quality parameter (V4L2_CID_JPEG_COMPRESSION_QUALITY). Such table would either be hardcoded at the driver or at the codec's firmware. The output stream will not contain the headers. Userspace will need to figure out what it was used by the Kernel in order to add the JPEG header. I'm not sure if we can do it on a transparent way for all devices. In any case, one input produces one output. > > So, 3) it will QBUF a JPEG buffer, and 4) get the V4L2_EVENT_SOURCE_CHANGE > event in order to move forward. This implies a stateful machinery in the > driver. Like I said, more complex API, more complex implementation... > and only to avoid a new pixelformat? Doesn't sound a good deal :-) I really can't understand what you're meaning by that... I can't see any need to generate an event to move forward. > > In comparison, a proper stateless decoder driver will expose > a simpler API, provide a 1:1 output/capture buffer relationship, > and expose the full capabilities to userspace. > > > So, from my side, at least for the first version of this driver, > > please make it produce a full JPEG frame with the headers, and > > use V4L2_PIX_FMT_JPEG format. > > > > This means using some form of JPEG quantization tables and huffman > tables in the kernel, such as in drivers/media/usb/gspca/topro.c. > > Really doesn't sound like the best choice, violating the > mechanism/policy separation, as now the kernel is deciding > the tables for the user. What the Huffmann tables has to do with it? No matter if the JPEG has headers or not, the codec should be doing RLE encoding/decoding. For that, it needs to know the Huffmann tables. Anyway, you can use a control to pass huffmann tables at runtime, using the request API, no matter if JPEG has headers or not. > The hardware can take any quantization table, and allowing userspace > flexible usage, but we are going to restrict that to just a specific > usage and a specific set of tables... again, only to avoid a new > pixelformat? Same here: passing quantization tables to the driver has nothing to do with having headers or not. For decoding, having headers is actually better, as the table will be at the JPEG header. For encoding, a V4L2 array control. could be used. > > That should allow us to add this driver on staging without adding > > a new JPEG_RAW format. > > > > AFAIK, Rockchip and Allwinner JPEG codecs handle so-called JPEG_RAW > format, i.e. the payload of a SOF segment. And so, almost all other JPEG devices we have already. > I think having these > two cases should be enough reason to add a new format. I'm not against adding it. Just against adding it without further discussions, in a way that it would potentially cause troubles for userspace to encode/decode JPEG. > > If you still think we'll need a V4L2_PIX_FMT_JPEG_RAW, I'd like > > to merge it by the time we also have a decoder for libv4l. I > > would also like to see, at the Kernel patch series, the addition > > of such support for already existing non-staging camera > > drivers that internally receive headless JPEG headers from the > > hardware. This will allow us to test the userspace code with > > different models and proof that a generic JPEG_RAW decoder on > > userspace is doable. > > > > We can add a libv4l2 plugin for this decoder if that helps moving > forward. Although I suspect it will just fall into oblivion, > as chromeos and android will do their own thing. I don't have any control what people do on forked Kernels. > > JPEG is a very well known format, and it's trivial for applications > to add the headers. With the current documentation you added at the V4L2 core patch, I seriously doubt that it would be possible to do it in a way that it would work with different vendors. Also, I don't know any current V4L2 generic application that would do that. Having it at libv4l would allow those apps to work with that. Thanks, Mauro
On Fri, 2018-10-05 at 13:49 -0300, Mauro Carvalho Chehab wrote: > Em Fri, 05 Oct 2018 12:37:43 -0300 > Ezequiel Garcia <ezequiel@collabora.com> escreveu: > > > On Fri, 2018-10-05 at 09:10 -0300, Mauro Carvalho Chehab wrote: > > > Em Thu, 04 Oct 2018 20:39:31 -0300 > > > Ezequiel Garcia <ezequiel@collabora.com> escreveu: > > > > > > > On Mon, 2018-10-01 at 14:54 -0300, Ezequiel Garcia wrote: > > > > > On Fri, 2018-09-28 at 14:33 +0200, Hans Verkuil wrote: > > > > > > On 09/17/2018 07:30 PM, Ezequiel Garcia wrote: > > > > > > > This series adds support for JPEG encoding via the VPU block > > > > > > > present in Rockchip platforms. Currently, support for RK3288 > > > > > > > and RK3399 is included. > > > > > > > > > > > > > > Please, see the previous versions of this cover letter for > > > > > > > more information. > > > > > > > > > > > > > > Compared to v5, the only change is in the V4L2_CID_JPEG_QUANTIZATION > > > > > > > control. We've decided to support only baseline profile, > > > > > > > and only add 8-bit luma and chroma tables. > > > > > > > > > > > > > > struct v4l2_ctrl_jpeg_quantization { > > > > > > > __u8 luma_quantization_matrix[64]; > > > > > > > __u8 chroma_quantization_matrix[64]; > > > > > > > }; > > > > > > > > > > > > > > By doing this, it's clear that we don't currently support anything > > > > > > > but baseline. > > > > > > > > > > > > > > This series should apply cleanly on top of > > > > > > > > > > > > > > git://linuxtv.org/hverkuil/media_tree.git br-cedrus tag. > > > > > > > > > > > > > > If everyone is happy with this series, I'd like to route the devicetree > > > > > > > changes through the rockchip tree, and the rest via the media subsystem. > > > > > > > > > > > > OK, so I have what is no doubt an annoying question: do we really need > > > > > > a JPEG_RAW format? > > > > > > > > > > > > > > > > Not annoying, as it helps clarify a few things :-) > > > > > I think we do need the JPEG_RAW format. The way I see it, using > > > > > JPEG opens a can of worms... > > > > > > > > > > > The JPEG header is really easy to parse for a decoder and really easy to > > > > > > prepend to the compressed image for the encoder. > > > > > > > > > > > > The only reason I can see for a JPEG_RAW is if the image must start at > > > > > > some specific address alignment. Although even then you can just pad the > > > > > > JPEG header that you will add according to the alignment requirements. > > > > > > > > > > > > I know I am very late with this question, but I never looked all that > > > > > > closely at what a JPEG header looks like. But this helped: > > > > > > > > > > > > https://en.wikipedia.org/wiki/JPEG_File_Interchange_Format > > > > > > > > > > > > and it doesn't seem difficult at all to parse or create the header. > > > > > > > > > > > > > > > > > > > > > > ... I think that having JPEG_RAW as the compressed format > > > > > is much more clear for userspace, as it explicitly specifies > > > > > what is expected. > > > > > > > > > > This way, for a stateless encoder, applications are required > > > > > to set quantization and/or entropy tables, and are then in > > > > > charge of using the compressed JPEG_RAW payload in whatever form > > > > > they want. Stupid simple. > > > > > > > > > > On the other side, if the stateless encoder driver supports > > > > > JPEG (creating headers in-kernel), it means that: > > > > > > > > > > *) applications should pass a quality control, if supported, > > > > > and the driver will use hardcoded tables or... > > > > > > > > > > *) applications pass quantization control and, if supported, > > > > > entropy control. The kernel uses them to create the JPEG frame. > > > > > But also, some drivers (e.g. Rockchip), use default entropy > > > > > tables, which should now be in the kernel. > > > > > > > > > > So the application would have to query controls to find out > > > > > what to do. Not exactly hard, but I think having the JPEG_RAW > > > > > is much simpler and more clear. > > > > > > > > > > Now, for stateless decoders, supporting JPEG_RAW means > > > > > the application has to pass quantization and entropy > > > > > controls, probably using the Request API. > > > > > Given the application has parsed the JPEG, > > > > > it knows the width and height and can request > > > > > buffers accordingly. > > > > > > > > > > The hardware is stateless, and so is the driver. > > > > > > > > > > On the other hand, supporting JPEG would mean that > > > > > drivers will have to parse the image, extracting > > > > > the quantization and entropy tables. > > > > > > > > > > Format negotiation is now more complex, > > > > > either we follow the stateful spec, introducing a little > > > > > state machine in the driver... or we use the Request API, > > > > > but that means parsing on both sides kernel and application. > > > > > > > > > > Either way, using JPEG_RAW is just waaay simpler and puts > > > > > things where they belong. > > > > > > > > > > > > > As discussed on IRC, I'm sending a v7 for this series, > > > > fixing only the checkpatch issues and the extra line in the > > > > binding document. > > > > > > > > We've agreed to move forward with JPEG_RAW, for the reasons > > > > stated above. > > > > > > > > Plus, we've agreed to keep this in staging until the userspace > > > > support for JPEG_RAW format is clear. > > > > > > The problem is that a new format is actually a V4L2 core change, and > > > not something for staging. > > > > > > IMHO, it is prudent to wait for an userspace code before hushing > > > merging this upstream. > > > > > > - > > > > > > I'm concerned about adding a JPEG_RAW format. We had this discussion > > > in the past. On that time, it was opted to add a code inside the > > > drivers that will be adding the jpeg headers, on the cases where > > > the hardware was providing headerless frames. > > > > > > > I suspect such discussion was held on completely different hardware. > > What we are trying to support here has specific needs, and thus > > why we need new formats. > > > > > Among the reasons, different drivers may be requiring different > > > logic to add the jpeg headers, depending on how they interpret > > > the quality parameters and on the Huffmann tables used by an > > > specific codec. > > > > > > I can't see why this couldn't be done on a stateless codec. I mean: > > > in the specific case of JPEG, each frame is independently coded. > > > It doesn't need to store anything from a previous frame in order to > > > produce a JPEG header. > > > > > > > Well, this can be done. But since it's the wrong architecture, > > it will bring a more complex API and more complex implementation. > > I'm not saying that this is a final solution. > > We're talking about a driver that it is intended to be merged > (for now) at staging. "Polluting" V4L2 core with a new almost identical > format that has a potential of bringing problems due to a staging > driver is something that doesn't sound right. > > Before changing the core, we need to be able to test your proposal on > more drivers, on both directions (capture and output). > > > In the case of a stateless decoder, it will mean a really cumbersome > > API. As I have explained above, using JPEG format on hardware > > than can only deal with JPEG_RAW format means: > > > > 1) Userspace sets JPEG format on the OUTPUT queue (sink) > > I'm assuming that you're actually talking about a M2M device whose > output stream receives a JPEG frame, and its capture stream outputs > something like RGB, right? > Right, RGB or more likely YUVish. > Also, I'm assuming the description you gave for the "raw JPEG" as a > JPEG image without JPEG headers. > > > 2) Userspace needs to know the raw format, but it has > > to wait for the driver to parse it first. > > Why would userspace need to know the raw format? It doesn't really > matter if the output is raw JPEG or full JPEG. > > For the output stream scenario: > > 1) for a headerless-JPEG (raw JPEG is really a crappy name): > > Userspace will use the headers to extract some information useful > for the decoder to work, passing them via controls. Then, it will > queue a buffer. Request API is needed, as the parameters will change > frame by frame. > > 2) for "true" JPEG (e. g. a jpeg frame that follows the standard and > has headers): > > Kernelspace will parse the jpeg headers, extracting the parameters > they need. No need to use request API, as each frame will already > contain the needed metadata. > > No matter if (1) or (2) is used, a M2M device will receive one JPEG > frame, decode it and produce one RGB frame. No need to store any > state. > > The opposite would equally work, and it is identical to what we > have already in kernel, for the "true" JPEG output. > > A M2M device whose output stream is RGB and capture is MPEG: > > 1) for "true" JPEG: > > The driver will use a Huffmann table based on a quality parameter > (V4L2_CID_JPEG_COMPRESSION_QUALITY). Such table would either > be hardcoded at the driver or at the codec's firmware. The > output stream will contain the headers. > > 1) for a headerless-JPEG: > > The driver will use a Huffmann table based on a quality parameter > (V4L2_CID_JPEG_COMPRESSION_QUALITY). Such table would either > be hardcoded at the driver or at the codec's firmware. The > output stream will not contain the headers. > > Userspace will need to figure out what it was used by the Kernel > in order to add the JPEG header. I'm not sure if we can do it > on a transparent way for all devices. > > In any case, one input produces one output. > > > > > So, 3) it will QBUF a JPEG buffer, and 4) get the V4L2_EVENT_SOURCE_CHANGE > > event in order to move forward. This implies a stateful machinery in the > > driver. Like I said, more complex API, more complex implementation... > > and only to avoid a new pixelformat? Doesn't sound a good deal :-) > > I really can't understand what you're meaning by that... I can't > see any need to generate an event to move forward. > It is needed for userspace to negotiate the format on the CAPTURE queue. See for instance: https://github.com/torvalds/linux/blob/master/drivers/media/platform/mtk-jpeg/mtk_jpeg_core.c#L692 Which shows how the driver needs to have an internal state, and how it needs to produce an event on the .buf_queue. This sequence is specified in the stateful codec spec, and is not needed by stateless codecs such as this one. > > > > In comparison, a proper stateless decoder driver will expose > > a simpler API, provide a 1:1 output/capture buffer relationship, > > and expose the full capabilities to userspace. > > > > > So, from my side, at least for the first version of this driver, > > > please make it produce a full JPEG frame with the headers, and > > > use V4L2_PIX_FMT_JPEG format. > > > > > > > This means using some form of JPEG quantization tables and huffman > > tables in the kernel, such as in drivers/media/usb/gspca/topro.c. > > > > Really doesn't sound like the best choice, violating the > > mechanism/policy separation, as now the kernel is deciding > > the tables for the user. > > What the Huffmann tables has to do with it? No matter if the > JPEG has headers or not, the codec should be doing RLE > encoding/decoding. For that, it needs to know the Huffmann tables. > > Anyway, you can use a control to pass huffmann tables at runtime, > using the request API, no matter if JPEG has headers or not. > > > The hardware can take any quantization table, and allowing userspace > > flexible usage, but we are going to restrict that to just a specific > > usage and a specific set of tables... again, only to avoid a new > > pixelformat? > > Same here: passing quantization tables to the driver has nothing > to do with having headers or not. > > For decoding, having headers is actually better, as the table > will be at the JPEG header. For encoding, a V4L2 array control. > could be used. > > > > That should allow us to add this driver on staging without adding > > > a new JPEG_RAW format. > > > > > > > AFAIK, Rockchip and Allwinner JPEG codecs handle so-called JPEG_RAW > > format, i.e. the payload of a SOF segment. > > And so, almost all other JPEG devices we have already. > This is not true. JPEG decoder drivers currently supported, such as coda, mtk-jpeg and similar, seem to pass the JPEG frame (with all its headers) to the hardware. Rockhip and Allwinner are different, where the hardware only takes the SOF payload. We are adding a new pixelformat, because these devices handle a different format. > > I think having these > > two cases should be enough reason to add a new format. > > I'm not against adding it. Just against adding it without further > discussions, in a way that it would potentially cause troubles for > userspace to encode/decode JPEG. > > > > If you still think we'll need a V4L2_PIX_FMT_JPEG_RAW, I'd like > > > to merge it by the time we also have a decoder for libv4l. I > > > would also like to see, at the Kernel patch series, the addition > > > of such support for already existing non-staging camera > > > drivers that internally receive headless JPEG headers from the > > > hardware. This will allow us to test the userspace code with > > > different models and proof that a generic JPEG_RAW decoder on > > > userspace is doable. > > > > > > > We can add a libv4l2 plugin for this decoder if that helps moving > > forward. Although I suspect it will just fall into oblivion, > > as chromeos and android will do their own thing. > > I don't have any control what people do on forked Kernels. > > > > > JPEG is a very well known format, and it's trivial for applications > > to add the headers. > > With the current documentation you added at the V4L2 core patch, > I seriously doubt that it would be possible to do it in a way that > it would work with different vendors. > > Also, I don't know any current V4L2 generic application that would > do that. Having it at libv4l would allow those apps to work with that. > OK, perhaps we can try adding the libv4l2 plugin in a v8 submission. Regards, Eze