Message ID | 20190827092339.8858-12-jacopo@jmondi.org (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | media: Report camera sensor properties | expand |
On 8/27/19 11:23 AM, Jacopo Mondi wrote: > Report the native pixel array size and the crop bounds for the ov5670 > sensor driver. > > Signed-off-by: Jacopo Mondi <jacopo@jmondi.org> > --- > drivers/media/i2c/ov5670.c | 20 ++++++++++++++++++++ > 1 file changed, 20 insertions(+) > > diff --git a/drivers/media/i2c/ov5670.c b/drivers/media/i2c/ov5670.c > index 2bc57e85f721..3e22fe9ccad1 100644 > --- a/drivers/media/i2c/ov5670.c > +++ b/drivers/media/i2c/ov5670.c > @@ -2258,6 +2258,25 @@ static int ov5670_set_pad_format(struct v4l2_subdev *sd, > return 0; > } > > +static int ov5670_get_selection(struct v4l2_subdev *sd, > + struct v4l2_subdev_pad_config *cfg, > + struct v4l2_subdev_selection *sel) > +{ > + switch (sel->target) { > + case V4L2_SEL_TGT_CROP_BOUNDS: > + case V4L2_SEL_TGT_NATIVE_SIZE: > + sel->r.left = 0; > + sel->r.top = 0; > + sel->r.width = 2592; > + sel->r.height = 1944; Why do you need this? Since the format can change for this and the next driver I think CROP_BOUNDS at least should match the current format. I don't think this patch and the next have anything to do with the location/rotate support. I would split it off from this series. Regards, Hans > + break; > + default: > + return -EINVAL; > + } > + > + return 0; > +} > + > static int ov5670_get_skip_frames(struct v4l2_subdev *sd, u32 *frames) > { > *frames = OV5670_NUM_OF_SKIP_FRAMES; > @@ -2425,6 +2444,7 @@ static const struct v4l2_subdev_pad_ops ov5670_pad_ops = { > .enum_mbus_code = ov5670_enum_mbus_code, > .get_fmt = ov5670_get_pad_format, > .set_fmt = ov5670_set_pad_format, > + .get_selection = ov5670_get_selection, > .enum_frame_size = ov5670_enum_frame_size, > }; > >
HI Hans, On Thu, Aug 29, 2019 at 12:20:18PM +0200, Hans Verkuil wrote: > On 8/27/19 11:23 AM, Jacopo Mondi wrote: > > Report the native pixel array size and the crop bounds for the ov5670 > > sensor driver. > > > > Signed-off-by: Jacopo Mondi <jacopo@jmondi.org> > > --- > > drivers/media/i2c/ov5670.c | 20 ++++++++++++++++++++ > > 1 file changed, 20 insertions(+) > > > > diff --git a/drivers/media/i2c/ov5670.c b/drivers/media/i2c/ov5670.c > > index 2bc57e85f721..3e22fe9ccad1 100644 > > --- a/drivers/media/i2c/ov5670.c > > +++ b/drivers/media/i2c/ov5670.c > > @@ -2258,6 +2258,25 @@ static int ov5670_set_pad_format(struct v4l2_subdev *sd, > > return 0; > > } > > > > +static int ov5670_get_selection(struct v4l2_subdev *sd, > > + struct v4l2_subdev_pad_config *cfg, > > + struct v4l2_subdev_selection *sel) > > +{ > > + switch (sel->target) { > > + case V4L2_SEL_TGT_CROP_BOUNDS: > > + case V4L2_SEL_TGT_NATIVE_SIZE: > > + sel->r.left = 0; > > + sel->r.top = 0; > > + sel->r.width = 2592; > > + sel->r.height = 1944; > > Why do you need this? > I need to expose the pixel array size and the active pixel area size, and I interpreted the two above targets as the right place where to do so. At least for NATIVE_SIZE the documentation seems to match my understanding: "The native size of the device, e.g. a sensor’s pixel array. left and top fields are zero for this target." > Since the format can change for this and the next driver I think CROP_BOUNDS > at least should match the current format. > Ah, does it? It was not clear to me from the documentation, as it suggested to me that target actually identifies the active pixel area "Bounds of the crop rectangle. All valid crop rectangles fit inside the crop bounds rectangle." It does not mention format, should this be updated? How would you report the two information I need? > I don't think this patch and the next have anything to do with the location/rotate > support. I would split it off from this series. > Agreed, they were split in v1, maybe it has not been a wise move to make a single series out of them. I'll split again. Thanks j > Regards, > > Hans > > > + break; > > + default: > > + return -EINVAL; > > + } > > + > > + return 0; > > +} > > + > > static int ov5670_get_skip_frames(struct v4l2_subdev *sd, u32 *frames) > > { > > *frames = OV5670_NUM_OF_SKIP_FRAMES; > > @@ -2425,6 +2444,7 @@ static const struct v4l2_subdev_pad_ops ov5670_pad_ops = { > > .enum_mbus_code = ov5670_enum_mbus_code, > > .get_fmt = ov5670_get_pad_format, > > .set_fmt = ov5670_set_pad_format, > > + .get_selection = ov5670_get_selection, > > .enum_frame_size = ov5670_enum_frame_size, > > }; > > > > >
On 8/29/19 2:40 PM, Jacopo Mondi wrote: > HI Hans, > > On Thu, Aug 29, 2019 at 12:20:18PM +0200, Hans Verkuil wrote: >> On 8/27/19 11:23 AM, Jacopo Mondi wrote: >>> Report the native pixel array size and the crop bounds for the ov5670 >>> sensor driver. >>> >>> Signed-off-by: Jacopo Mondi <jacopo@jmondi.org> >>> --- >>> drivers/media/i2c/ov5670.c | 20 ++++++++++++++++++++ >>> 1 file changed, 20 insertions(+) >>> >>> diff --git a/drivers/media/i2c/ov5670.c b/drivers/media/i2c/ov5670.c >>> index 2bc57e85f721..3e22fe9ccad1 100644 >>> --- a/drivers/media/i2c/ov5670.c >>> +++ b/drivers/media/i2c/ov5670.c >>> @@ -2258,6 +2258,25 @@ static int ov5670_set_pad_format(struct v4l2_subdev *sd, >>> return 0; >>> } >>> >>> +static int ov5670_get_selection(struct v4l2_subdev *sd, >>> + struct v4l2_subdev_pad_config *cfg, >>> + struct v4l2_subdev_selection *sel) >>> +{ >>> + switch (sel->target) { >>> + case V4L2_SEL_TGT_CROP_BOUNDS: >>> + case V4L2_SEL_TGT_NATIVE_SIZE: >>> + sel->r.left = 0; >>> + sel->r.top = 0; >>> + sel->r.width = 2592; >>> + sel->r.height = 1944; >> >> Why do you need this? >> > > I need to expose the pixel array size and the active pixel area size, > and I interpreted the two above targets as the right place where to do > so. That didn't answer my question :-) Why do you need to expose this? Who uses it for what purpose? > > At least for NATIVE_SIZE the documentation seems to match my > understanding: > > "The native size of the device, e.g. a sensor’s pixel array. left and top > fields are zero for this target." Correct. > > >> Since the format can change for this and the next driver I think CROP_BOUNDS >> at least should match the current format. >> > > Ah, does it? It was not clear to me from the documentation, as it > suggested to me that target actually identifies the active pixel area > > "Bounds of the crop rectangle. All valid crop rectangles fit inside the > crop bounds rectangle." > > It does not mention format, should this be updated? The problem is that for sensors it is indeed not clear. In principle the crop bounds matches the resolution that the sensor uses pre-scaling. So yes, that means that it is equal to the native size. But many sensors have a discrete list of supported formats and it is not clear how they map each format to the actual sensor. If it is clearly just done through binning or averaging, then all is fine. If the formats have different aspect ratios, then it becomes a bit more difficult how to relate the crop bounds with the format since you may not know to which sensor area the current format corresponds. I have no clear answer for you, to be honest, but it can get tricky. > > How would you report the two information I need? It depends on my original question: what do you need this information for? Regards, Hans > >> I don't think this patch and the next have anything to do with the location/rotate >> support. I would split it off from this series. >> > > Agreed, they were split in v1, maybe it has not been a wise move to > make a single series out of them. I'll split again. > > Thanks > j > >> Regards, >> >> Hans >> >>> + break; >>> + default: >>> + return -EINVAL; >>> + } >>> + >>> + return 0; >>> +} >>> + >>> static int ov5670_get_skip_frames(struct v4l2_subdev *sd, u32 *frames) >>> { >>> *frames = OV5670_NUM_OF_SKIP_FRAMES; >>> @@ -2425,6 +2444,7 @@ static const struct v4l2_subdev_pad_ops ov5670_pad_ops = { >>> .enum_mbus_code = ov5670_enum_mbus_code, >>> .get_fmt = ov5670_get_pad_format, >>> .set_fmt = ov5670_set_pad_format, >>> + .get_selection = ov5670_get_selection, >>> .enum_frame_size = ov5670_enum_frame_size, >>> }; >>> >>> >>
Hi Hans On Thu, Aug 29, 2019 at 02:55:30PM +0200, Hans Verkuil wrote: > On 8/29/19 2:40 PM, Jacopo Mondi wrote: > > HI Hans, > > > > On Thu, Aug 29, 2019 at 12:20:18PM +0200, Hans Verkuil wrote: > >> On 8/27/19 11:23 AM, Jacopo Mondi wrote: > >>> Report the native pixel array size and the crop bounds for the ov5670 > >>> sensor driver. > >>> > >>> Signed-off-by: Jacopo Mondi <jacopo@jmondi.org> > >>> --- > >>> drivers/media/i2c/ov5670.c | 20 ++++++++++++++++++++ > >>> 1 file changed, 20 insertions(+) > >>> > >>> diff --git a/drivers/media/i2c/ov5670.c b/drivers/media/i2c/ov5670.c > >>> index 2bc57e85f721..3e22fe9ccad1 100644 > >>> --- a/drivers/media/i2c/ov5670.c > >>> +++ b/drivers/media/i2c/ov5670.c > >>> @@ -2258,6 +2258,25 @@ static int ov5670_set_pad_format(struct v4l2_subdev *sd, > >>> return 0; > >>> } > >>> > >>> +static int ov5670_get_selection(struct v4l2_subdev *sd, > >>> + struct v4l2_subdev_pad_config *cfg, > >>> + struct v4l2_subdev_selection *sel) > >>> +{ > >>> + switch (sel->target) { > >>> + case V4L2_SEL_TGT_CROP_BOUNDS: > >>> + case V4L2_SEL_TGT_NATIVE_SIZE: > >>> + sel->r.left = 0; > >>> + sel->r.top = 0; > >>> + sel->r.width = 2592; > >>> + sel->r.height = 1944; > >> > >> Why do you need this? > >> > > > > I need to expose the pixel array size and the active pixel area size, > > and I interpreted the two above targets as the right place where to do > > so. > > That didn't answer my question :-) > > Why do you need to expose this? Who uses it for what purpose? > Ah, ok :) Quoting the cover letter of the series: Retrieving the following camera static information is a requirement for the implementation of the Android-compatiblity layer of libcamera, but I'm sure it might prove useful for other user-space applications and libraries as well. In other words, in order to report to the android camera stack those two information (among -many- others) they should be somehow retrieved from the kernel interface, and after considering adding two more read-only controls to expose them as I did for rotation and location, I went for using the selection API. > > > > At least for NATIVE_SIZE the documentation seems to match my > > understanding: > > > > "The native size of the device, e.g. a sensor’s pixel array. left and top > > fields are zero for this target." > > Correct. > > > > > > >> Since the format can change for this and the next driver I think CROP_BOUNDS > >> at least should match the current format. > >> > > > > Ah, does it? It was not clear to me from the documentation, as it > > suggested to me that target actually identifies the active pixel area > > > > "Bounds of the crop rectangle. All valid crop rectangles fit inside the > > crop bounds rectangle." > > > > It does not mention format, should this be updated? > > The problem is that for sensors it is indeed not clear. > > In principle the crop bounds matches the resolution that the sensor uses > pre-scaling. So yes, that means that it is equal to the native size. native size != active areas size. Did you mean the latter here? > > But many sensors have a discrete list of supported formats and it is not > clear how they map each format to the actual sensor. If it is clearly just > done through binning or averaging, then all is fine. > > If the formats have different aspect ratios, then it becomes a bit more > difficult how to relate the crop bounds with the format since you may not > know to which sensor area the current format corresponds. I see. We've had the same discussion in the libcamera list, as indeed the crop bounds might depend on the aspect ratio as you said. Maybe the crop_bound target is not the right place to report the active area, or maybe the concept of active area is ill-defined if the sensor is cross-shaped: /--------------------/ -> Pixel array size ---------------------- |x | |--- x|------> 4:3 crop bounds |x------------------x| |x| | | |x| |x| | | |x|------> 21:9 crop bounds |x| | | |x| |x------------------x| |x |________| x| ---------------------- | | ---------------------------> Black pixels colums > > I have no clear answer for you, to be honest, but it can get tricky. > Indeed, but assuming a simpler square-shaped sensor, is the crop bound target the right one to report the active pixel area ? > > > > How would you report the two information I need? > > It depends on my original question: what do you need this information for? > Please note that it's for the android camera stack in this case, but it's an information that userspace might want to know for several different reasons. Calibration and FOV calculation come to mind. Does this makes sense to you? Thanks j > Regards, > > Hans > > > > >> I don't think this patch and the next have anything to do with the location/rotate > >> support. I would split it off from this series. > >> > > > > Agreed, they were split in v1, maybe it has not been a wise move to > > make a single series out of them. I'll split again. > > > > Thanks > > j > > > >> Regards, > >> > >> Hans > >> > >>> + break; > >>> + default: > >>> + return -EINVAL; > >>> + } > >>> + > >>> + return 0; > >>> +} > >>> + > >>> static int ov5670_get_skip_frames(struct v4l2_subdev *sd, u32 *frames) > >>> { > >>> *frames = OV5670_NUM_OF_SKIP_FRAMES; > >>> @@ -2425,6 +2444,7 @@ static const struct v4l2_subdev_pad_ops ov5670_pad_ops = { > >>> .enum_mbus_code = ov5670_enum_mbus_code, > >>> .get_fmt = ov5670_get_pad_format, > >>> .set_fmt = ov5670_set_pad_format, > >>> + .get_selection = ov5670_get_selection, > >>> .enum_frame_size = ov5670_enum_frame_size, > >>> }; > >>> > >>> > >> >
On 8/29/19 3:18 PM, Jacopo Mondi wrote: > Hi Hans > > On Thu, Aug 29, 2019 at 02:55:30PM +0200, Hans Verkuil wrote: >> On 8/29/19 2:40 PM, Jacopo Mondi wrote: >>> HI Hans, >>> >>> On Thu, Aug 29, 2019 at 12:20:18PM +0200, Hans Verkuil wrote: >>>> On 8/27/19 11:23 AM, Jacopo Mondi wrote: >>>>> Report the native pixel array size and the crop bounds for the ov5670 >>>>> sensor driver. >>>>> >>>>> Signed-off-by: Jacopo Mondi <jacopo@jmondi.org> >>>>> --- >>>>> drivers/media/i2c/ov5670.c | 20 ++++++++++++++++++++ >>>>> 1 file changed, 20 insertions(+) >>>>> >>>>> diff --git a/drivers/media/i2c/ov5670.c b/drivers/media/i2c/ov5670.c >>>>> index 2bc57e85f721..3e22fe9ccad1 100644 >>>>> --- a/drivers/media/i2c/ov5670.c >>>>> +++ b/drivers/media/i2c/ov5670.c >>>>> @@ -2258,6 +2258,25 @@ static int ov5670_set_pad_format(struct v4l2_subdev *sd, >>>>> return 0; >>>>> } >>>>> >>>>> +static int ov5670_get_selection(struct v4l2_subdev *sd, >>>>> + struct v4l2_subdev_pad_config *cfg, >>>>> + struct v4l2_subdev_selection *sel) >>>>> +{ >>>>> + switch (sel->target) { >>>>> + case V4L2_SEL_TGT_CROP_BOUNDS: >>>>> + case V4L2_SEL_TGT_NATIVE_SIZE: >>>>> + sel->r.left = 0; >>>>> + sel->r.top = 0; >>>>> + sel->r.width = 2592; >>>>> + sel->r.height = 1944; >>>> >>>> Why do you need this? >>>> >>> >>> I need to expose the pixel array size and the active pixel area size, >>> and I interpreted the two above targets as the right place where to do >>> so. >> >> That didn't answer my question :-) >> >> Why do you need to expose this? Who uses it for what purpose? >> > > Ah, ok :) > > Quoting the cover letter of the series: > > Retrieving the following camera static information is a requirement for the > implementation of the Android-compatiblity layer of libcamera, but I'm sure > it might prove useful for other user-space applications and libraries as well. > > In other words, in order to report to the android camera stack those > two information (among -many- others) they should be somehow retrieved > from the kernel interface, and after considering adding two more > read-only controls to expose them as I did for rotation and location, > I went for using the selection API. > >>> >>> At least for NATIVE_SIZE the documentation seems to match my >>> understanding: >>> >>> "The native size of the device, e.g. a sensor’s pixel array. left and top >>> fields are zero for this target." >> >> Correct. >> >>> >>> >>>> Since the format can change for this and the next driver I think CROP_BOUNDS >>>> at least should match the current format. >>>> >>> >>> Ah, does it? It was not clear to me from the documentation, as it >>> suggested to me that target actually identifies the active pixel area >>> >>> "Bounds of the crop rectangle. All valid crop rectangles fit inside the >>> crop bounds rectangle." >>> >>> It does not mention format, should this be updated? >> >> The problem is that for sensors it is indeed not clear. >> >> In principle the crop bounds matches the resolution that the sensor uses >> pre-scaling. So yes, that means that it is equal to the native size. > > native size != active areas size. > Did you mean the latter here? I mean the latter, but I have to say that the spec doesn't make it clear if TGT_NATIVE_SIZE equals the active area or also includes inactive pixels. What exactly does Android expect? > >> >> But many sensors have a discrete list of supported formats and it is not >> clear how they map each format to the actual sensor. If it is clearly just >> done through binning or averaging, then all is fine. >> >> If the formats have different aspect ratios, then it becomes a bit more >> difficult how to relate the crop bounds with the format since you may not >> know to which sensor area the current format corresponds. > > I see. We've had the same discussion in the libcamera list, as indeed > the crop bounds might depend on the aspect ratio as you said. > > Maybe the crop_bound target is not the right place to report the > active area, or maybe the concept of active area is ill-defined if > the sensor is cross-shaped: > > /--------------------/ -> Pixel array size > ---------------------- > |x | |--- x|------> 4:3 crop bounds > |x------------------x| > |x| | | |x| > |x| | | |x|------> 21:9 crop bounds > |x| | | |x| > |x------------------x| > |x |________| x| > ---------------------- > | | > ---------------------------> Black pixels colums > >> >> I have no clear answer for you, to be honest, but it can get tricky. >> > > Indeed, but assuming a simpler square-shaped sensor, is the crop bound > target the right one to report the active pixel area ? I don't think so. The crop bounds defines the outer bounds of the area where you can put your crop rectangle. This can very well include areas where the pixels are covered and so always return black (sometimes used to obtain black levels, I believe). The default crop rectangle would be the one that reports the active area. The native size rectangle would be the full pixel array. So CROP_DEFAULT <= CROP_BOUNDS <= NATIVE_SIZE. For a cross-shaped sensor I would expect that the CROP_BOUNDS/DEFAULT depends on the format (aspect ratio). I think this makes sense. The specification definitely needs to be improved, patches are welcome... > >>> >>> How would you report the two information I need? >> >> It depends on my original question: what do you need this information for? >> > > Please note that it's for the android camera stack in this case, but > it's an information that userspace might want to know for several > different reasons. Calibration and FOV calculation come to mind. Does > this makes sense to you? Ah, so that's what it is used for :-) Which of the three targets above would match with what Android needs? Regards, Hans > > Thanks > j > >> Regards, >> >> Hans >> >>> >>>> I don't think this patch and the next have anything to do with the location/rotate >>>> support. I would split it off from this series. >>>> >>> >>> Agreed, they were split in v1, maybe it has not been a wise move to >>> make a single series out of them. I'll split again. >>> >>> Thanks >>> j >>> >>>> Regards, >>>> >>>> Hans >>>> >>>>> + break; >>>>> + default: >>>>> + return -EINVAL; >>>>> + } >>>>> + >>>>> + return 0; >>>>> +} >>>>> + >>>>> static int ov5670_get_skip_frames(struct v4l2_subdev *sd, u32 *frames) >>>>> { >>>>> *frames = OV5670_NUM_OF_SKIP_FRAMES; >>>>> @@ -2425,6 +2444,7 @@ static const struct v4l2_subdev_pad_ops ov5670_pad_ops = { >>>>> .enum_mbus_code = ov5670_enum_mbus_code, >>>>> .get_fmt = ov5670_get_pad_format, >>>>> .set_fmt = ov5670_set_pad_format, >>>>> + .get_selection = ov5670_get_selection, >>>>> .enum_frame_size = ov5670_enum_frame_size, >>>>> }; >>>>> >>>>> >>>> >>
Hi Hans, On Thu, Aug 29, 2019 at 03:39:43PM +0200, Hans Verkuil wrote: > On 8/29/19 3:18 PM, Jacopo Mondi wrote: > > Hi Hans > > > > On Thu, Aug 29, 2019 at 02:55:30PM +0200, Hans Verkuil wrote: > >> On 8/29/19 2:40 PM, Jacopo Mondi wrote: > >>> HI Hans, > >>> > >>> On Thu, Aug 29, 2019 at 12:20:18PM +0200, Hans Verkuil wrote: > >>>> On 8/27/19 11:23 AM, Jacopo Mondi wrote: > >>>>> Report the native pixel array size and the crop bounds for the ov5670 > >>>>> sensor driver. > >>>>> > >>>>> Signed-off-by: Jacopo Mondi <jacopo@jmondi.org> > >>>>> --- > >>>>> drivers/media/i2c/ov5670.c | 20 ++++++++++++++++++++ > >>>>> 1 file changed, 20 insertions(+) > >>>>> > >>>>> diff --git a/drivers/media/i2c/ov5670.c b/drivers/media/i2c/ov5670.c > >>>>> index 2bc57e85f721..3e22fe9ccad1 100644 > >>>>> --- a/drivers/media/i2c/ov5670.c > >>>>> +++ b/drivers/media/i2c/ov5670.c > >>>>> @@ -2258,6 +2258,25 @@ static int ov5670_set_pad_format(struct v4l2_subdev *sd, > >>>>> return 0; > >>>>> } > >>>>> > >>>>> +static int ov5670_get_selection(struct v4l2_subdev *sd, > >>>>> + struct v4l2_subdev_pad_config *cfg, > >>>>> + struct v4l2_subdev_selection *sel) > >>>>> +{ > >>>>> + switch (sel->target) { > >>>>> + case V4L2_SEL_TGT_CROP_BOUNDS: > >>>>> + case V4L2_SEL_TGT_NATIVE_SIZE: > >>>>> + sel->r.left = 0; > >>>>> + sel->r.top = 0; > >>>>> + sel->r.width = 2592; > >>>>> + sel->r.height = 1944; > >>>> > >>>> Why do you need this? > >>>> > >>> > >>> I need to expose the pixel array size and the active pixel area size, > >>> and I interpreted the two above targets as the right place where to do > >>> so. > >> > >> That didn't answer my question :-) > >> > >> Why do you need to expose this? Who uses it for what purpose? > >> > > > > Ah, ok :) > > > > Quoting the cover letter of the series: > > > > Retrieving the following camera static information is a requirement for the > > implementation of the Android-compatiblity layer of libcamera, but I'm sure > > it might prove useful for other user-space applications and libraries as well. > > > > In other words, in order to report to the android camera stack those > > two information (among -many- others) they should be somehow retrieved > > from the kernel interface, and after considering adding two more > > read-only controls to expose them as I did for rotation and location, > > I went for using the selection API. > > > >>> > >>> At least for NATIVE_SIZE the documentation seems to match my > >>> understanding: > >>> > >>> "The native size of the device, e.g. a sensor’s pixel array. left and top > >>> fields are zero for this target." > >> > >> Correct. > >> > >>> > >>> > >>>> Since the format can change for this and the next driver I think CROP_BOUNDS > >>>> at least should match the current format. > >>>> > >>> > >>> Ah, does it? It was not clear to me from the documentation, as it > >>> suggested to me that target actually identifies the active pixel area > >>> > >>> "Bounds of the crop rectangle. All valid crop rectangles fit inside the > >>> crop bounds rectangle." > >>> > >>> It does not mention format, should this be updated? > >> > >> The problem is that for sensors it is indeed not clear. > >> > >> In principle the crop bounds matches the resolution that the sensor uses > >> pre-scaling. So yes, that means that it is equal to the native size. > > > > native size != active areas size. > > Did you mean the latter here? > > I mean the latter, but I have to say that the spec doesn't make it clear > if TGT_NATIVE_SIZE equals the active area or also includes inactive pixels. > > What exactly does Android expect? > Ah! You asked for this, so be prepared... I started by trying to retrieve from the kernel interface the following parameters: The pixel array size https://jmondi.org/android_metadata_tags/docs.html#static_android.sensor.info.pixelArraySize The physical dimensions of the sensor (TODO) https://jmondi.org/android_metadata_tags/docs.html#static_android.sensor.info.physicalSize The active area size: https://jmondi.org/android_metadata_tags/docs.html#static_android.sensor.info.activeArraySize Sensor orientation: https://jmondi.org/android_metadata_tags/docs.html#static_android.sensor.orientation Location: https://jmondi.org/android_metadata_tags/docs.html#static_android.lens.facing For very simple cameras there are not many more informations about sensor/lens to be retrieved, but if you look at the end of the page you can see a list of other informations Android requires for more advanced use case (not so advanced actually, those are expected to be there for any recent modern phone device I guess) It's a long list, with many details about the sensor, lens and the capabilities of the camera. Just to provide an example list: android.lens.info.availableApertures (static) android.lens.info.availableFilterDensities (static) android.lens.info.availableFocalLengths (static) android.lens.info.availableOpticalStabilization (static) android.lens.info.minimumFocusDistance (static) android.lens.info.shadingMapSize (static) android.lens.info.focusDistanceCalibration (static) android.sensor.info.sensitivityRange (static) android.sensor.info.exposureTimeRange (static) android.sensor.info.maxFrameDuration (static) android.sensor.info.physicalSize (static) android.sensor.info.timestampSource (static) android.sensor.maxAnalogSensitivity (static) And they grow up to the point of describing the color transformation matrices and other low level details, if your device claim to support raw capture and manipulation: android.sensor.info.colorFilterArrangement (static) android.sensor.info.whiteLevel (static) android.sensor.info.preCorrectionActiveArraySize (static) android.sensor.referenceIlluminant1 (static) android.sensor.referenceIlluminant2 (static) android.sensor.calibrationTransform1 (static) android.sensor.calibrationTransform2 (static) android.sensor.colorTransform1 (static) android.sensor.colorTransform2 (static) android.sensor.forwardMatrix1 (static) android.sensor.forwardMatrix2 (static) android.sensor.blackLevelPattern (static) android.sensor.profileHueSatMapDimensions (static) So, going forward I cannot tell how much of this could actually come from V4L, ideally all of this, but as of today it's not possible, so vendors keep adding custom solutions to support this with out-of-tree drivers or find solutions to generate these information from configuration files in user space. For libcamera this is not possible, it aims to build on top of standard v4l2 interfaces, so I've started the process of adding small bits to v4l2 to at least retrieve the most basic information. > > > >> > >> But many sensors have a discrete list of supported formats and it is not > >> clear how they map each format to the actual sensor. If it is clearly just > >> done through binning or averaging, then all is fine. > >> > >> If the formats have different aspect ratios, then it becomes a bit more > >> difficult how to relate the crop bounds with the format since you may not > >> know to which sensor area the current format corresponds. > > > > I see. We've had the same discussion in the libcamera list, as indeed > > the crop bounds might depend on the aspect ratio as you said. > > > > Maybe the crop_bound target is not the right place to report the > > active area, or maybe the concept of active area is ill-defined if > > the sensor is cross-shaped: > > > > /--------------------/ -> Pixel array size > > ---------------------- > > |x | |--- x|------> 4:3 crop bounds > > |x------------------x| > > |x| | | |x| > > |x| | | |x|------> 21:9 crop bounds > > |x| | | |x| > > |x------------------x| > > |x |________| x| > > ---------------------- > > | | > > ---------------------------> Black pixels colums > > > >> > >> I have no clear answer for you, to be honest, but it can get tricky. > >> > > > > Indeed, but assuming a simpler square-shaped sensor, is the crop bound > > target the right one to report the active pixel area ? > > I don't think so. > > The crop bounds defines the outer bounds of the area where you can put > your crop rectangle. This can very well include areas where the pixels > are covered and so always return black (sometimes used to obtain black > levels, I believe). Ah, I didn't considered black pixels as part of a possible crop selection.. > > The default crop rectangle would be the one that reports the active > area. The native size rectangle would be the full pixel array. > > So CROP_DEFAULT <= CROP_BOUNDS <= NATIVE_SIZE. > I see thanks for clarifying this > For a cross-shaped sensor I would expect that the CROP_BOUNDS/DEFAULT > depends on the format (aspect ratio). Currently the selection API do not support to retrieve that information depending on the aspect-ratio though > > I think this makes sense. > > The specification definitely needs to be improved, patches are welcome... > > > > >>> > >>> How would you report the two information I need? > >> > >> It depends on my original question: what do you need this information for? > >> > > > > Please note that it's for the android camera stack in this case, but > > it's an information that userspace might want to know for several > > different reasons. Calibration and FOV calculation come to mind. Does > > this makes sense to you? > > Ah, so that's what it is used for :-) > > Which of the three targets above would match with what Android needs? > the full pixel array size would be reported by the NATIVE_SIZE target the active area size by the CROP_DEFAULT rectangle (regardless of the aspect ratio) It's still not clear to me what the CROP_BOUNDS rectangle would then be used for, unless it is made to be dependent on an aspect ratio but that would require userspace to provide a populated rectangle to G_SELECTION and drivers to inspect it and return an opportune CROP_BOUNDS one. Would this be acceptable ? Thanks j > Regards, > > Hans > > > > > Thanks > > j > > > >> Regards, > >> > >> Hans > >> > >>> > >>>> I don't think this patch and the next have anything to do with the location/rotate > >>>> support. I would split it off from this series. > >>>> > >>> > >>> Agreed, they were split in v1, maybe it has not been a wise move to > >>> make a single series out of them. I'll split again. > >>> > >>> Thanks > >>> j > >>> > >>>> Regards, > >>>> > >>>> Hans > >>>> > >>>>> + break; > >>>>> + default: > >>>>> + return -EINVAL; > >>>>> + } > >>>>> + > >>>>> + return 0; > >>>>> +} > >>>>> + > >>>>> static int ov5670_get_skip_frames(struct v4l2_subdev *sd, u32 *frames) > >>>>> { > >>>>> *frames = OV5670_NUM_OF_SKIP_FRAMES; > >>>>> @@ -2425,6 +2444,7 @@ static const struct v4l2_subdev_pad_ops ov5670_pad_ops = { > >>>>> .enum_mbus_code = ov5670_enum_mbus_code, > >>>>> .get_fmt = ov5670_get_pad_format, > >>>>> .set_fmt = ov5670_set_pad_format, > >>>>> + .get_selection = ov5670_get_selection, > >>>>> .enum_frame_size = ov5670_enum_frame_size, > >>>>> }; > >>>>> > >>>>> > >>>> > >> >
Hi Jacopo, On Thu, Aug 29, 2019 at 06:43:30PM +0200, Jacopo Mondi wrote: > On Thu, Aug 29, 2019 at 03:39:43PM +0200, Hans Verkuil wrote: > > On 8/29/19 3:18 PM, Jacopo Mondi wrote: > > > On Thu, Aug 29, 2019 at 02:55:30PM +0200, Hans Verkuil wrote: > > >> On 8/29/19 2:40 PM, Jacopo Mondi wrote: > > >>> On Thu, Aug 29, 2019 at 12:20:18PM +0200, Hans Verkuil wrote: > > >>>> On 8/27/19 11:23 AM, Jacopo Mondi wrote: > > >>>>> Report the native pixel array size and the crop bounds for the ov5670 > > >>>>> sensor driver. > > >>>>> > > >>>>> Signed-off-by: Jacopo Mondi <jacopo@jmondi.org> > > >>>>> --- > > >>>>> drivers/media/i2c/ov5670.c | 20 ++++++++++++++++++++ > > >>>>> 1 file changed, 20 insertions(+) > > >>>>> > > >>>>> diff --git a/drivers/media/i2c/ov5670.c b/drivers/media/i2c/ov5670.c > > >>>>> index 2bc57e85f721..3e22fe9ccad1 100644 > > >>>>> --- a/drivers/media/i2c/ov5670.c > > >>>>> +++ b/drivers/media/i2c/ov5670.c > > >>>>> @@ -2258,6 +2258,25 @@ static int ov5670_set_pad_format(struct v4l2_subdev *sd, > > >>>>> return 0; > > >>>>> } > > >>>>> > > >>>>> +static int ov5670_get_selection(struct v4l2_subdev *sd, > > >>>>> + struct v4l2_subdev_pad_config *cfg, > > >>>>> + struct v4l2_subdev_selection *sel) > > >>>>> +{ > > >>>>> + switch (sel->target) { > > >>>>> + case V4L2_SEL_TGT_CROP_BOUNDS: > > >>>>> + case V4L2_SEL_TGT_NATIVE_SIZE: > > >>>>> + sel->r.left = 0; > > >>>>> + sel->r.top = 0; > > >>>>> + sel->r.width = 2592; > > >>>>> + sel->r.height = 1944; > > >>>> > > >>>> Why do you need this? > > >>>> > > >>> > > >>> I need to expose the pixel array size and the active pixel area size, > > >>> and I interpreted the two above targets as the right place where to do > > >>> so. > > >> > > >> That didn't answer my question :-) > > >> > > >> Why do you need to expose this? Who uses it for what purpose? > > >> > > > > > > Ah, ok :) > > > > > > Quoting the cover letter of the series: > > > > > > Retrieving the following camera static information is a requirement for the > > > implementation of the Android-compatiblity layer of libcamera, but I'm sure > > > it might prove useful for other user-space applications and libraries as well. > > > > > > In other words, in order to report to the android camera stack those > > > two information (among -many- others) they should be somehow retrieved > > > from the kernel interface, and after considering adding two more > > > read-only controls to expose them as I did for rotation and location, > > > I went for using the selection API. > > > > > >>> > > >>> At least for NATIVE_SIZE the documentation seems to match my > > >>> understanding: > > >>> > > >>> "The native size of the device, e.g. a sensor’s pixel array. left and top > > >>> fields are zero for this target." > > >> > > >> Correct. > > >> > > >>> > > >>> > > >>>> Since the format can change for this and the next driver I think CROP_BOUNDS > > >>>> at least should match the current format. > > >>>> > > >>> > > >>> Ah, does it? It was not clear to me from the documentation, as it > > >>> suggested to me that target actually identifies the active pixel area > > >>> > > >>> "Bounds of the crop rectangle. All valid crop rectangles fit inside the > > >>> crop bounds rectangle." > > >>> > > >>> It does not mention format, should this be updated? > > >> > > >> The problem is that for sensors it is indeed not clear. > > >> > > >> In principle the crop bounds matches the resolution that the sensor uses > > >> pre-scaling. So yes, that means that it is equal to the native size. > > > > > > native size != active areas size. > > > Did you mean the latter here? > > > > I mean the latter, but I have to say that the spec doesn't make it clear > > if TGT_NATIVE_SIZE equals the active area or also includes inactive pixels. > > > > What exactly does Android expect? > > > Ah! You asked for this, so be prepared... > > I started by trying to retrieve from the kernel interface the > following parameters: > > The pixel array size > https://jmondi.org/android_metadata_tags/docs.html#static_android.sensor.info.pixelArraySize > The physical dimensions of the sensor (TODO) > https://jmondi.org/android_metadata_tags/docs.html#static_android.sensor.info.physicalSize > The active area size: > https://jmondi.org/android_metadata_tags/docs.html#static_android.sensor.info.activeArraySize > Sensor orientation: > https://jmondi.org/android_metadata_tags/docs.html#static_android.sensor.orientation > Location: > https://jmondi.org/android_metadata_tags/docs.html#static_android.lens.facing > > For very simple cameras there are not many more informations about > sensor/lens to be retrieved, but if you look at the end of the page you can > see a list of other informations Android requires for more advanced use case > (not so advanced actually, those are expected to be there for any > recent modern phone device I guess) > > It's a long list, with many details about the sensor, lens and the > capabilities of the camera. > > Just to provide an example list: > android.lens.info.availableApertures (static) > android.lens.info.availableFilterDensities (static) > android.lens.info.availableFocalLengths (static) > android.lens.info.availableOpticalStabilization (static) > android.lens.info.minimumFocusDistance (static) > android.lens.info.shadingMapSize (static) > android.lens.info.focusDistanceCalibration (static) > android.sensor.info.sensitivityRange (static) > android.sensor.info.exposureTimeRange (static) > android.sensor.info.maxFrameDuration (static) > android.sensor.info.physicalSize (static) > android.sensor.info.timestampSource (static) > android.sensor.maxAnalogSensitivity (static) > > And they grow up to the point of describing the color transformation > matrices and other low level details, if your device claim to support > raw capture and manipulation: > android.sensor.info.colorFilterArrangement (static) > android.sensor.info.whiteLevel (static) > android.sensor.info.preCorrectionActiveArraySize (static) > android.sensor.referenceIlluminant1 (static) > android.sensor.referenceIlluminant2 (static) > android.sensor.calibrationTransform1 (static) > android.sensor.calibrationTransform2 (static) > android.sensor.colorTransform1 (static) > android.sensor.colorTransform2 (static) > android.sensor.forwardMatrix1 (static) > android.sensor.forwardMatrix2 (static) > android.sensor.blackLevelPattern (static) > android.sensor.profileHueSatMapDimensions (static) > > So, going forward I cannot tell how much of this could actually come > from V4L, ideally all of this, but as of today it's not possible, so > vendors keep adding custom solutions to support this with out-of-tree > drivers or find solutions to generate these information from > configuration files in user space. > > For libcamera this is not possible, it aims to build on top of > standard v4l2 interfaces, so I've started the process of adding small > bits to v4l2 to at least retrieve the most basic information. Note that for properties that only depend on the sensor model, we could keep a database of sensors in userspace instead of pushing all that to the kernel. We could in theory do the same for the pixel array size and other related sizes, but as V4L2 already expose those, I would prefer using the V4L2 API to query them. > > >> But many sensors have a discrete list of supported formats and it is not > > >> clear how they map each format to the actual sensor. If it is clearly just > > >> done through binning or averaging, then all is fine. > > >> > > >> If the formats have different aspect ratios, then it becomes a bit more > > >> difficult how to relate the crop bounds with the format since you may not > > >> know to which sensor area the current format corresponds. > > > > > > I see. We've had the same discussion in the libcamera list, as indeed > > > the crop bounds might depend on the aspect ratio as you said. > > > > > > Maybe the crop_bound target is not the right place to report the > > > active area, or maybe the concept of active area is ill-defined if > > > the sensor is cross-shaped: > > > > > > /--------------------/ -> Pixel array size > > > ---------------------- > > > |x | |--- x|------> 4:3 crop bounds > > > |x------------------x| > > > |x| | | |x| > > > |x| | | |x|------> 21:9 crop bounds > > > |x| | | |x| > > > |x------------------x| > > > |x |________| x| > > > ---------------------- > > > | | > > > ---------------------------> Black pixels colums > > > > > >> > > >> I have no clear answer for you, to be honest, but it can get tricky. > > >> > > > > > > Indeed, but assuming a simpler square-shaped sensor, is the crop bound > > > target the right one to report the active pixel area ? > > > > I don't think so. > > > > The crop bounds defines the outer bounds of the area where you can put > > your crop rectangle. This can very well include areas where the pixels > > are covered and so always return black (sometimes used to obtain black > > levels, I believe). > > Ah, I didn't considered black pixels as part of a possible crop > selection.. The whole point of black pixels is that they can be captured (or at least part of them), for the ISP to calculate an average black level and subtract it from the active pixels. Let's thus make sure we clearly define every selection rectangle in the specification. > > The default crop rectangle would be the one that reports the active > > area. The native size rectangle would be the full pixel array. > > > > So CROP_DEFAULT <= CROP_BOUNDS <= NATIVE_SIZE. > > I see thanks for clarifying this > > > For a cross-shaped sensor I would expect that the CROP_BOUNDS/DEFAULT > > depends on the format (aspect ratio). > > Currently the selection API do not support to retrieve that > information depending on the aspect-ratio though > > > I think this makes sense. > > > > The specification definitely needs to be improved, patches are welcome... > > > > >>> > > >>> How would you report the two information I need? > > >> > > >> It depends on my original question: what do you need this information for? > > > > > > Please note that it's for the android camera stack in this case, but > > > it's an information that userspace might want to know for several > > > different reasons. Calibration and FOV calculation come to mind. Does > > > this makes sense to you? > > > > Ah, so that's what it is used for :-) > > > > Which of the three targets above would match with what Android needs? > > the full pixel array size would be reported by the NATIVE_SIZE target > the active area size by the CROP_DEFAULT rectangle (regardless of the > aspect ratio) > > It's still not clear to me what the CROP_BOUNDS rectangle would then > be used for, unless it is made to be dependent on an aspect ratio but > that would require userspace to provide a populated rectangle to > G_SELECTION and drivers to inspect it and return an opportune > CROP_BOUNDS one. Would this be acceptable ? > > > >>>> I don't think this patch and the next have anything to do with the location/rotate > > >>>> support. I would split it off from this series. > > >>>> > > >>> > > >>> Agreed, they were split in v1, maybe it has not been a wise move to > > >>> make a single series out of them. I'll split again. > > >>> > > >>>>> + break; > > >>>>> + default: > > >>>>> + return -EINVAL; > > >>>>> + } > > >>>>> + > > >>>>> + return 0; > > >>>>> +} > > >>>>> + > > >>>>> static int ov5670_get_skip_frames(struct v4l2_subdev *sd, u32 *frames) > > >>>>> { > > >>>>> *frames = OV5670_NUM_OF_SKIP_FRAMES; > > >>>>> @@ -2425,6 +2444,7 @@ static const struct v4l2_subdev_pad_ops ov5670_pad_ops = { > > >>>>> .enum_mbus_code = ov5670_enum_mbus_code, > > >>>>> .get_fmt = ov5670_get_pad_format, > > >>>>> .set_fmt = ov5670_set_pad_format, > > >>>>> + .get_selection = ov5670_get_selection, > > >>>>> .enum_frame_size = ov5670_enum_frame_size, > > >>>>> }; > > >>>>>
Hi Hans, Jacopo, On Thu, Aug 29, 2019 at 02:55:30PM +0200, Hans Verkuil wrote: > On 8/29/19 2:40 PM, Jacopo Mondi wrote: > > HI Hans, > > > > On Thu, Aug 29, 2019 at 12:20:18PM +0200, Hans Verkuil wrote: > >> On 8/27/19 11:23 AM, Jacopo Mondi wrote: > >>> Report the native pixel array size and the crop bounds for the ov5670 > >>> sensor driver. > >>> > >>> Signed-off-by: Jacopo Mondi <jacopo@jmondi.org> > >>> --- > >>> drivers/media/i2c/ov5670.c | 20 ++++++++++++++++++++ > >>> 1 file changed, 20 insertions(+) > >>> > >>> diff --git a/drivers/media/i2c/ov5670.c b/drivers/media/i2c/ov5670.c > >>> index 2bc57e85f721..3e22fe9ccad1 100644 > >>> --- a/drivers/media/i2c/ov5670.c > >>> +++ b/drivers/media/i2c/ov5670.c > >>> @@ -2258,6 +2258,25 @@ static int ov5670_set_pad_format(struct v4l2_subdev *sd, > >>> return 0; > >>> } > >>> > >>> +static int ov5670_get_selection(struct v4l2_subdev *sd, > >>> + struct v4l2_subdev_pad_config *cfg, > >>> + struct v4l2_subdev_selection *sel) > >>> +{ > >>> + switch (sel->target) { > >>> + case V4L2_SEL_TGT_CROP_BOUNDS: > >>> + case V4L2_SEL_TGT_NATIVE_SIZE: > >>> + sel->r.left = 0; > >>> + sel->r.top = 0; > >>> + sel->r.width = 2592; > >>> + sel->r.height = 1944; > >> > >> Why do you need this? > >> > > > > I need to expose the pixel array size and the active pixel area size, > > and I interpreted the two above targets as the right place where to do > > so. > > That didn't answer my question :-) > > Why do you need to expose this? Who uses it for what purpose? > > > > > At least for NATIVE_SIZE the documentation seems to match my > > understanding: > > > > "The native size of the device, e.g. a sensor’s pixel array. left and top > > fields are zero for this target." > > Correct. > > > > > > >> Since the format can change for this and the next driver I think CROP_BOUNDS > >> at least should match the current format. > >> > > > > Ah, does it? It was not clear to me from the documentation, as it > > suggested to me that target actually identifies the active pixel area > > > > "Bounds of the crop rectangle. All valid crop rectangles fit inside the > > crop bounds rectangle." > > > > It does not mention format, should this be updated? > > The problem is that for sensors it is indeed not clear. > > In principle the crop bounds matches the resolution that the sensor uses > pre-scaling. So yes, that means that it is equal to the native size. > > But many sensors have a discrete list of supported formats and it is not > clear how they map each format to the actual sensor. If it is clearly just > done through binning or averaging, then all is fine. Sensor drivers do; sensors themselves support much, much more than most drivers allow. But this is due to the nature of information available to the sensor driver developers, not really something that is trivial to change. > > If the formats have different aspect ratios, then it becomes a bit more > difficult how to relate the crop bounds with the format since you may not > know to which sensor area the current format corresponds. > > I have no clear answer for you, to be honest, but it can get tricky. I've suggested earlier that the crop and compose selection targets to be used to convey the cropping and binning (or scaling) that is done on the sensor, in that order. In reality, there are usually more than one (sometimes three) inside a sensor to crop, and often more than one place to scale as well. So the driver would need to accommodate this. The modes come with both cropping and scaling configuration, and V4L2 only allows specifying one at a time. In principle an output size may be achieved by scaling and cropping by different amounts, and as most drivers use only the output format (size) in mode selection, the result could be ambiguous. In practice this hasn't been an actual issue. Better sensor drivers (such as smiapp) do not have this problem as the configurations (cropping in three different places as well as binning and scaling) can be all independently configured. So with some luck this problem could disappear in time with newer hardware and better hardware documentation. I have no objections to providing the cropping and scaling information to the user space using the selection rectangles, albeit it's somewhat against the semantics currently. This approach would also require using compose rectangles on the source pads which is not supported (documentation-wise) at the moment, but it's better that way: it can be added now. There are other, older, drivers such as omap3isp that configure scaling based on the source format configuration only. Perhaps a new selection flag telling the selection is read-only?
Hi Sakari, Hans, On Tue, Sep 03, 2019 at 04:06:26PM +0300, Sakari Ailus wrote: > Hi Hans, Jacopo, > > On Thu, Aug 29, 2019 at 02:55:30PM +0200, Hans Verkuil wrote: > > On 8/29/19 2:40 PM, Jacopo Mondi wrote: > > > HI Hans, > > > > > > On Thu, Aug 29, 2019 at 12:20:18PM +0200, Hans Verkuil wrote: > > >> On 8/27/19 11:23 AM, Jacopo Mondi wrote: > > >>> Report the native pixel array size and the crop bounds for the ov5670 > > >>> sensor driver. > > >>> > > >>> Signed-off-by: Jacopo Mondi <jacopo@jmondi.org> > > >>> --- > > >>> drivers/media/i2c/ov5670.c | 20 ++++++++++++++++++++ > > >>> 1 file changed, 20 insertions(+) > > >>> > > >>> diff --git a/drivers/media/i2c/ov5670.c b/drivers/media/i2c/ov5670.c > > >>> index 2bc57e85f721..3e22fe9ccad1 100644 > > >>> --- a/drivers/media/i2c/ov5670.c > > >>> +++ b/drivers/media/i2c/ov5670.c > > >>> @@ -2258,6 +2258,25 @@ static int ov5670_set_pad_format(struct v4l2_subdev *sd, > > >>> return 0; > > >>> } > > >>> > > >>> +static int ov5670_get_selection(struct v4l2_subdev *sd, > > >>> + struct v4l2_subdev_pad_config *cfg, > > >>> + struct v4l2_subdev_selection *sel) > > >>> +{ > > >>> + switch (sel->target) { > > >>> + case V4L2_SEL_TGT_CROP_BOUNDS: > > >>> + case V4L2_SEL_TGT_NATIVE_SIZE: > > >>> + sel->r.left = 0; > > >>> + sel->r.top = 0; > > >>> + sel->r.width = 2592; > > >>> + sel->r.height = 1944; > > >> > > >> Why do you need this? > > >> > > > > > > I need to expose the pixel array size and the active pixel area size, > > > and I interpreted the two above targets as the right place where to do > > > so. > > > > That didn't answer my question :-) > > > > Why do you need to expose this? Who uses it for what purpose? > > > > > > > > At least for NATIVE_SIZE the documentation seems to match my > > > understanding: > > > > > > "The native size of the device, e.g. a sensor’s pixel array. left and top > > > fields are zero for this target." > > > > Correct. > > > > > > > > > > >> Since the format can change for this and the next driver I think CROP_BOUNDS > > >> at least should match the current format. > > >> > > > > > > Ah, does it? It was not clear to me from the documentation, as it > > > suggested to me that target actually identifies the active pixel area > > > > > > "Bounds of the crop rectangle. All valid crop rectangles fit inside the > > > crop bounds rectangle." > > > > > > It does not mention format, should this be updated? > > > > The problem is that for sensors it is indeed not clear. > > > > In principle the crop bounds matches the resolution that the sensor uses > > pre-scaling. So yes, that means that it is equal to the native size. > > > > But many sensors have a discrete list of supported formats and it is not > > clear how they map each format to the actual sensor. If it is clearly just > > done through binning or averaging, then all is fine. > > Sensor drivers do; sensors themselves support much, much more than most > drivers allow. But this is due to the nature of information available to > the sensor driver developers, not really something that is trivial to > change. > > > > > If the formats have different aspect ratios, then it becomes a bit more > > difficult how to relate the crop bounds with the format since you may not > > know to which sensor area the current format corresponds. > > > > I have no clear answer for you, to be honest, but it can get tricky. > > I've suggested earlier that the crop and compose selection targets to be > used to convey the cropping and binning (or scaling) that is done on the > sensor, in that order. In reality, there are usually more than one > (sometimes three) inside a sensor to crop, and often more than one place to > scale as well. So the driver would need to accommodate this. > > The modes come with both cropping and scaling configuration, and V4L2 only > allows specifying one at a time. In principle an output size may be > achieved by scaling and cropping by different amounts, and as most drivers > use only the output format (size) in mode selection, the result could be > ambiguous. In practice this hasn't been an actual issue. > > Better sensor drivers (such as smiapp) do not have this problem as the > configurations (cropping in three different places as well as binning and > scaling) can be all independently configured. So with some luck this > problem could disappear in time with newer hardware and better hardware > documentation. > > I have no objections to providing the cropping and scaling information to > the user space using the selection rectangles, albeit it's somewhat against > the semantics currently. This approach would also require using compose > rectangles on the source pads which is not supported (documentation-wise) > at the moment, but it's better that way: it can be added now. There are > other, older, drivers such as omap3isp that configure scaling based on the > source format configuration only. Thanks for all information here, but I think we've gone a bit far from my original point. The cropping and scaling informations you mention are, in my understanding, the portion of the pixel array which is fed to the ISP before scaling, and the result of the ISP binning/averaging respectively. Those information indeed depends on the desired capture resolution, the driver implementation, and the sensor capabilities. What I was interested in was just reporting to userspace the physical size of the active pixel array area, which should reasonably be defined as a sub-portion of the native pixel array, excluding the back calibration pixels. In one of the previous emails Hans suggested to use CROP_DEFAULT for that purpose, but in the documentation it is reported not to apply to subdevice :( Do we need a new target, similar to V4L2_SEL_TGT_NATIVE_SIZE that could maybe report multiple rectangles to accommodate cross-shaped sensors? Are the selection API the right place to report these information? With a control, it might be easier to report such a static information... Thanks j > > Perhaps a new selection flag telling the selection is read-only? > > -- > Regards, > > Sakari Ailus > sakari.ailus@linux.intel.com
On 9/3/19 6:49 PM, Jacopo Mondi wrote: > Hi Sakari, Hans, > > On Tue, Sep 03, 2019 at 04:06:26PM +0300, Sakari Ailus wrote: >> Hi Hans, Jacopo, >> >> On Thu, Aug 29, 2019 at 02:55:30PM +0200, Hans Verkuil wrote: >>> On 8/29/19 2:40 PM, Jacopo Mondi wrote: >>>> HI Hans, >>>> >>>> On Thu, Aug 29, 2019 at 12:20:18PM +0200, Hans Verkuil wrote: >>>>> On 8/27/19 11:23 AM, Jacopo Mondi wrote: >>>>>> Report the native pixel array size and the crop bounds for the ov5670 >>>>>> sensor driver. >>>>>> >>>>>> Signed-off-by: Jacopo Mondi <jacopo@jmondi.org> >>>>>> --- >>>>>> drivers/media/i2c/ov5670.c | 20 ++++++++++++++++++++ >>>>>> 1 file changed, 20 insertions(+) >>>>>> >>>>>> diff --git a/drivers/media/i2c/ov5670.c b/drivers/media/i2c/ov5670.c >>>>>> index 2bc57e85f721..3e22fe9ccad1 100644 >>>>>> --- a/drivers/media/i2c/ov5670.c >>>>>> +++ b/drivers/media/i2c/ov5670.c >>>>>> @@ -2258,6 +2258,25 @@ static int ov5670_set_pad_format(struct v4l2_subdev *sd, >>>>>> return 0; >>>>>> } >>>>>> >>>>>> +static int ov5670_get_selection(struct v4l2_subdev *sd, >>>>>> + struct v4l2_subdev_pad_config *cfg, >>>>>> + struct v4l2_subdev_selection *sel) >>>>>> +{ >>>>>> + switch (sel->target) { >>>>>> + case V4L2_SEL_TGT_CROP_BOUNDS: >>>>>> + case V4L2_SEL_TGT_NATIVE_SIZE: >>>>>> + sel->r.left = 0; >>>>>> + sel->r.top = 0; >>>>>> + sel->r.width = 2592; >>>>>> + sel->r.height = 1944; >>>>> >>>>> Why do you need this? >>>>> >>>> >>>> I need to expose the pixel array size and the active pixel area size, >>>> and I interpreted the two above targets as the right place where to do >>>> so. >>> >>> That didn't answer my question :-) >>> >>> Why do you need to expose this? Who uses it for what purpose? >>> >>>> >>>> At least for NATIVE_SIZE the documentation seems to match my >>>> understanding: >>>> >>>> "The native size of the device, e.g. a sensor’s pixel array. left and top >>>> fields are zero for this target." >>> >>> Correct. >>> >>>> >>>> >>>>> Since the format can change for this and the next driver I think CROP_BOUNDS >>>>> at least should match the current format. >>>>> >>>> >>>> Ah, does it? It was not clear to me from the documentation, as it >>>> suggested to me that target actually identifies the active pixel area >>>> >>>> "Bounds of the crop rectangle. All valid crop rectangles fit inside the >>>> crop bounds rectangle." >>>> >>>> It does not mention format, should this be updated? >>> >>> The problem is that for sensors it is indeed not clear. >>> >>> In principle the crop bounds matches the resolution that the sensor uses >>> pre-scaling. So yes, that means that it is equal to the native size. >>> >>> But many sensors have a discrete list of supported formats and it is not >>> clear how they map each format to the actual sensor. If it is clearly just >>> done through binning or averaging, then all is fine. >> >> Sensor drivers do; sensors themselves support much, much more than most >> drivers allow. But this is due to the nature of information available to >> the sensor driver developers, not really something that is trivial to >> change. >> >>> >>> If the formats have different aspect ratios, then it becomes a bit more >>> difficult how to relate the crop bounds with the format since you may not >>> know to which sensor area the current format corresponds. >>> >>> I have no clear answer for you, to be honest, but it can get tricky. >> >> I've suggested earlier that the crop and compose selection targets to be >> used to convey the cropping and binning (or scaling) that is done on the >> sensor, in that order. In reality, there are usually more than one >> (sometimes three) inside a sensor to crop, and often more than one place to >> scale as well. So the driver would need to accommodate this. >> >> The modes come with both cropping and scaling configuration, and V4L2 only >> allows specifying one at a time. In principle an output size may be >> achieved by scaling and cropping by different amounts, and as most drivers >> use only the output format (size) in mode selection, the result could be >> ambiguous. In practice this hasn't been an actual issue. >> >> Better sensor drivers (such as smiapp) do not have this problem as the >> configurations (cropping in three different places as well as binning and >> scaling) can be all independently configured. So with some luck this >> problem could disappear in time with newer hardware and better hardware >> documentation. >> >> I have no objections to providing the cropping and scaling information to >> the user space using the selection rectangles, albeit it's somewhat against >> the semantics currently. This approach would also require using compose >> rectangles on the source pads which is not supported (documentation-wise) >> at the moment, but it's better that way: it can be added now. There are >> other, older, drivers such as omap3isp that configure scaling based on the >> source format configuration only. > > Thanks for all information here, but I think we've gone a bit far > from my original point. The cropping and scaling informations you > mention are, in my understanding, the portion of the pixel array which > is fed to the ISP before scaling, and the result of the ISP > binning/averaging respectively. Those information indeed depends on the > desired capture resolution, the driver implementation, and the sensor > capabilities. > > What I was interested in was just reporting to userspace the physical > size of the active pixel array area, which should reasonably be > defined as a sub-portion of the native pixel array, excluding the back > calibration pixels. > > In one of the previous emails Hans suggested to use CROP_DEFAULT for > that purpose, but in the documentation it is reported not to apply to > subdevice :( > > Do we need a new target, similar to V4L2_SEL_TGT_NATIVE_SIZE that > could maybe report multiple rectangles to accommodate cross-shaped > sensors? > > Are the selection API the right place to report these information? > With a control, it might be easier to report such a static > information... Sorry for the late follow-up. I am uncomfortable with hacking something here. I strongly feel that we are missing a proper API to configure a sensor. If I look at video receivers (SDTV or HDTV), then in both cases you set up the incoming video resolution with S_STD or S_DV_TIMINGS. This explicitly defines the incoming image size and all crop/compose/scale operations operate on that image size. In my opinion we are missing an equivalent to that for sensors. I think the original thinking was that sensors have a fixed pixel array, so there is nothing to configure. But with cross-shaped sensors, binning/skipping, sensors that just hide all the inner details and just report a set of discrete framesizes, this is not actually trivial anymore. And the odd thing is that our API does have discovery (at least to some extent) of the various options through ENUM_FRAMESIZES, but does not let you choose a specific variant. We abuse S_FMT for this, which is really not the right ioctl to use since it defines what you want to receive in memory after cropping/composing/scaling, not the initial image size before all these operations take place. Regarding binning and skipping: I never understood why this wasn't configured using controls. Is there a reason why it was never implemented like that? If we had a Set Source Size operation (either a control, new ioctl or selection target) then you can use that to select a specific source size based on what ENUM_FRAMESIZES reports. Sensor drivers can choose to either enumerate variants based on the cross-shape and binning/skipping, or just enumerate variants based on the cross-shape and add binning/skipping controls to explicitly set this up. In any case, since you now know exactly which image size the sensor produces, you can set up all the selection rectangles associated with that correctly. Since I am not a sensor expert I no doubt missed things, but in my view we really need to see a sensor not as a fixed array, but as something more akin to video receiver, i.e. you need to explicitly configure the video source before cropping/composing/scaling takes place. So in the example of a cross-shaped sensor ENUM_FRAMESIZES would report two sizes (landscape/portrait), binning/cropping controls (optional, this might be implicit in the enumerated framesizes), and after configuring all this using the new API 'Set Source Size' (or whatever it will be called) and possible binning/skipping controls, the user can read various selection targets to get all the needed information. Any cropping/composing/scaling will be done based on the selected image size + binning/skipping. Re binning/skipping: one other option is to add a flags field to v4l2_frmsizeenum that reports if the reported size was achieved via binning and/or skipping. You wouldn't need controls in that case. Regards, Hans
Hi Hans, thanks for the lenghty reply On Thu, Sep 26, 2019 at 10:11:51AM +0200, Hans Verkuil wrote: > On 9/3/19 6:49 PM, Jacopo Mondi wrote: > > Hi Sakari, Hans, > > > > On Tue, Sep 03, 2019 at 04:06:26PM +0300, Sakari Ailus wrote: > >> Hi Hans, Jacopo, > >> > >> On Thu, Aug 29, 2019 at 02:55:30PM +0200, Hans Verkuil wrote: > >>> On 8/29/19 2:40 PM, Jacopo Mondi wrote: > >>>> HI Hans, > >>>> > >>>> On Thu, Aug 29, 2019 at 12:20:18PM +0200, Hans Verkuil wrote: > >>>>> On 8/27/19 11:23 AM, Jacopo Mondi wrote: > >>>>>> Report the native pixel array size and the crop bounds for the ov5670 > >>>>>> sensor driver. > >>>>>> > >>>>>> Signed-off-by: Jacopo Mondi <jacopo@jmondi.org> > >>>>>> --- > >>>>>> drivers/media/i2c/ov5670.c | 20 ++++++++++++++++++++ > >>>>>> 1 file changed, 20 insertions(+) > >>>>>> > >>>>>> diff --git a/drivers/media/i2c/ov5670.c b/drivers/media/i2c/ov5670.c > >>>>>> index 2bc57e85f721..3e22fe9ccad1 100644 > >>>>>> --- a/drivers/media/i2c/ov5670.c > >>>>>> +++ b/drivers/media/i2c/ov5670.c > >>>>>> @@ -2258,6 +2258,25 @@ static int ov5670_set_pad_format(struct v4l2_subdev *sd, > >>>>>> return 0; > >>>>>> } > >>>>>> > >>>>>> +static int ov5670_get_selection(struct v4l2_subdev *sd, > >>>>>> + struct v4l2_subdev_pad_config *cfg, > >>>>>> + struct v4l2_subdev_selection *sel) > >>>>>> +{ > >>>>>> + switch (sel->target) { > >>>>>> + case V4L2_SEL_TGT_CROP_BOUNDS: > >>>>>> + case V4L2_SEL_TGT_NATIVE_SIZE: > >>>>>> + sel->r.left = 0; > >>>>>> + sel->r.top = 0; > >>>>>> + sel->r.width = 2592; > >>>>>> + sel->r.height = 1944; > >>>>> > >>>>> Why do you need this? > >>>>> > >>>> > >>>> I need to expose the pixel array size and the active pixel area size, > >>>> and I interpreted the two above targets as the right place where to do > >>>> so. > >>> > >>> That didn't answer my question :-) > >>> > >>> Why do you need to expose this? Who uses it for what purpose? > >>> > >>>> > >>>> At least for NATIVE_SIZE the documentation seems to match my > >>>> understanding: > >>>> > >>>> "The native size of the device, e.g. a sensor’s pixel array. left and top > >>>> fields are zero for this target." > >>> > >>> Correct. > >>> > >>>> > >>>> > >>>>> Since the format can change for this and the next driver I think CROP_BOUNDS > >>>>> at least should match the current format. > >>>>> > >>>> > >>>> Ah, does it? It was not clear to me from the documentation, as it > >>>> suggested to me that target actually identifies the active pixel area > >>>> > >>>> "Bounds of the crop rectangle. All valid crop rectangles fit inside the > >>>> crop bounds rectangle." > >>>> > >>>> It does not mention format, should this be updated? > >>> > >>> The problem is that for sensors it is indeed not clear. > >>> > >>> In principle the crop bounds matches the resolution that the sensor uses > >>> pre-scaling. So yes, that means that it is equal to the native size. > >>> > >>> But many sensors have a discrete list of supported formats and it is not > >>> clear how they map each format to the actual sensor. If it is clearly just > >>> done through binning or averaging, then all is fine. > >> > >> Sensor drivers do; sensors themselves support much, much more than most > >> drivers allow. But this is due to the nature of information available to > >> the sensor driver developers, not really something that is trivial to > >> change. > >> > >>> > >>> If the formats have different aspect ratios, then it becomes a bit more > >>> difficult how to relate the crop bounds with the format since you may not > >>> know to which sensor area the current format corresponds. > >>> > >>> I have no clear answer for you, to be honest, but it can get tricky. > >> > >> I've suggested earlier that the crop and compose selection targets to be > >> used to convey the cropping and binning (or scaling) that is done on the > >> sensor, in that order. In reality, there are usually more than one > >> (sometimes three) inside a sensor to crop, and often more than one place to > >> scale as well. So the driver would need to accommodate this. > >> > >> The modes come with both cropping and scaling configuration, and V4L2 only > >> allows specifying one at a time. In principle an output size may be > >> achieved by scaling and cropping by different amounts, and as most drivers > >> use only the output format (size) in mode selection, the result could be > >> ambiguous. In practice this hasn't been an actual issue. > >> > >> Better sensor drivers (such as smiapp) do not have this problem as the > >> configurations (cropping in three different places as well as binning and > >> scaling) can be all independently configured. So with some luck this > >> problem could disappear in time with newer hardware and better hardware > >> documentation. > >> > >> I have no objections to providing the cropping and scaling information to > >> the user space using the selection rectangles, albeit it's somewhat against > >> the semantics currently. This approach would also require using compose > >> rectangles on the source pads which is not supported (documentation-wise) > >> at the moment, but it's better that way: it can be added now. There are > >> other, older, drivers such as omap3isp that configure scaling based on the > >> source format configuration only. > > > > Thanks for all information here, but I think we've gone a bit far > > from my original point. The cropping and scaling informations you > > mention are, in my understanding, the portion of the pixel array which > > is fed to the ISP before scaling, and the result of the ISP > > binning/averaging respectively. Those information indeed depends on the > > desired capture resolution, the driver implementation, and the sensor > > capabilities. > > > > What I was interested in was just reporting to userspace the physical > > size of the active pixel array area, which should reasonably be > > defined as a sub-portion of the native pixel array, excluding the back > > calibration pixels. > > > > In one of the previous emails Hans suggested to use CROP_DEFAULT for > > that purpose, but in the documentation it is reported not to apply to > > subdevice :( > > > > Do we need a new target, similar to V4L2_SEL_TGT_NATIVE_SIZE that > > could maybe report multiple rectangles to accommodate cross-shaped > > sensors? > > > > Are the selection API the right place to report these information? > > With a control, it might be easier to report such a static > > information... > > Sorry for the late follow-up. > > I am uncomfortable with hacking something here. I strongly feel that we > are missing a proper API to configure a sensor. > > If I look at video receivers (SDTV or HDTV), then in both cases you set up > the incoming video resolution with S_STD or S_DV_TIMINGS. This explicitly > defines the incoming image size and all crop/compose/scale operations > operate on that image size. > > In my opinion we are missing an equivalent to that for sensors. I think the > original thinking was that sensors have a fixed pixel array, so there is > nothing to configure. But with cross-shaped sensors, binning/skipping, > sensors that just hide all the inner details and just report a set of > discrete framesizes, this is not actually trivial anymore. > > And the odd thing is that our API does have discovery (at least to some > extent) of the various options through ENUM_FRAMESIZES, but does not let > you choose a specific variant. We abuse S_FMT for this, which is really > not the right ioctl to use since it defines what you want to receive in > memory after cropping/composing/scaling, not the initial image size before > all these operations take place. > > Regarding binning and skipping: I never understood why this wasn't > configured using controls. Is there a reason why it was never implemented > like that? > > If we had a Set Source Size operation (either a control, new ioctl or selection > target) then you can use that to select a specific source size based on what > ENUM_FRAMESIZES reports. Sensor drivers can choose to either enumerate > variants based on the cross-shape and binning/skipping, or just enumerate variants > based on the cross-shape and add binning/skipping controls to explicitly set > this up. > > In any case, since you now know exactly which image size the sensor produces, > you can set up all the selection rectangles associated with that correctly. > > Since I am not a sensor expert I no doubt missed things, but in my view we > really need to see a sensor not as a fixed array, but as something more akin > to video receiver, i.e. you need to explicitly configure the video source > before cropping/composing/scaling takes place. > > So in the example of a cross-shaped sensor ENUM_FRAMESIZES would report > two sizes (landscape/portrait), binning/cropping controls (optional, this > might be implicit in the enumerated framesizes), and after configuring all > this using the new API 'Set Source Size' (or whatever it will be called) and > possible binning/skipping controls, the user can read various selection > targets to get all the needed information. > > Any cropping/composing/scaling will be done based on the selected image size > + binning/skipping. If I'm not mistaken smiapp sensor driver models this by exposing multiple subdevices, one for the pixel array and additional ones for the scaler (and the CSI-2 TX port, iirc from my multiplexed stream support adventures). Sakari knows better for sure here... Anyway, going towards a model where sensor expose several subdeves would allow to have one for each configurable part of the sensor processing pipeline, such as one for the raw pixel array (where one could support cross-shaped sensors as you have suggested), one for ISP input where to select the area of the pixel array to feed to the ISP, and one to model the final processing stage where to to get to the final image from, obtained through binning/skipping/averaging whatever and possibly by applying a composition rectangle or selecting the desired output format, if the sensor supports so. Am I over-simplifying things here? I fear this is mostly frowned upon by the constant lack of decent documentation from sensor manufacturers, as most drivers still relies on magic blobs 'guaranteed to work' and often produced by some 'certified' tuning applications.. And yes, more burden on userspace, as right now most sensors are a single entity that gives you images in one format/resolution and that's it, but this clearly is showing limits. On this side, libcamera could be a great help and sensor-specific configurations will be offloaded there. One quick consideration from my limited experience with sensors: the method used to obtain the final image from the portion of the pixel array fed to the ISP is not something often freely selectable. From the sensor I've seen the manuals of, to obtain a certain resolution you use averaging/binning, for another one skipping or direct crop of the full pixel array etc.. So I'm not sure a control would fit well here. Again, I might be really simplifying things :) > > Re binning/skipping: one other option is to add a flags field to v4l2_frmsizeenum > that reports if the reported size was achieved via binning and/or skipping. > Yes, but in this case that would be a read-only information, not sure what you would use it for Thanks j > You wouldn't need controls in that case. > > Regards, > > Hans
On Fri, Sep 27, 2019 at 3:55 AM Jacopo Mondi <jacopo@jmondi.org> wrote: > > Hi Hans, thanks for the lenghty reply > > On Thu, Sep 26, 2019 at 10:11:51AM +0200, Hans Verkuil wrote: > > On 9/3/19 6:49 PM, Jacopo Mondi wrote: > > > Hi Sakari, Hans, > > > > > > On Tue, Sep 03, 2019 at 04:06:26PM +0300, Sakari Ailus wrote: > > >> Hi Hans, Jacopo, > > >> > > >> On Thu, Aug 29, 2019 at 02:55:30PM +0200, Hans Verkuil wrote: > > >>> On 8/29/19 2:40 PM, Jacopo Mondi wrote: > > >>>> HI Hans, > > >>>> > > >>>> On Thu, Aug 29, 2019 at 12:20:18PM +0200, Hans Verkuil wrote: > > >>>>> On 8/27/19 11:23 AM, Jacopo Mondi wrote: > > >>>>>> Report the native pixel array size and the crop bounds for the ov5670 > > >>>>>> sensor driver. > > >>>>>> > > >>>>>> Signed-off-by: Jacopo Mondi <jacopo@jmondi.org> > > >>>>>> --- > > >>>>>> drivers/media/i2c/ov5670.c | 20 ++++++++++++++++++++ > > >>>>>> 1 file changed, 20 insertions(+) > > >>>>>> > > >>>>>> diff --git a/drivers/media/i2c/ov5670.c b/drivers/media/i2c/ov5670.c > > >>>>>> index 2bc57e85f721..3e22fe9ccad1 100644 > > >>>>>> --- a/drivers/media/i2c/ov5670.c > > >>>>>> +++ b/drivers/media/i2c/ov5670.c > > >>>>>> @@ -2258,6 +2258,25 @@ static int ov5670_set_pad_format(struct v4l2_subdev *sd, > > >>>>>> return 0; > > >>>>>> } > > >>>>>> > > >>>>>> +static int ov5670_get_selection(struct v4l2_subdev *sd, > > >>>>>> + struct v4l2_subdev_pad_config *cfg, > > >>>>>> + struct v4l2_subdev_selection *sel) > > >>>>>> +{ > > >>>>>> + switch (sel->target) { > > >>>>>> + case V4L2_SEL_TGT_CROP_BOUNDS: > > >>>>>> + case V4L2_SEL_TGT_NATIVE_SIZE: > > >>>>>> + sel->r.left = 0; > > >>>>>> + sel->r.top = 0; > > >>>>>> + sel->r.width = 2592; > > >>>>>> + sel->r.height = 1944; > > >>>>> > > >>>>> Why do you need this? > > >>>>> > > >>>> > > >>>> I need to expose the pixel array size and the active pixel area size, > > >>>> and I interpreted the two above targets as the right place where to do > > >>>> so. > > >>> > > >>> That didn't answer my question :-) > > >>> > > >>> Why do you need to expose this? Who uses it for what purpose? > > >>> > > >>>> > > >>>> At least for NATIVE_SIZE the documentation seems to match my > > >>>> understanding: > > >>>> > > >>>> "The native size of the device, e.g. a sensor’s pixel array. left and top > > >>>> fields are zero for this target." > > >>> > > >>> Correct. > > >>> > > >>>> > > >>>> > > >>>>> Since the format can change for this and the next driver I think CROP_BOUNDS > > >>>>> at least should match the current format. > > >>>>> > > >>>> > > >>>> Ah, does it? It was not clear to me from the documentation, as it > > >>>> suggested to me that target actually identifies the active pixel area > > >>>> > > >>>> "Bounds of the crop rectangle. All valid crop rectangles fit inside the > > >>>> crop bounds rectangle." > > >>>> > > >>>> It does not mention format, should this be updated? > > >>> > > >>> The problem is that for sensors it is indeed not clear. > > >>> > > >>> In principle the crop bounds matches the resolution that the sensor uses > > >>> pre-scaling. So yes, that means that it is equal to the native size. > > >>> > > >>> But many sensors have a discrete list of supported formats and it is not > > >>> clear how they map each format to the actual sensor. If it is clearly just > > >>> done through binning or averaging, then all is fine. > > >> > > >> Sensor drivers do; sensors themselves support much, much more than most > > >> drivers allow. But this is due to the nature of information available to > > >> the sensor driver developers, not really something that is trivial to > > >> change. > > >> > > >>> > > >>> If the formats have different aspect ratios, then it becomes a bit more > > >>> difficult how to relate the crop bounds with the format since you may not > > >>> know to which sensor area the current format corresponds. > > >>> > > >>> I have no clear answer for you, to be honest, but it can get tricky. > > >> > > >> I've suggested earlier that the crop and compose selection targets to be > > >> used to convey the cropping and binning (or scaling) that is done on the > > >> sensor, in that order. In reality, there are usually more than one > > >> (sometimes three) inside a sensor to crop, and often more than one place to > > >> scale as well. So the driver would need to accommodate this. > > >> > > >> The modes come with both cropping and scaling configuration, and V4L2 only > > >> allows specifying one at a time. In principle an output size may be > > >> achieved by scaling and cropping by different amounts, and as most drivers > > >> use only the output format (size) in mode selection, the result could be > > >> ambiguous. In practice this hasn't been an actual issue. > > >> > > >> Better sensor drivers (such as smiapp) do not have this problem as the > > >> configurations (cropping in three different places as well as binning and > > >> scaling) can be all independently configured. So with some luck this > > >> problem could disappear in time with newer hardware and better hardware > > >> documentation. > > >> > > >> I have no objections to providing the cropping and scaling information to > > >> the user space using the selection rectangles, albeit it's somewhat against > > >> the semantics currently. This approach would also require using compose > > >> rectangles on the source pads which is not supported (documentation-wise) > > >> at the moment, but it's better that way: it can be added now. There are > > >> other, older, drivers such as omap3isp that configure scaling based on the > > >> source format configuration only. > > > > > > Thanks for all information here, but I think we've gone a bit far > > > from my original point. The cropping and scaling informations you > > > mention are, in my understanding, the portion of the pixel array which > > > is fed to the ISP before scaling, and the result of the ISP > > > binning/averaging respectively. Those information indeed depends on the > > > desired capture resolution, the driver implementation, and the sensor > > > capabilities. > > > > > > What I was interested in was just reporting to userspace the physical > > > size of the active pixel array area, which should reasonably be > > > defined as a sub-portion of the native pixel array, excluding the back > > > calibration pixels. > > > > > > In one of the previous emails Hans suggested to use CROP_DEFAULT for > > > that purpose, but in the documentation it is reported not to apply to > > > subdevice :( > > > > > > Do we need a new target, similar to V4L2_SEL_TGT_NATIVE_SIZE that > > > could maybe report multiple rectangles to accommodate cross-shaped > > > sensors? > > > > > > Are the selection API the right place to report these information? > > > With a control, it might be easier to report such a static > > > information... > > > > Sorry for the late follow-up. > > > > I am uncomfortable with hacking something here. I strongly feel that we > > are missing a proper API to configure a sensor. > > > > If I look at video receivers (SDTV or HDTV), then in both cases you set up > > the incoming video resolution with S_STD or S_DV_TIMINGS. This explicitly > > defines the incoming image size and all crop/compose/scale operations > > operate on that image size. > > > > In my opinion we are missing an equivalent to that for sensors. I think the > > original thinking was that sensors have a fixed pixel array, so there is > > nothing to configure. But with cross-shaped sensors, binning/skipping, > > sensors that just hide all the inner details and just report a set of > > discrete framesizes, this is not actually trivial anymore. > > > > And the odd thing is that our API does have discovery (at least to some > > extent) of the various options through ENUM_FRAMESIZES, but does not let > > you choose a specific variant. We abuse S_FMT for this, which is really > > not the right ioctl to use since it defines what you want to receive in > > memory after cropping/composing/scaling, not the initial image size before > > all these operations take place. > > > > Regarding binning and skipping: I never understood why this wasn't > > configured using controls. Is there a reason why it was never implemented > > like that? > > > > If we had a Set Source Size operation (either a control, new ioctl or selection > > target) then you can use that to select a specific source size based on what > > ENUM_FRAMESIZES reports. Sensor drivers can choose to either enumerate > > variants based on the cross-shape and binning/skipping, or just enumerate variants > > based on the cross-shape and add binning/skipping controls to explicitly set > > this up. > > > > In any case, since you now know exactly which image size the sensor produces, > > you can set up all the selection rectangles associated with that correctly. > > > > Since I am not a sensor expert I no doubt missed things, but in my view we > > really need to see a sensor not as a fixed array, but as something more akin > > to video receiver, i.e. you need to explicitly configure the video source > > before cropping/composing/scaling takes place. > > > > So in the example of a cross-shaped sensor ENUM_FRAMESIZES would report > > two sizes (landscape/portrait), binning/cropping controls (optional, this > > might be implicit in the enumerated framesizes), and after configuring all > > this using the new API 'Set Source Size' (or whatever it will be called) and > > possible binning/skipping controls, the user can read various selection > > targets to get all the needed information. > > > > Any cropping/composing/scaling will be done based on the selected image size > > + binning/skipping. > > If I'm not mistaken smiapp sensor driver models this by exposing > multiple subdevices, one for the pixel array and additional ones for > the scaler (and the CSI-2 TX port, iirc from my multiplexed stream > support adventures). Sakari knows better for sure here... > > Anyway, going towards a model where sensor expose several subdeves > would allow to have one for each configurable part of the sensor > processing pipeline, such as one for the raw pixel array (where one could > support cross-shaped sensors as you have suggested), one for ISP input > where to select the area of the pixel array to feed to the ISP, and one > to model the final processing stage where to to get to the final > image from, obtained through binning/skipping/averaging whatever and > possibly by applying a composition rectangle or selecting the > desired output format, if the sensor supports so. > > Am I over-simplifying things here? I fear this is mostly frowned upon by > the constant lack of decent documentation from sensor manufacturers, > as most drivers still relies on magic blobs 'guaranteed to work' and > often produced by some 'certified' tuning applications.. > I can only add that even though we usually have the right documentation, there are still specific blessed (certified?) register combinations that sensor and SoC vendor support and if we use anything else, we're on our own. > And yes, more burden on userspace, as right now most sensors are a single > entity that gives you images in one format/resolution and that's it, > but this clearly is showing limits. On this side, libcamera could be a > great help and sensor-specific configurations will be offloaded there. > > One quick consideration from my limited experience with sensors: the > method used to obtain the final image from the portion of the pixel > array fed to the ISP is not something often freely selectable. From > the sensor I've seen the manuals of, to obtain a certain resolution > you use averaging/binning, for another one skipping or direct crop of > the full pixel array etc.. So I'm not sure a control would fit well here. > > Again, I might be really simplifying things :) > Modelling various processing bits with a proper topology surely sounds reasonable. However I think we may need to lay some educational groundwork first to switch the mindsets from register array-based pseudo-drivers to proper drivers. Best regards, Tomasz > > > > Re binning/skipping: one other option is to add a flags field to v4l2_frmsizeenum > > that reports if the reported size was achieved via binning and/or skipping. > > > > Yes, but in this case that would be a read-only information, not sure > what you would use it for > > Thanks > j > > > You wouldn't need controls in that case. > > > > Regards, > > > > Hans
diff --git a/drivers/media/i2c/ov5670.c b/drivers/media/i2c/ov5670.c index 2bc57e85f721..3e22fe9ccad1 100644 --- a/drivers/media/i2c/ov5670.c +++ b/drivers/media/i2c/ov5670.c @@ -2258,6 +2258,25 @@ static int ov5670_set_pad_format(struct v4l2_subdev *sd, return 0; } +static int ov5670_get_selection(struct v4l2_subdev *sd, + struct v4l2_subdev_pad_config *cfg, + struct v4l2_subdev_selection *sel) +{ + switch (sel->target) { + case V4L2_SEL_TGT_CROP_BOUNDS: + case V4L2_SEL_TGT_NATIVE_SIZE: + sel->r.left = 0; + sel->r.top = 0; + sel->r.width = 2592; + sel->r.height = 1944; + break; + default: + return -EINVAL; + } + + return 0; +} + static int ov5670_get_skip_frames(struct v4l2_subdev *sd, u32 *frames) { *frames = OV5670_NUM_OF_SKIP_FRAMES; @@ -2425,6 +2444,7 @@ static const struct v4l2_subdev_pad_ops ov5670_pad_ops = { .enum_mbus_code = ov5670_enum_mbus_code, .get_fmt = ov5670_get_pad_format, .set_fmt = ov5670_set_pad_format, + .get_selection = ov5670_get_selection, .enum_frame_size = ov5670_enum_frame_size, };
Report the native pixel array size and the crop bounds for the ov5670 sensor driver. Signed-off-by: Jacopo Mondi <jacopo@jmondi.org> --- drivers/media/i2c/ov5670.c | 20 ++++++++++++++++++++ 1 file changed, 20 insertions(+)