[v1] staging: media: starfive: Renamed capture_raw to capture_dump
Commit Message
The pixel formats captured by the capture_raw video device depends on
what pixel formats come from the source device. It is actually dump
the source device data. So renamed it to capture_dump.
Signed-off-by: Changhuang Liang <changhuang.liang@starfivetech.com>
---
.../admin-guide/media/starfive_camss.rst | 2 +-
.../media/starfive_camss_graph.dot | 2 +-
.../staging/media/starfive/camss/stf-camss.c | 6 ++---
.../media/starfive/camss/stf-capture.c | 26 +++++++++----------
.../staging/media/starfive/camss/stf-video.h | 2 +-
5 files changed, 19 insertions(+), 19 deletions(-)
--
2.25.1
Comments
Hi Changhuang,
On 01/03/2024 8:00 am, Changhuang Liang wrote:
> The pixel formats captured by the capture_raw video device depends on
> what pixel formats come from the source device. It is actually dump
> the source device data. So renamed it to capture_dump.
I don't think 'dump' is a great name. I think what you really want to
say is 'unprocessed'. And usually that is indeed called 'raw capture'.
>
> Signed-off-by: Changhuang Liang <changhuang.liang@starfivetech.com>
> ---
> .../admin-guide/media/starfive_camss.rst | 2 +-
> .../media/starfive_camss_graph.dot | 2 +-
> .../staging/media/starfive/camss/stf-camss.c | 6 ++---
> .../media/starfive/camss/stf-capture.c | 26 +++++++++----------
> .../staging/media/starfive/camss/stf-video.h | 2 +-
> 5 files changed, 19 insertions(+), 19 deletions(-)
>
> diff --git a/Documentation/admin-guide/media/starfive_camss.rst b/Documentation/admin-guide/media/starfive_camss.rst
> index ca42e9447c47..c224e6123042 100644
> --- a/Documentation/admin-guide/media/starfive_camss.rst
> +++ b/Documentation/admin-guide/media/starfive_camss.rst
> @@ -60,7 +60,7 @@ The media controller pipeline graph is as follows:
>
> The driver has 2 video devices:
>
> -- capture_raw: The capture device, capturing image data directly from a sensor.
> +- capture_dump: The capture device, capturing image data directly from a sensor.
So perhaps rather than renaming everything, it would be better to explain it
better here:
- capture_raw: The capture device, capturing image data directly from a sensor, bypassing
the ISP module.
Regards,
Hans
> - capture_yuv: The capture device, capturing YUV frame data processed by the
> ISP module
>
> diff --git a/Documentation/admin-guide/media/starfive_camss_graph.dot b/Documentation/admin-guide/media/starfive_camss_graph.dot
> index 8eff1f161ac7..5e8731e27701 100644
> --- a/Documentation/admin-guide/media/starfive_camss_graph.dot
> +++ b/Documentation/admin-guide/media/starfive_camss_graph.dot
> @@ -2,7 +2,7 @@ digraph board {
> rankdir=TB
> n00000001 [label="{{<port0> 0} | stf_isp\n/dev/v4l-subdev0 | {<port1> 1}}", shape=Mrecord, style=filled, fillcolor=green]
> n00000001:port1 -> n00000008 [style=dashed]
> - n00000004 [label="capture_raw\n/dev/video0", shape=box, style=filled, fillcolor=yellow]
> + n00000004 [label="capture_dump\n/dev/video0", shape=box, style=filled, fillcolor=yellow]
> n00000008 [label="capture_yuv\n/dev/video1", shape=box, style=filled, fillcolor=yellow]
> n0000000e [label="{{<port0> 0} | cdns_csi2rx.19800000.csi-bridge\n | {<port1> 1 | <port2> 2 | <port3> 3 | <port4> 4}}", shape=Mrecord, style=filled, fillcolor=green]
> n0000000e:port1 -> n00000001:port0 [style=dashed]
> diff --git a/drivers/staging/media/starfive/camss/stf-camss.c b/drivers/staging/media/starfive/camss/stf-camss.c
> index a587f860101a..81fc39f20615 100644
> --- a/drivers/staging/media/starfive/camss/stf-camss.c
> +++ b/drivers/staging/media/starfive/camss/stf-camss.c
> @@ -176,7 +176,7 @@ static int stfcamss_subdev_notifier_bound(struct v4l2_async_notifier *async,
> container_of(asc, struct stfcamss_async_subdev, asd);
> enum stf_port_num port = csd->port;
> struct stf_isp_dev *isp_dev = &stfcamss->isp_dev;
> - struct stf_capture *cap_raw = &stfcamss->captures[STF_CAPTURE_RAW];
> + struct stf_capture *cap_dump = &stfcamss->captures[STF_CAPTURE_DUMP];
> struct media_pad *pad;
> int ret;
>
> @@ -192,12 +192,12 @@ static int stfcamss_subdev_notifier_bound(struct v4l2_async_notifier *async,
> return ret;
>
> ret = media_create_pad_link(&subdev->entity, 1,
> - &cap_raw->video.vdev.entity, 0, 0);
> + &cap_dump->video.vdev.entity, 0, 0);
> if (ret)
> return ret;
>
> isp_dev->source_subdev = subdev;
> - cap_raw->video.source_subdev = subdev;
> + cap_dump->video.source_subdev = subdev;
>
> return 0;
> }
> diff --git a/drivers/staging/media/starfive/camss/stf-capture.c b/drivers/staging/media/starfive/camss/stf-capture.c
> index ec5169e7b391..5c91126d5132 100644
> --- a/drivers/staging/media/starfive/camss/stf-capture.c
> +++ b/drivers/staging/media/starfive/camss/stf-capture.c
> @@ -10,7 +10,7 @@
> #include "stf-camss.h"
>
> static const char * const stf_cap_names[] = {
> - "capture_raw",
> + "capture_dump",
> "capture_yuv",
> };
>
> @@ -60,7 +60,7 @@ static inline struct stf_capture *to_stf_capture(struct stfcamss_video *video)
> return container_of(video, struct stf_capture, video);
> }
>
> -static void stf_set_raw_addr(struct stfcamss *stfcamss, dma_addr_t addr)
> +static void stf_set_dump_addr(struct stfcamss *stfcamss, dma_addr_t addr)
> {
> stf_syscon_reg_write(stfcamss, VIN_START_ADDR_O, (long)addr);
> stf_syscon_reg_write(stfcamss, VIN_START_ADDR_N, (long)addr);
> @@ -87,8 +87,8 @@ static void stf_init_addrs(struct stfcamss_video *video)
> addr0 = output->buf[0]->addr[0];
> addr1 = output->buf[0]->addr[1];
>
> - if (cap->type == STF_CAPTURE_RAW)
> - stf_set_raw_addr(video->stfcamss, addr0);
> + if (cap->type == STF_CAPTURE_DUMP)
> + stf_set_dump_addr(video->stfcamss, addr0);
> else if (cap->type == STF_CAPTURE_YUV)
> stf_set_yuv_addr(video->stfcamss, addr0, addr1);
> }
> @@ -179,7 +179,7 @@ static void stf_channel_set(struct stfcamss_video *video)
> struct stfcamss *stfcamss = cap->video.stfcamss;
> u32 val;
>
> - if (cap->type == STF_CAPTURE_RAW) {
> + if (cap->type == STF_CAPTURE_DUMP) {
> val = stf_syscon_reg_read(stfcamss, VIN_CHANNEL_SEL_EN);
> val &= ~U0_VIN_CHANNEL_SEL_MASK;
> val |= CHANNEL(0);
> @@ -219,7 +219,7 @@ static void stf_capture_start(struct stfcamss_video *video)
> struct stf_capture *cap = to_stf_capture(video);
>
> stf_channel_set(video);
> - if (cap->type == STF_CAPTURE_RAW) {
> + if (cap->type == STF_CAPTURE_DUMP) {
> stf_wr_irq_enable(video);
> stf_wr_data_en(video);
> }
> @@ -231,7 +231,7 @@ static void stf_capture_stop(struct stfcamss_video *video)
> {
> struct stf_capture *cap = to_stf_capture(video);
>
> - if (cap->type == STF_CAPTURE_RAW)
> + if (cap->type == STF_CAPTURE_DUMP)
> stf_wr_irq_disable(video);
>
> stf_cap_s_cleanup(video);
> @@ -252,7 +252,7 @@ static void stf_capture_init(struct stfcamss *stfcamss, struct stf_capture *cap)
> cap->video.stfcamss = stfcamss;
> cap->video.bpl_alignment = 16 * 8;
>
> - if (cap->type == STF_CAPTURE_RAW) {
> + if (cap->type == STF_CAPTURE_DUMP) {
> cap->video.formats = stf_wr_fmts;
> cap->video.nformats = ARRAY_SIZE(stf_wr_fmts);
> cap->video.bpl_alignment = 8;
> @@ -437,8 +437,8 @@ static void stf_change_buffer(struct stf_v_buf *output)
> if (output->state == STF_OUTPUT_STOPPING) {
> output->last_buffer = ready_buf;
> } else {
> - if (cap->type == STF_CAPTURE_RAW)
> - stf_set_raw_addr(stfcamss, new_addr[0]);
> + if (cap->type == STF_CAPTURE_DUMP)
> + stf_set_dump_addr(stfcamss, new_addr[0]);
> else if (cap->type == STF_CAPTURE_YUV)
> stf_set_yuv_addr(stfcamss, new_addr[0], new_addr[1]);
>
> @@ -452,7 +452,7 @@ static void stf_change_buffer(struct stf_v_buf *output)
> irqreturn_t stf_wr_irq_handler(int irq, void *priv)
> {
> struct stfcamss *stfcamss = priv;
> - struct stf_capture *cap = &stfcamss->captures[STF_CAPTURE_RAW];
> + struct stf_capture *cap = &stfcamss->captures[STF_CAPTURE_DUMP];
>
> if (atomic_dec_if_positive(&cap->buffers.frame_skip) < 0) {
> stf_change_buffer(&cap->buffers);
> @@ -569,10 +569,10 @@ static void stf_capture_unregister_one(struct stf_capture *cap)
>
> void stf_capture_unregister(struct stfcamss *stfcamss)
> {
> - struct stf_capture *cap_raw = &stfcamss->captures[STF_CAPTURE_RAW];
> + struct stf_capture *cap_dump = &stfcamss->captures[STF_CAPTURE_DUMP];
> struct stf_capture *cap_yuv = &stfcamss->captures[STF_CAPTURE_YUV];
>
> - stf_capture_unregister_one(cap_raw);
> + stf_capture_unregister_one(cap_dump);
> stf_capture_unregister_one(cap_yuv);
> }
>
> diff --git a/drivers/staging/media/starfive/camss/stf-video.h b/drivers/staging/media/starfive/camss/stf-video.h
> index 8052b77e3ad8..90c73c0dee89 100644
> --- a/drivers/staging/media/starfive/camss/stf-video.h
> +++ b/drivers/staging/media/starfive/camss/stf-video.h
> @@ -35,7 +35,7 @@ enum stf_v_line_id {
> };
>
> enum stf_capture_type {
> - STF_CAPTURE_RAW = 0,
> + STF_CAPTURE_DUMP = 0,
> STF_CAPTURE_YUV,
> STF_CAPTURE_NUM,
> };
> --
> 2.25.1
>
Hi, Hans
Thanks for your comments.
> On 01/03/2024 8:00 am, Changhuang Liang wrote:
> > The pixel formats captured by the capture_raw video device depends on
> > what pixel formats come from the source device. It is actually dump
> > the source device data. So renamed it to capture_dump.
>
> I don't think 'dump' is a great name. I think what you really want to say is
> 'unprocessed'. And usually that is indeed called 'raw capture'.
>
StarFive JH7110 SoC also supports the output of the raw format data from ISP port, see:
https://lore.kernel.org/all/20240306093334.9321-1-changhuang.liang@starfivetech.com/
So I think it is better to call the ISP output port "capture_raw".
On the other hand, Old "capture_raw" video devices can also support YUV color space,
mainly depending on the camera output format. There may be a confusion called it "capture_raw".
Regards,
Changhuang
On 21/03/2024 1:17 pm, Changhuang Liang wrote:
> Hi, Hans
>
> Thanks for your comments.
>
>> On 01/03/2024 8:00 am, Changhuang Liang wrote:
>>> The pixel formats captured by the capture_raw video device depends on
>>> what pixel formats come from the source device. It is actually dump
>>> the source device data. So renamed it to capture_dump.
>>
>> I don't think 'dump' is a great name. I think what you really want to say is
>> 'unprocessed'. And usually that is indeed called 'raw capture'.
>>
>
> StarFive JH7110 SoC also supports the output of the raw format data from ISP port, see:
>
> https://lore.kernel.org/all/20240306093334.9321-1-changhuang.liang@starfivetech.com/
>
> So I think it is better to call the ISP output port "capture_raw".
If it is processed by the ISP, how can it be 'raw'? Or do you mean that this uses Bayer pixelformat?
The normal convention AFAIK is that the term 'raw capture' generally means that
the sensor bayer data is captured without being processed (or only minimal processing).
Looking at:
https://lore.kernel.org/all/20240306093334.9321-6-changhuang.liang@starfivetech.com/
I think you indeed mean that you can get the processed ISP data in bayer format
through a separate video device. So I would call that port 'capture_bayer' and keep the
'capture_raw' name for the port from the CSI receiver.
Regards,
Hans
>
> On the other hand, Old "capture_raw" video devices can also support YUV color space,
> mainly depending on the camera output format. There may be a confusion called it "capture_raw".
>
> Regards,
> Changhuang
>
>
> On 21/03/2024 1:17 pm, Changhuang Liang wrote:
> > Hi, Hans
> >
> > Thanks for your comments.
> >
> >> On 01/03/2024 8:00 am, Changhuang Liang wrote:
> >>> The pixel formats captured by the capture_raw video device depends
> >>> on what pixel formats come from the source device. It is actually
> >>> dump the source device data. So renamed it to capture_dump.
> >>
> >> I don't think 'dump' is a great name. I think what you really want to
> >> say is 'unprocessed'. And usually that is indeed called 'raw capture'.
> >>
> >
> > StarFive JH7110 SoC also supports the output of the raw format data from
> ISP port, see:
> >
> > https://lore.kernel.org/all/20240306093334.9321-1-changhuang.liang@sta
> > rfivetech.com/
> >
> > So I think it is better to call the ISP output port "capture_raw".
>
> If it is processed by the ISP, how can it be 'raw'? Or do you mean that this uses
> Bayer pixelformat?
>
> The normal convention AFAIK is that the term 'raw capture' generally means
> that the sensor bayer data is captured without being processed (or only
> minimal processing).
>
> Looking at:
>
> https://lore.kernel.org/all/20240306093334.9321-6-changhuang.liang@starfi
> vetech.com/
>
> I think you indeed mean that you can get the processed ISP data in bayer
> format through a separate video device. So I would call that port
> 'capture_bayer' and keep the 'capture_raw' name for the port from the CSI
> receiver.
>
OK, will use this way.
Regards,
Changhuang
@@ -60,7 +60,7 @@ The media controller pipeline graph is as follows:
The driver has 2 video devices:
-- capture_raw: The capture device, capturing image data directly from a sensor.
+- capture_dump: The capture device, capturing image data directly from a sensor.
- capture_yuv: The capture device, capturing YUV frame data processed by the
ISP module
@@ -2,7 +2,7 @@ digraph board {
rankdir=TB
n00000001 [label="{{<port0> 0} | stf_isp\n/dev/v4l-subdev0 | {<port1> 1}}", shape=Mrecord, style=filled, fillcolor=green]
n00000001:port1 -> n00000008 [style=dashed]
- n00000004 [label="capture_raw\n/dev/video0", shape=box, style=filled, fillcolor=yellow]
+ n00000004 [label="capture_dump\n/dev/video0", shape=box, style=filled, fillcolor=yellow]
n00000008 [label="capture_yuv\n/dev/video1", shape=box, style=filled, fillcolor=yellow]
n0000000e [label="{{<port0> 0} | cdns_csi2rx.19800000.csi-bridge\n | {<port1> 1 | <port2> 2 | <port3> 3 | <port4> 4}}", shape=Mrecord, style=filled, fillcolor=green]
n0000000e:port1 -> n00000001:port0 [style=dashed]
@@ -176,7 +176,7 @@ static int stfcamss_subdev_notifier_bound(struct v4l2_async_notifier *async,
container_of(asc, struct stfcamss_async_subdev, asd);
enum stf_port_num port = csd->port;
struct stf_isp_dev *isp_dev = &stfcamss->isp_dev;
- struct stf_capture *cap_raw = &stfcamss->captures[STF_CAPTURE_RAW];
+ struct stf_capture *cap_dump = &stfcamss->captures[STF_CAPTURE_DUMP];
struct media_pad *pad;
int ret;
@@ -192,12 +192,12 @@ static int stfcamss_subdev_notifier_bound(struct v4l2_async_notifier *async,
return ret;
ret = media_create_pad_link(&subdev->entity, 1,
- &cap_raw->video.vdev.entity, 0, 0);
+ &cap_dump->video.vdev.entity, 0, 0);
if (ret)
return ret;
isp_dev->source_subdev = subdev;
- cap_raw->video.source_subdev = subdev;
+ cap_dump->video.source_subdev = subdev;
return 0;
}
@@ -10,7 +10,7 @@
#include "stf-camss.h"
static const char * const stf_cap_names[] = {
- "capture_raw",
+ "capture_dump",
"capture_yuv",
};
@@ -60,7 +60,7 @@ static inline struct stf_capture *to_stf_capture(struct stfcamss_video *video)
return container_of(video, struct stf_capture, video);
}
-static void stf_set_raw_addr(struct stfcamss *stfcamss, dma_addr_t addr)
+static void stf_set_dump_addr(struct stfcamss *stfcamss, dma_addr_t addr)
{
stf_syscon_reg_write(stfcamss, VIN_START_ADDR_O, (long)addr);
stf_syscon_reg_write(stfcamss, VIN_START_ADDR_N, (long)addr);
@@ -87,8 +87,8 @@ static void stf_init_addrs(struct stfcamss_video *video)
addr0 = output->buf[0]->addr[0];
addr1 = output->buf[0]->addr[1];
- if (cap->type == STF_CAPTURE_RAW)
- stf_set_raw_addr(video->stfcamss, addr0);
+ if (cap->type == STF_CAPTURE_DUMP)
+ stf_set_dump_addr(video->stfcamss, addr0);
else if (cap->type == STF_CAPTURE_YUV)
stf_set_yuv_addr(video->stfcamss, addr0, addr1);
}
@@ -179,7 +179,7 @@ static void stf_channel_set(struct stfcamss_video *video)
struct stfcamss *stfcamss = cap->video.stfcamss;
u32 val;
- if (cap->type == STF_CAPTURE_RAW) {
+ if (cap->type == STF_CAPTURE_DUMP) {
val = stf_syscon_reg_read(stfcamss, VIN_CHANNEL_SEL_EN);
val &= ~U0_VIN_CHANNEL_SEL_MASK;
val |= CHANNEL(0);
@@ -219,7 +219,7 @@ static void stf_capture_start(struct stfcamss_video *video)
struct stf_capture *cap = to_stf_capture(video);
stf_channel_set(video);
- if (cap->type == STF_CAPTURE_RAW) {
+ if (cap->type == STF_CAPTURE_DUMP) {
stf_wr_irq_enable(video);
stf_wr_data_en(video);
}
@@ -231,7 +231,7 @@ static void stf_capture_stop(struct stfcamss_video *video)
{
struct stf_capture *cap = to_stf_capture(video);
- if (cap->type == STF_CAPTURE_RAW)
+ if (cap->type == STF_CAPTURE_DUMP)
stf_wr_irq_disable(video);
stf_cap_s_cleanup(video);
@@ -252,7 +252,7 @@ static void stf_capture_init(struct stfcamss *stfcamss, struct stf_capture *cap)
cap->video.stfcamss = stfcamss;
cap->video.bpl_alignment = 16 * 8;
- if (cap->type == STF_CAPTURE_RAW) {
+ if (cap->type == STF_CAPTURE_DUMP) {
cap->video.formats = stf_wr_fmts;
cap->video.nformats = ARRAY_SIZE(stf_wr_fmts);
cap->video.bpl_alignment = 8;
@@ -437,8 +437,8 @@ static void stf_change_buffer(struct stf_v_buf *output)
if (output->state == STF_OUTPUT_STOPPING) {
output->last_buffer = ready_buf;
} else {
- if (cap->type == STF_CAPTURE_RAW)
- stf_set_raw_addr(stfcamss, new_addr[0]);
+ if (cap->type == STF_CAPTURE_DUMP)
+ stf_set_dump_addr(stfcamss, new_addr[0]);
else if (cap->type == STF_CAPTURE_YUV)
stf_set_yuv_addr(stfcamss, new_addr[0], new_addr[1]);
@@ -452,7 +452,7 @@ static void stf_change_buffer(struct stf_v_buf *output)
irqreturn_t stf_wr_irq_handler(int irq, void *priv)
{
struct stfcamss *stfcamss = priv;
- struct stf_capture *cap = &stfcamss->captures[STF_CAPTURE_RAW];
+ struct stf_capture *cap = &stfcamss->captures[STF_CAPTURE_DUMP];
if (atomic_dec_if_positive(&cap->buffers.frame_skip) < 0) {
stf_change_buffer(&cap->buffers);
@@ -569,10 +569,10 @@ static void stf_capture_unregister_one(struct stf_capture *cap)
void stf_capture_unregister(struct stfcamss *stfcamss)
{
- struct stf_capture *cap_raw = &stfcamss->captures[STF_CAPTURE_RAW];
+ struct stf_capture *cap_dump = &stfcamss->captures[STF_CAPTURE_DUMP];
struct stf_capture *cap_yuv = &stfcamss->captures[STF_CAPTURE_YUV];
- stf_capture_unregister_one(cap_raw);
+ stf_capture_unregister_one(cap_dump);
stf_capture_unregister_one(cap_yuv);
}
@@ -35,7 +35,7 @@ enum stf_v_line_id {
};
enum stf_capture_type {
- STF_CAPTURE_RAW = 0,
+ STF_CAPTURE_DUMP = 0,
STF_CAPTURE_YUV,
STF_CAPTURE_NUM,
};