Patch Detail
Show a patch.
GET /api/patches/16743/?format=api
{ "id": 16743, "url": "https://patchwork.libcamera.org/api/patches/16743/?format=api", "web_url": "https://patchwork.libcamera.org/patch/16743/", "project": { "id": 1, "url": "https://patchwork.libcamera.org/api/projects/1/?format=api", "name": "libcamera", "link_name": "libcamera", "list_id": "libcamera_core", "list_email": "libcamera-devel@lists.libcamera.org", "web_url": "", "scm_url": "", "webscm_url": "" }, "msgid": "<20220723095330.43542-5-jacopo@jmondi.org>", "date": "2022-07-23T09:53:26", "name": "[libcamera-devel,v2,4/8] libcamera: v4l2_videodevice: Reintroduce toV4L2PixelFormat()", "commit_ref": null, "pull_url": null, "state": "accepted", "archived": false, "hash": "1cbfa3e5e82847a83661ab0ee342f53e137737df", "submitter": { "id": 3, "url": "https://patchwork.libcamera.org/api/people/3/?format=api", "name": "Jacopo Mondi", "email": "jacopo@jmondi.org" }, "delegate": null, "mbox": "https://patchwork.libcamera.org/patch/16743/mbox/", "series": [ { "id": 3320, "url": "https://patchwork.libcamera.org/api/series/3320/?format=api", "web_url": "https://patchwork.libcamera.org/project/libcamera/list/?series=3320", "date": "2022-07-23T09:53:22", "name": "libcamera: Map multiple V4L2 formats to a single libcamera::format", "version": 2, "mbox": "https://patchwork.libcamera.org/series/3320/mbox/" } ], "comments": "https://patchwork.libcamera.org/api/patches/16743/comments/", "check": "pending", "checks": "https://patchwork.libcamera.org/api/patches/16743/checks/", "tags": {}, "headers": { "Return-Path": "<libcamera-devel-bounces@lists.libcamera.org>", "X-Original-To": "parsemail@patchwork.libcamera.org", "Delivered-To": "parsemail@patchwork.libcamera.org", "Received": [ "from lancelot.ideasonboard.com (lancelot.ideasonboard.com\n\t[92.243.16.209])\n\tby patchwork.libcamera.org (Postfix) with ESMTPS id 28774BE173\n\tfor <parsemail@patchwork.libcamera.org>;\n\tSat, 23 Jul 2022 09:54:00 +0000 (UTC)", "from lancelot.ideasonboard.com (localhost [IPv6:::1])\n\tby lancelot.ideasonboard.com (Postfix) with ESMTP id DD8DD63315;\n\tSat, 23 Jul 2022 11:53:59 +0200 (CEST)", "from relay8-d.mail.gandi.net (relay8-d.mail.gandi.net\n\t[217.70.183.201])\n\tby lancelot.ideasonboard.com (Postfix) with ESMTPS id EE25C63316\n\tfor <libcamera-devel@lists.libcamera.org>;\n\tSat, 23 Jul 2022 11:53:57 +0200 (CEST)", "(Authenticated sender: jacopo@jmondi.org)\n\tby mail.gandi.net (Postfix) with ESMTPSA id 06D501BF20A;\n\tSat, 23 Jul 2022 09:53:55 +0000 (UTC)" ], "DKIM-Signature": "v=1; a=rsa-sha256; c=relaxed/simple; d=libcamera.org;\n\ts=mail; t=1658570039;\n\tbh=ZRDFbJH/EIccVvGQT805+JELa2SJPgXnKK4tqWwT1Ok=;\n\th=To:Date:In-Reply-To:References:Subject:List-Id:List-Unsubscribe:\n\tList-Archive:List-Post:List-Help:List-Subscribe:From:Reply-To:Cc:\n\tFrom;\n\tb=ReTgWiXxQBwX4i21ruW2J7MmtSQqOpVOWo1Li9YM4YCMfY0XkpYknmoVy6QR3wKS4\n\tBKmFqCCcHQjB58iVAYPgMiszSrYOxSiM39cYf4dhUVEu4taJ0RipUvUoOBTmIrXCbt\n\t5O/gXvy+9je2rkdVk8Pq16FidoJvGUXt8Tk5x+5ORVdU0PwCzCFRUAWqiDXKq5g4H4\n\tPQgEcSNFHmUbpuVl+3IiaJomHQaAx7ZWINc3RLc4M9L6cypKwTWqj+SMw9qnW6J4Ng\n\tGkfyTnRmG8a2bE33f+zXHUN9mumHw8wFyzNHBj6mgJyuJQLy3+l1vtQiw8PTV3l/GI\n\tHpZVr3ME+mw3g==", "To": "libcamera-devel@lists.libcamera.org", "Date": "Sat, 23 Jul 2022 11:53:26 +0200", "Message-Id": "<20220723095330.43542-5-jacopo@jmondi.org>", "X-Mailer": "git-send-email 2.37.1", "In-Reply-To": "<20220723095330.43542-1-jacopo@jmondi.org>", "References": "<20220723095330.43542-1-jacopo@jmondi.org>", "MIME-Version": "1.0", "Content-Transfer-Encoding": "8bit", "Subject": "[libcamera-devel] [PATCH v2 4/8] libcamera: v4l2_videodevice:\n\tReintroduce toV4L2PixelFormat()", "X-BeenThere": "libcamera-devel@lists.libcamera.org", "X-Mailman-Version": "2.1.29", "Precedence": "list", "List-Id": "<libcamera-devel.lists.libcamera.org>", "List-Unsubscribe": "<https://lists.libcamera.org/options/libcamera-devel>,\n\t<mailto:libcamera-devel-request@lists.libcamera.org?subject=unsubscribe>", "List-Archive": "<https://lists.libcamera.org/pipermail/libcamera-devel/>", "List-Post": "<mailto:libcamera-devel@lists.libcamera.org>", "List-Help": "<mailto:libcamera-devel-request@lists.libcamera.org?subject=help>", "List-Subscribe": "<https://lists.libcamera.org/listinfo/libcamera-devel>,\n\t<mailto:libcamera-devel-request@lists.libcamera.org?subject=subscribe>", "From": "Jacopo Mondi via libcamera-devel <libcamera-devel@lists.libcamera.org>", "Reply-To": "Jacopo Mondi <jacopo@jmondi.org>", "Cc": "jozef@mlich.cz, Pavel Machek <pavel@ucw.cz>", "Errors-To": "libcamera-devel-bounces@lists.libcamera.org", "Sender": "\"libcamera-devel\" <libcamera-devel-bounces@lists.libcamera.org>" }, "content": "This is a partial revert of commit 395d43d6d75b (\"libcamera:\nv4l2_videodevice: Drop toV4L2PixelFormat()\")\n\nThe function was removed because it incorrectly maps non-contigous V4L2\nformat variants (ie V4L2_PIX_FMT_YUV420M) to the API version supported\nby the video device (singleplanar API and multiplanar API). It was\ndecided at the time to remove the function and let its users call\ndirectly V4L2PixelFormat::fromPixelFormat() which accepts a\n'multiplanar' flags.\n\nAs we now aim to associate multiple V4L2PixelFormat to a single\nlibcamera format, it is necessary to verify which of them is actually\nsupported by the video device.\n\nHence re-introduce V4L2VideoDevice::toV4L2PixelFormat() and convert all\nthe V4L2PixelFormat::fromPixelFormat() users to use it.\n\nThe V4L2 compatibility layer is the only outliner as it doesn't have a\nvideo device to poke, hence it still uses\nV4L2PixelFormat::fromPixelFormat() and defaults to the first format.\n\nNext patches will implement the device format matching logic and handle\nthe non-contiguous plane issue in V4L2VideoDevice::toV4L2PixelFormat().\n\nSigned-off-by: Jacopo Mondi <jacopo@jmondi.org>\n---\n Documentation/guides/pipeline-handler.rst | 7 ++--\n include/libcamera/internal/v4l2_videodevice.h | 2 ++\n src/libcamera/pipeline/ipu3/cio2.cpp | 2 +-\n src/libcamera/pipeline/ipu3/imgu.cpp | 2 +-\n .../pipeline/raspberrypi/raspberrypi.cpp | 34 +++++++++++--------\n src/libcamera/pipeline/rkisp1/rkisp1_path.cpp | 6 ++--\n src/libcamera/pipeline/simple/converter.cpp | 10 +++---\n src/libcamera/pipeline/simple/simple.cpp | 2 +-\n src/libcamera/pipeline/uvcvideo/uvcvideo.cpp | 4 +--\n src/libcamera/pipeline/vimc/vimc.cpp | 8 ++---\n src/libcamera/v4l2_videodevice.cpp | 17 ++++++++++\n test/libtest/buffer_source.cpp | 2 +-\n 12 files changed, 61 insertions(+), 35 deletions(-)", "diff": "diff --git a/Documentation/guides/pipeline-handler.rst b/Documentation/guides/pipeline-handler.rst\nindex 2d55666d094f..dcb6971e46f0 100644\n--- a/Documentation/guides/pipeline-handler.rst\n+++ b/Documentation/guides/pipeline-handler.rst\n@@ -971,7 +971,8 @@ with the fourcc and size attributes to apply directly to the capture device\n node. The fourcc attribute is a `V4L2PixelFormat`_ and differs from the\n ``libcamera::PixelFormat``. Converting the format requires knowledge of the\n plane configuration for multiplanar formats, so you must explicitly convert it\n-using the helper ``V4L2PixelFormat::fromPixelFormat()``.\n+using the helper ``V4L2VideoDevice::toV4L2PixelFormat()`` provided by the\n+V4L2VideoDevice instance of which the format will be applied on.\n \n .. _V4L2DeviceFormat: https://libcamera.org/api-html/classlibcamera_1_1V4L2DeviceFormat.html\n .. _V4L2PixelFormat: https://libcamera.org/api-html/classlibcamera_1_1V4L2PixelFormat.html\n@@ -981,7 +982,7 @@ Add the following code beneath the code from above:\n .. code-block:: cpp\n \n V4L2DeviceFormat format = {};\n- format.fourcc = V4L2PixelFormat::fromPixelFormat(cfg.pixelFormat);\n+ format.fourcc = data->video_->toV4L2PixelFormat(cfg.pixelFormat);\n format.size = cfg.size;\n \n Set the video device format defined above using the\n@@ -1001,7 +1002,7 @@ Continue the implementation with the following code:\n return ret;\n \n if (format.size != cfg.size ||\n- format.fourcc != V4L2PixelFormat::fromPixelFormat(cfg.pixelFormat))\n+ format.fourcc != data->video_->toV4L2PixelFormat(cfg.pixelFormat))\n return -EINVAL;\n \n Finally, store and set stream-specific data reflecting the state of the stream.\ndiff --git a/include/libcamera/internal/v4l2_videodevice.h b/include/libcamera/internal/v4l2_videodevice.h\nindex 8525acbc558d..29fa0bbaf670 100644\n--- a/include/libcamera/internal/v4l2_videodevice.h\n+++ b/include/libcamera/internal/v4l2_videodevice.h\n@@ -228,6 +228,8 @@ public:\n \tstatic std::unique_ptr<V4L2VideoDevice>\n \tfromEntityName(const MediaDevice *media, const std::string &entity);\n \n+\tV4L2PixelFormat toV4L2PixelFormat(const PixelFormat &pixelFormat) const;\n+\n protected:\n \tstd::string logPrefix() const override;\n \ndiff --git a/src/libcamera/pipeline/ipu3/cio2.cpp b/src/libcamera/pipeline/ipu3/cio2.cpp\nindex 4a13d16973e9..d4e523af24b4 100644\n--- a/src/libcamera/pipeline/ipu3/cio2.cpp\n+++ b/src/libcamera/pipeline/ipu3/cio2.cpp\n@@ -203,7 +203,7 @@ int CIO2Device::configure(const Size &size, V4L2DeviceFormat *outputFormat)\n \tif (itInfo == mbusCodesToPixelFormat.end())\n \t\treturn -EINVAL;\n \n-\toutputFormat->fourcc = V4L2PixelFormat::fromPixelFormat(itInfo->second)[0];\n+\toutputFormat->fourcc = output_->toV4L2PixelFormat(itInfo->second);\n \toutputFormat->size = sensorFormat.size;\n \toutputFormat->planesCount = 1;\n \ndiff --git a/src/libcamera/pipeline/ipu3/imgu.cpp b/src/libcamera/pipeline/ipu3/imgu.cpp\nindex 251164001ffd..531879f18183 100644\n--- a/src/libcamera/pipeline/ipu3/imgu.cpp\n+++ b/src/libcamera/pipeline/ipu3/imgu.cpp\n@@ -558,7 +558,7 @@ int ImgUDevice::configureVideoDevice(V4L2VideoDevice *dev, unsigned int pad,\n \t\treturn 0;\n \n \t*outputFormat = {};\n-\toutputFormat->fourcc = V4L2PixelFormat::fromPixelFormat(formats::NV12)[0];\n+\toutputFormat->fourcc = dev->toV4L2PixelFormat(formats::NV12);\n \toutputFormat->size = cfg.size;\n \toutputFormat->planesCount = 2;\n \ndiff --git a/src/libcamera/pipeline/raspberrypi/raspberrypi.cpp b/src/libcamera/pipeline/raspberrypi/raspberrypi.cpp\nindex 6e813bcb094a..4f60224729e3 100644\n--- a/src/libcamera/pipeline/raspberrypi/raspberrypi.cpp\n+++ b/src/libcamera/pipeline/raspberrypi/raspberrypi.cpp\n@@ -90,13 +90,14 @@ PixelFormat mbusCodeToPixelFormat(unsigned int mbus_code,\n \treturn pix;\n }\n \n-V4L2DeviceFormat toV4L2DeviceFormat(const V4L2SubdeviceFormat &format,\n+V4L2DeviceFormat toV4L2DeviceFormat(const V4L2VideoDevice *dev,\n+\t\t\t\t const V4L2SubdeviceFormat &format,\n \t\t\t\t BayerFormat::Packing packingReq)\n {\n \tconst PixelFormat pix = mbusCodeToPixelFormat(format.mbus_code, packingReq);\n \tV4L2DeviceFormat deviceFormat;\n \n-\tdeviceFormat.fourcc = V4L2PixelFormat::fromPixelFormat(pix)[0];\n+\tdeviceFormat.fourcc = dev->toV4L2PixelFormat(pix);\n \tdeviceFormat.size = format.size;\n \tdeviceFormat.colorSpace = format.colorSpace;\n \treturn deviceFormat;\n@@ -422,15 +423,15 @@ CameraConfiguration::Status RPiCameraConfiguration::validate()\n \t\t\t * Calculate the best sensor mode we can use based on\n \t\t\t * the user request.\n \t\t\t */\n+\t\t\tV4L2VideoDevice *unicam = data_->unicam_[Unicam::Image].dev();\n \t\t\tconst PixelFormatInfo &info = PixelFormatInfo::info(cfg.pixelFormat);\n \t\t\tunsigned int bitDepth = info.isValid() ? info.bitsPerPixel : defaultRawBitDepth;\n \t\t\tV4L2SubdeviceFormat sensorFormat = findBestFormat(data_->sensorFormats_, cfg.size, bitDepth);\n \t\t\tBayerFormat::Packing packing = BayerFormat::Packing::CSI2;\n \t\t\tif (info.isValid() && !info.packed)\n \t\t\t\tpacking = BayerFormat::Packing::None;\n-\t\t\tV4L2DeviceFormat unicamFormat = toV4L2DeviceFormat(sensorFormat,\n-\t\t\t\t\t\t\t\t\t packing);\n-\t\t\tint ret = data_->unicam_[Unicam::Image].dev()->tryFormat(&unicamFormat);\n+\t\t\tV4L2DeviceFormat unicamFormat = toV4L2DeviceFormat(unicam, sensorFormat, packing);\n+\t\t\tint ret = unicam->tryFormat(&unicamFormat);\n \t\t\tif (ret)\n \t\t\t\treturn Invalid;\n \n@@ -516,14 +517,14 @@ CameraConfiguration::Status RPiCameraConfiguration::validate()\n \n \t\tV4L2VideoDevice::Formats fmts = dev->formats();\n \n-\t\tif (fmts.find(V4L2PixelFormat::fromPixelFormat(cfgPixFmt)[0]) == fmts.end()) {\n+\t\tif (fmts.find(dev->toV4L2PixelFormat(cfgPixFmt)) == fmts.end()) {\n \t\t\t/* If we cannot find a native format, use a default one. */\n \t\t\tcfgPixFmt = formats::NV12;\n \t\t\tstatus = Adjusted;\n \t\t}\n \n \t\tV4L2DeviceFormat format;\n-\t\tformat.fourcc = V4L2PixelFormat::fromPixelFormat(cfgPixFmt)[0];\n+\t\tformat.fourcc = dev->toV4L2PixelFormat(cfgPixFmt);\n \t\tformat.size = cfg.size;\n \t\tformat.colorSpace = cfg.colorSpace;\n \n@@ -751,8 +752,9 @@ int PipelineHandlerRPi::configure(Camera *camera, CameraConfiguration *config)\n \tif (ret)\n \t\treturn ret;\n \n-\tV4L2DeviceFormat unicamFormat = toV4L2DeviceFormat(sensorFormat, packing);\n-\tret = data->unicam_[Unicam::Image].dev()->setFormat(&unicamFormat);\n+\tV4L2VideoDevice *unicam = data->unicam_[Unicam::Image].dev();\n+\tV4L2DeviceFormat unicamFormat = toV4L2DeviceFormat(unicam, sensorFormat, packing);\n+\tret = unicam->setFormat(&unicamFormat);\n \tif (ret)\n \t\treturn ret;\n \n@@ -783,7 +785,7 @@ int PipelineHandlerRPi::configure(Camera *camera, CameraConfiguration *config)\n \t\tRPi::Stream *stream = i == maxIndex ? &data->isp_[Isp::Output0]\n \t\t\t\t\t\t : &data->isp_[Isp::Output1];\n \n-\t\tV4L2PixelFormat fourcc = V4L2PixelFormat::fromPixelFormat(cfg.pixelFormat)[0];\n+\t\tV4L2PixelFormat fourcc = stream->dev()->toV4L2PixelFormat(cfg.pixelFormat);\n \t\tformat.size = cfg.size;\n \t\tformat.fourcc = fourcc;\n \t\tformat.colorSpace = cfg.colorSpace;\n@@ -826,13 +828,15 @@ int PipelineHandlerRPi::configure(Camera *camera, CameraConfiguration *config)\n \t * statistics coming from the hardware.\n \t */\n \tif (!output0Set) {\n+\t\tV4L2VideoDevice *dev = data->isp_[Isp::Output0].dev();\n+\n \t\tmaxSize = Size(320, 240);\n \t\tformat = {};\n \t\tformat.size = maxSize;\n-\t\tformat.fourcc = V4L2PixelFormat::fromPixelFormat(formats::YUV420)[0];\n+\t\tformat.fourcc = dev->toV4L2PixelFormat(formats::YUV420);\n \t\t/* No one asked for output, so the color space doesn't matter. */\n \t\tformat.colorSpace = ColorSpace::Jpeg;\n-\t\tret = data->isp_[Isp::Output0].dev()->setFormat(&format);\n+\t\tret = dev->setFormat(&format);\n \t\tif (ret) {\n \t\t\tLOG(RPI, Error)\n \t\t\t\t<< \"Failed to set default format on ISP Output0: \"\n@@ -856,18 +860,20 @@ int PipelineHandlerRPi::configure(Camera *camera, CameraConfiguration *config)\n \t * colour denoise will not run.\n \t */\n \tif (!output1Set) {\n+\t\tV4L2VideoDevice *dev = data->isp_[Isp::Output1].dev();\n+\n \t\tV4L2DeviceFormat output1Format;\n \t\tconstexpr Size maxDimensions(1200, 1200);\n \t\tconst Size limit = maxDimensions.boundedToAspectRatio(format.size);\n \n \t\toutput1Format.size = (format.size / 2).boundedTo(limit).alignedDownTo(2, 2);\n \t\toutput1Format.colorSpace = format.colorSpace;\n-\t\toutput1Format.fourcc = V4L2PixelFormat::fromPixelFormat(formats::YUV420)[0];\n+\t\toutput1Format.fourcc = dev->toV4L2PixelFormat(formats::YUV420);\n \n \t\tLOG(RPI, Debug) << \"Setting ISP Output1 (internal) to \"\n \t\t\t\t<< output1Format;\n \n-\t\tret = data->isp_[Isp::Output1].dev()->setFormat(&output1Format);\n+\t\tret = dev->setFormat(&output1Format);\n \t\tif (ret) {\n \t\t\tLOG(RPI, Error) << \"Failed to set format on ISP Output1: \"\n \t\t\t\t\t<< ret;\ndiff --git a/src/libcamera/pipeline/rkisp1/rkisp1_path.cpp b/src/libcamera/pipeline/rkisp1/rkisp1_path.cpp\nindex f4e71bab00ed..2d38f0fb37ab 100644\n--- a/src/libcamera/pipeline/rkisp1/rkisp1_path.cpp\n+++ b/src/libcamera/pipeline/rkisp1/rkisp1_path.cpp\n@@ -81,7 +81,7 @@ CameraConfiguration::Status RkISP1Path::validate(StreamConfiguration *cfg)\n \tcfg->bufferCount = RKISP1_BUFFER_COUNT;\n \n \tV4L2DeviceFormat format;\n-\tformat.fourcc = V4L2PixelFormat::fromPixelFormat(cfg->pixelFormat)[0];\n+\tformat.fourcc = video_->toV4L2PixelFormat(cfg->pixelFormat);\n \tformat.size = cfg->size;\n \n \tint ret = video_->tryFormat(&format);\n@@ -147,7 +147,7 @@ int RkISP1Path::configure(const StreamConfiguration &config,\n \n \tconst PixelFormatInfo &info = PixelFormatInfo::info(config.pixelFormat);\n \tV4L2DeviceFormat outputFormat;\n-\toutputFormat.fourcc = V4L2PixelFormat::fromPixelFormat(config.pixelFormat)[0];\n+\toutputFormat.fourcc = video_->toV4L2PixelFormat(config.pixelFormat);\n \toutputFormat.size = config.size;\n \toutputFormat.planesCount = info.numPlanes();\n \n@@ -156,7 +156,7 @@ int RkISP1Path::configure(const StreamConfiguration &config,\n \t\treturn ret;\n \n \tif (outputFormat.size != config.size ||\n-\t outputFormat.fourcc != V4L2PixelFormat::fromPixelFormat(config.pixelFormat)[0]) {\n+\t outputFormat.fourcc != video_->toV4L2PixelFormat(config.pixelFormat)) {\n \t\tLOG(RkISP1, Error)\n \t\t\t<< \"Unable to configure capture in \" << config.toString();\n \t\treturn -EINVAL;\ndiff --git a/src/libcamera/pipeline/simple/converter.cpp b/src/libcamera/pipeline/simple/converter.cpp\nindex 47e2150b4332..acaaa64c949a 100644\n--- a/src/libcamera/pipeline/simple/converter.cpp\n+++ b/src/libcamera/pipeline/simple/converter.cpp\n@@ -46,7 +46,7 @@ int SimpleConverter::Stream::configure(const StreamConfiguration &inputCfg,\n \t\t\t\t const StreamConfiguration &outputCfg)\n {\n \tV4L2PixelFormat videoFormat =\n-\t\tV4L2PixelFormat::fromPixelFormat(inputCfg.pixelFormat)[0];\n+\t\tm2m_->output()->toV4L2PixelFormat(inputCfg.pixelFormat);\n \n \tV4L2DeviceFormat format;\n \tformat.fourcc = videoFormat;\n@@ -71,7 +71,7 @@ int SimpleConverter::Stream::configure(const StreamConfiguration &inputCfg,\n \t}\n \n \t/* Set the pixel format and size on the output. */\n-\tvideoFormat = V4L2PixelFormat::fromPixelFormat(outputCfg.pixelFormat)[0];\n+\tvideoFormat = m2m_->capture()->toV4L2PixelFormat(outputCfg.pixelFormat);\n \tformat = {};\n \tformat.fourcc = videoFormat;\n \tformat.size = outputCfg.size;\n@@ -210,7 +210,7 @@ std::vector<PixelFormat> SimpleConverter::formats(PixelFormat input)\n \t * enumerate the conversion capabilities on its output (V4L2 capture).\n \t */\n \tV4L2DeviceFormat v4l2Format;\n-\tv4l2Format.fourcc = V4L2PixelFormat::fromPixelFormat(input)[0];\n+\tv4l2Format.fourcc = m2m_->output()->toV4L2PixelFormat(input);\n \tv4l2Format.size = { 1, 1 };\n \n \tint ret = m2m_->output()->setFormat(&v4l2Format);\n@@ -220,7 +220,7 @@ std::vector<PixelFormat> SimpleConverter::formats(PixelFormat input)\n \t\treturn {};\n \t}\n \n-\tif (v4l2Format.fourcc != V4L2PixelFormat::fromPixelFormat(input)) {\n+\tif (v4l2Format.fourcc != m2m_->output()->toV4L2PixelFormat(input)) {\n \t\tLOG(SimplePipeline, Debug)\n \t\t\t<< \"Input format \" << input << \" not supported.\";\n \t\treturn {};\n@@ -287,7 +287,7 @@ SimpleConverter::strideAndFrameSize(const PixelFormat &pixelFormat,\n \t\t\t\t const Size &size)\n {\n \tV4L2DeviceFormat format;\n-\tformat.fourcc = V4L2PixelFormat::fromPixelFormat(pixelFormat)[0];\n+\tformat.fourcc = m2m_->capture()->toV4L2PixelFormat(pixelFormat);\n \tformat.size = size;\n \n \tint ret = m2m_->capture()->tryFormat(&format);\ndiff --git a/src/libcamera/pipeline/simple/simple.cpp b/src/libcamera/pipeline/simple/simple.cpp\nindex 33a7b4e57deb..65a483c14379 100644\n--- a/src/libcamera/pipeline/simple/simple.cpp\n+++ b/src/libcamera/pipeline/simple/simple.cpp\n@@ -918,7 +918,7 @@ CameraConfiguration::Status SimpleCameraConfiguration::validate()\n \t\t\t\treturn Invalid;\n \t\t} else {\n \t\t\tV4L2DeviceFormat format;\n-\t\t\tformat.fourcc = V4L2PixelFormat::fromPixelFormat(cfg.pixelFormat)[0];\n+\t\t\tformat.fourcc = data_->video_->toV4L2PixelFormat(cfg.pixelFormat);\n \t\t\tformat.size = cfg.size;\n \n \t\t\tint ret = data_->video_->tryFormat(&format);\ndiff --git a/src/libcamera/pipeline/uvcvideo/uvcvideo.cpp b/src/libcamera/pipeline/uvcvideo/uvcvideo.cpp\nindex 338aff02aff0..cf870d26e3c3 100644\n--- a/src/libcamera/pipeline/uvcvideo/uvcvideo.cpp\n+++ b/src/libcamera/pipeline/uvcvideo/uvcvideo.cpp\n@@ -149,7 +149,7 @@ CameraConfiguration::Status UVCCameraConfiguration::validate()\n \tcfg.bufferCount = 4;\n \n \tV4L2DeviceFormat format;\n-\tformat.fourcc = V4L2PixelFormat::fromPixelFormat(cfg.pixelFormat)[0];\n+\tformat.fourcc = data_->video_->toV4L2PixelFormat(cfg.pixelFormat);\n \tformat.size = cfg.size;\n \n \tint ret = data_->video_->tryFormat(&format);\n@@ -205,7 +205,7 @@ int PipelineHandlerUVC::configure(Camera *camera, CameraConfiguration *config)\n \tint ret;\n \n \tV4L2DeviceFormat format;\n-\tformat.fourcc = V4L2PixelFormat::fromPixelFormat(cfg.pixelFormat)[0];\n+\tformat.fourcc = data->video_->toV4L2PixelFormat(cfg.pixelFormat);\n \tformat.size = cfg.size;\n \n \tret = data->video_->setFormat(&format);\ndiff --git a/src/libcamera/pipeline/vimc/vimc.cpp b/src/libcamera/pipeline/vimc/vimc.cpp\nindex b88b30fff108..153cf849bb18 100644\n--- a/src/libcamera/pipeline/vimc/vimc.cpp\n+++ b/src/libcamera/pipeline/vimc/vimc.cpp\n@@ -171,7 +171,7 @@ CameraConfiguration::Status VimcCameraConfiguration::validate()\n \tcfg.bufferCount = 4;\n \n \tV4L2DeviceFormat format;\n-\tformat.fourcc = V4L2PixelFormat::fromPixelFormat(cfg.pixelFormat)[0];\n+\tformat.fourcc = data_->video_->toV4L2PixelFormat(cfg.pixelFormat);\n \tformat.size = cfg.size;\n \n \tint ret = data_->video_->tryFormat(&format);\n@@ -275,7 +275,7 @@ int PipelineHandlerVimc::configure(Camera *camera, CameraConfiguration *config)\n \t\treturn ret;\n \n \tV4L2DeviceFormat format;\n-\tformat.fourcc = V4L2PixelFormat::fromPixelFormat(cfg.pixelFormat)[0];\n+\tformat.fourcc = data->video_->toV4L2PixelFormat(cfg.pixelFormat);\n \tformat.size = cfg.size;\n \n \tret = data->video_->setFormat(&format);\n@@ -283,7 +283,7 @@ int PipelineHandlerVimc::configure(Camera *camera, CameraConfiguration *config)\n \t\treturn ret;\n \n \tif (format.size != cfg.size ||\n-\t format.fourcc != V4L2PixelFormat::fromPixelFormat(cfg.pixelFormat)[0])\n+\t format.fourcc != data->video_->toV4L2PixelFormat(cfg.pixelFormat))\n \t\treturn -EINVAL;\n \n \t/*\n@@ -598,7 +598,7 @@ int VimcCameraData::allocateMockIPABuffers()\n \tconstexpr unsigned int kBufCount = 2;\n \n \tV4L2DeviceFormat format;\n-\tformat.fourcc = V4L2PixelFormat::fromPixelFormat(formats::BGR888)[0];\n+\tformat.fourcc = video_->toV4L2PixelFormat(formats::BGR888);\n \tformat.size = Size (160, 120);\n \n \tint ret = video_->setFormat(&format);\ndiff --git a/src/libcamera/v4l2_videodevice.cpp b/src/libcamera/v4l2_videodevice.cpp\nindex 63911339f96e..43c3d0f69266 100644\n--- a/src/libcamera/v4l2_videodevice.cpp\n+++ b/src/libcamera/v4l2_videodevice.cpp\n@@ -1989,6 +1989,23 @@ V4L2VideoDevice::fromEntityName(const MediaDevice *media,\n \treturn std::make_unique<V4L2VideoDevice>(mediaEntity);\n }\n \n+/**\n+ * \\brief Convert \\a PixelFormat to its corresponding V4L2 FourCC\n+ * \\param[in] pixelFormat The PixelFormat to convert\n+ *\n+ * For multiplanar formats, the V4L2 format variant (contiguous or\n+ * non-contiguous planes) is selected automatically based on the capabilities\n+ * of the video device. If the video device supports the V4L2 multiplanar API,\n+ * non-contiguous formats are preferred.\n+ *\n+ * \\return The V4L2_PIX_FMT_* pixel format code corresponding to \\a pixelFormat\n+ */\n+V4L2PixelFormat V4L2VideoDevice::toV4L2PixelFormat(const PixelFormat &pixelFormat) const\n+{\n+\treturn V4L2PixelFormat::fromPixelFormat(pixelFormat,\n+\t\t\t\t\t\tcaps_.isMultiplanar())[0];\n+}\n+\n /**\n * \\class V4L2M2MDevice\n * \\brief Memory-to-Memory video device\ndiff --git a/test/libtest/buffer_source.cpp b/test/libtest/buffer_source.cpp\nindex 84f1169c9f5f..dde11f365e43 100644\n--- a/test/libtest/buffer_source.cpp\n+++ b/test/libtest/buffer_source.cpp\n@@ -72,7 +72,7 @@ int BufferSource::allocate(const StreamConfiguration &config)\n \t}\n \n \tformat.size = config.size;\n-\tformat.fourcc = V4L2PixelFormat::fromPixelFormat(config.pixelFormat)[0];\n+\tformat.fourcc = video->toV4L2PixelFormat(config.pixelFormat);\n \tif (video->setFormat(&format)) {\n \t\tstd::cout << \"Failed to set format on output device\" << std::endl;\n \t\treturn TestFail;\n", "prefixes": [ "libcamera-devel", "v2", "4/8" ] }