[{"id":3251,"web_url":"https://patchwork.libcamera.org/comment/3251/","msgid":"<20191215020147.GH16015@pendragon.ideasonboard.com>","date":"2019-12-15T02:01:47","subject":"Re: [libcamera-devel] [PATCH 21/30] libcamera: pipelines:\n\tExplicitly allocate streams","submitter":{"id":2,"url":"https://patchwork.libcamera.org/api/people/2/","name":"Laurent Pinchart","email":"laurent.pinchart@ideasonboard.com"},"content":"Hi Niklas,\n\nThank you for the patch.\n\nOn Wed, Nov 27, 2019 at 12:36:11AM +0100, Niklas Söderlund wrote:\n> Prepare for sub-classing the Stream class with a V4L2 specific\n> implementation which will need to be constructed later when knowledge\n> about the V4L2 video device is available.\n> \n> Signed-off-by: Niklas Söderlund <niklas.soderlund@ragnatech.se>\n> Reviewed-by: Jacopo Mondi <jacopo@jmondi.org>\n\nAs stated in the review of 23/30, I would merge the two patches. The\nsubject line of this patch should then mention switching to V4L2Stream\ninstead of explicitly allocating streams and the body of the commit\nmessage should be updated accordingly.\n\n> ---\n>  src/libcamera/pipeline/ipu3/ipu3.cpp     | 64 +++++++++++++-----------\n>  src/libcamera/pipeline/rkisp1/rkisp1.cpp | 19 ++++---\n>  src/libcamera/pipeline/uvcvideo.cpp      | 13 +++--\n>  src/libcamera/pipeline/vimc.cpp          | 14 ++++--\n>  4 files changed, 62 insertions(+), 48 deletions(-)\n> \n> diff --git a/src/libcamera/pipeline/ipu3/ipu3.cpp b/src/libcamera/pipeline/ipu3/ipu3.cpp\n> index 8ba08351c950f5e2..094c4db6bc32b683 100644\n> --- a/src/libcamera/pipeline/ipu3/ipu3.cpp\n> +++ b/src/libcamera/pipeline/ipu3/ipu3.cpp\n> @@ -138,8 +138,8 @@ public:\n>  class IPU3Stream : public Stream\n>  {\n>  public:\n> -\tIPU3Stream()\n> -\t\t: active_(false), device_(nullptr)\n> +\tIPU3Stream(ImgUDevice::ImgUOutput *device, const std::string &name)\n> +\t\t: active_(false), name_(name), device_(device)\n>  \t{\n>  \t}\n>  \n> @@ -152,10 +152,16 @@ class IPU3CameraData : public CameraData\n>  {\n>  public:\n>  \tIPU3CameraData(PipelineHandler *pipe)\n> -\t\t: CameraData(pipe)\n> +\t\t: CameraData(pipe), outStream_(nullptr), vfStream_(nullptr)\n>  \t{\n>  \t}\n>  \n> +\t~IPU3CameraData()\n> +\t{\n> +\t\tdelete outStream_;\n> +\t\tdelete vfStream_;\n> +\t}\n> +\n>  \tvoid imguOutputBufferReady(Buffer *buffer);\n>  \tvoid imguInputBufferReady(Buffer *buffer);\n>  \tvoid cio2BufferReady(Buffer *buffer);\n> @@ -163,8 +169,8 @@ public:\n>  \tCIO2Device cio2_;\n>  \tImgUDevice *imgu_;\n>  \n> -\tIPU3Stream outStream_;\n> -\tIPU3Stream vfStream_;\n> +\tIPU3Stream *outStream_;\n> +\tIPU3Stream *vfStream_;\n>  \n>  \tstd::vector<std::unique_ptr<Buffer>> rawBuffers_;\n>  };\n> @@ -343,8 +349,8 @@ CameraConfiguration::Status IPU3CameraConfiguration::validate()\n>  \t * stream otherwise.\n>  \t */\n>  \tstd::set<const IPU3Stream *> availableStreams = {\n> -\t\t&data_->outStream_,\n> -\t\t&data_->vfStream_,\n> +\t\tdata_->outStream_,\n> +\t\tdata_->vfStream_,\n>  \t};\n>  \n>  \tstreams_.clear();\n> @@ -357,9 +363,9 @@ CameraConfiguration::Status IPU3CameraConfiguration::validate()\n>  \t\tconst IPU3Stream *stream;\n>  \n>  \t\tif (cfg.size == sensorFormat_.size)\n> -\t\t\tstream = &data_->outStream_;\n> +\t\t\tstream = data_->outStream_;\n>  \t\telse\n> -\t\t\tstream = &data_->vfStream_;\n> +\t\t\tstream = data_->vfStream_;\n>  \n>  \t\tif (availableStreams.find(stream) == availableStreams.end())\n>  \t\t\tstream = *availableStreams.begin();\n> @@ -367,7 +373,7 @@ CameraConfiguration::Status IPU3CameraConfiguration::validate()\n>  \t\tLOG(IPU3, Debug)\n>  \t\t\t<< \"Assigned '\" << stream->name_ << \"' to stream \" << i;\n>  \n> -\t\tbool scale = stream == &data_->vfStream_;\n> +\t\tbool scale = stream == data_->vfStream_;\n>  \t\tadjustStream(config_[i], scale);\n>  \n>  \t\tif (cfg.pixelFormat != pixelFormat || cfg.size != size) {\n> @@ -395,8 +401,8 @@ CameraConfiguration *PipelineHandlerIPU3::generateConfiguration(Camera *camera,\n>  \tIPU3CameraData *data = cameraData(camera);\n>  \tIPU3CameraConfiguration *config;\n>  \tstd::set<IPU3Stream *> streams = {\n> -\t\t&data->outStream_,\n> -\t\t&data->vfStream_,\n> +\t\tdata->outStream_,\n> +\t\tdata->vfStream_,\n>  \t};\n>  \n>  \tconfig = new IPU3CameraConfiguration(camera, data);\n> @@ -414,10 +420,10 @@ CameraConfiguration *PipelineHandlerIPU3::generateConfiguration(Camera *camera,\n>  \t\t\t * and VideoRecording roles are not allowed on\n>  \t\t\t * the output stream.\n>  \t\t\t */\n> -\t\t\tif (streams.find(&data->outStream_) != streams.end()) {\n> -\t\t\t\tstream = &data->outStream_;\n> -\t\t\t} else if (streams.find(&data->vfStream_) != streams.end()) {\n> -\t\t\t\tstream = &data->vfStream_;\n> +\t\t\tif (streams.find(data->outStream_) != streams.end()) {\n> +\t\t\t\tstream = data->outStream_;\n> +\t\t\t} else if (streams.find(data->vfStream_) != streams.end()) {\n> +\t\t\t\tstream = data->vfStream_;\n>  \t\t\t} else {\n>  \t\t\t\tLOG(IPU3, Error)\n>  \t\t\t\t\t<< \"No stream available for requested role \"\n> @@ -447,14 +453,14 @@ CameraConfiguration *PipelineHandlerIPU3::generateConfiguration(Camera *camera,\n>  \t\t\t * \\todo This is an artificial limitation until we\n>  \t\t\t * figure out the exact capabilities of the hardware.\n>  \t\t\t */\n> -\t\t\tif (streams.find(&data->vfStream_) == streams.end()) {\n> +\t\t\tif (streams.find(data->vfStream_) == streams.end()) {\n>  \t\t\t\tLOG(IPU3, Error)\n>  \t\t\t\t\t<< \"No stream available for requested role \"\n>  \t\t\t\t\t<< role;\n>  \t\t\t\tbreak;\n>  \t\t\t}\n>  \n> -\t\t\tstream = &data->vfStream_;\n> +\t\t\tstream = data->vfStream_;\n>  \n>  \t\t\t/*\n>  \t\t\t * Align the default viewfinder size to the maximum\n> @@ -495,8 +501,8 @@ int PipelineHandlerIPU3::configure(Camera *camera, CameraConfiguration *c)\n>  \tIPU3CameraConfiguration *config =\n>  \t\tstatic_cast<IPU3CameraConfiguration *>(c);\n>  \tIPU3CameraData *data = cameraData(camera);\n> -\tIPU3Stream *outStream = &data->outStream_;\n> -\tIPU3Stream *vfStream = &data->vfStream_;\n> +\tIPU3Stream *outStream = data->outStream_;\n> +\tIPU3Stream *vfStream = data->vfStream_;\n>  \tCIO2Device *cio2 = &data->cio2_;\n>  \tImgUDevice *imgu = data->imgu_;\n>  \tint ret;\n> @@ -630,8 +636,8 @@ int PipelineHandlerIPU3::allocateBuffers(Camera *camera,\n>  \t\t\t\t\t const std::set<Stream *> &streams)\n>  {\n>  \tIPU3CameraData *data = cameraData(camera);\n> -\tIPU3Stream *outStream = &data->outStream_;\n> -\tIPU3Stream *vfStream = &data->vfStream_;\n> +\tIPU3Stream *outStream = data->outStream_;\n> +\tIPU3Stream *vfStream = data->vfStream_;\n>  \tCIO2Device *cio2 = &data->cio2_;\n>  \tImgUDevice *imgu = data->imgu_;\n>  \tunsigned int bufferCount;\n> @@ -855,10 +861,6 @@ int PipelineHandlerIPU3::registerCameras()\n>  \tfor (unsigned int id = 0; id < 4 && numCameras < 2; ++id) {\n>  \t\tstd::unique_ptr<IPU3CameraData> data =\n>  \t\t\tutils::make_unique<IPU3CameraData>(this);\n> -\t\tstd::set<Stream *> streams = {\n> -\t\t\t&data->outStream_,\n> -\t\t\t&data->vfStream_,\n> -\t\t};\n>  \t\tCIO2Device *cio2 = &data->cio2_;\n>  \n>  \t\tret = cio2->init(cio2MediaDev_, id);\n> @@ -872,10 +874,8 @@ int PipelineHandlerIPU3::registerCameras()\n>  \t\t * second.\n>  \t\t */\n>  \t\tdata->imgu_ = numCameras ? &imgu1_ : &imgu0_;\n> -\t\tdata->outStream_.device_ = &data->imgu_->output_;\n> -\t\tdata->outStream_.name_ = \"output\";\n> -\t\tdata->vfStream_.device_ = &data->imgu_->viewfinder_;\n> -\t\tdata->vfStream_.name_ = \"viewfinder\";\n> +\t\tdata->outStream_ = new IPU3Stream(&data->imgu_->output_, \"output\");\n> +\t\tdata->vfStream_ = new IPU3Stream(&data->imgu_->viewfinder_, \"viewfinder\");\n>  \n>  \t\t/*\n>  \t\t * Connect video devices' 'bufferReady' signals to their\n> @@ -895,6 +895,10 @@ int PipelineHandlerIPU3::registerCameras()\n>  \t\t\t\t\t&IPU3CameraData::imguOutputBufferReady);\n>  \n>  \t\t/* Create and register the Camera instance. */\n> +\t\tstd::set<Stream *> streams = {\n> +\t\t\tdata->outStream_,\n> +\t\t\tdata->vfStream_,\n> +\t\t};\n>  \t\tstd::string cameraName = cio2->sensor_->entity()->name() + \" \"\n>  \t\t\t\t       + std::to_string(id);\n>  \t\tstd::shared_ptr<Camera> camera = Camera::create(this,\n> diff --git a/src/libcamera/pipeline/rkisp1/rkisp1.cpp b/src/libcamera/pipeline/rkisp1/rkisp1.cpp\n> index 6ad9b57d8353896c..1ae50bec0d401b3f 100644\n> --- a/src/libcamera/pipeline/rkisp1/rkisp1.cpp\n> +++ b/src/libcamera/pipeline/rkisp1/rkisp1.cpp\n> @@ -117,19 +117,20 @@ class RkISP1CameraData : public CameraData\n>  {\n>  public:\n>  \tRkISP1CameraData(PipelineHandler *pipe)\n> -\t\t: CameraData(pipe), sensor_(nullptr), frame_(0),\n> -\t\t  frameInfo_(pipe)\n> +\t\t: CameraData(pipe), stream_(nullptr), sensor_(nullptr),\n> +\t\t  frame_(0), frameInfo_(pipe)\n>  \t{\n>  \t}\n>  \n>  \t~RkISP1CameraData()\n>  \t{\n> +\t\tdelete stream_;\n>  \t\tdelete sensor_;\n>  \t}\n>  \n>  \tint loadIPA();\n>  \n> -\tStream stream_;\n> +\tStream *stream_;\n>  \tCameraSensor *sensor_;\n>  \tunsigned int frame_;\n>  \tstd::vector<IPABuffer> ipaBuffers_;\n> @@ -651,7 +652,7 @@ int PipelineHandlerRkISP1::configure(Camera *camera, CameraConfiguration *c)\n>  \tif (ret)\n>  \t\treturn ret;\n>  \n> -\tcfg.setStream(&data->stream_);\n> +\tcfg.setStream(data->stream_);\n>  \n>  \treturn 0;\n>  }\n> @@ -785,8 +786,8 @@ int PipelineHandlerRkISP1::start(Camera *camera)\n>  \t/* Inform IPA of stream configuration and sensor controls. */\n>  \tstd::map<unsigned int, IPAStream> streamConfig;\n>  \tstreamConfig[0] = {\n> -\t\t.pixelFormat = data->stream_.configuration().pixelFormat,\n> -\t\t.size = data->stream_.configuration().size,\n> +\t\t.pixelFormat = data->stream_->configuration().pixelFormat,\n> +\t\t.size = data->stream_->configuration().size,\n>  \t};\n>  \n>  \tstd::map<unsigned int, const ControlInfoMap &> entityControls;\n> @@ -826,7 +827,7 @@ int PipelineHandlerRkISP1::queueRequestHardware(Camera *camera,\n>  \t\t\t\t\t\tRequest *request)\n>  {\n>  \tRkISP1CameraData *data = cameraData(camera);\n> -\tStream *stream = &data->stream_;\n> +\tStream *stream = data->stream_;\n>  \n>  \tRkISP1FrameInfo *info = data->frameInfo_.create(data->frame_, request,\n>  \t\t\t\t\t\t\tstream);\n> @@ -887,6 +888,8 @@ int PipelineHandlerRkISP1::createCamera(MediaEntity *sensor)\n>  \tstd::unique_ptr<RkISP1CameraData> data =\n>  \t\tutils::make_unique<RkISP1CameraData>(this);\n>  \n> +\tdata->stream_ = new Stream();\n> +\n>  \tControlInfoMap::Map ctrls;\n>  \tctrls.emplace(std::piecewise_construct,\n>  \t\t      std::forward_as_tuple(&controls::AeEnable),\n> @@ -903,7 +906,7 @@ int PipelineHandlerRkISP1::createCamera(MediaEntity *sensor)\n>  \tif (ret)\n>  \t\treturn ret;\n>  \n> -\tstd::set<Stream *> streams{ &data->stream_ };\n> +\tstd::set<Stream *> streams{ data->stream_ };\n>  \tstd::shared_ptr<Camera> camera =\n>  \t\tCamera::create(this, sensor->name(), streams);\n>  \tregisterCamera(std::move(camera), std::move(data));\n> diff --git a/src/libcamera/pipeline/uvcvideo.cpp b/src/libcamera/pipeline/uvcvideo.cpp\n> index 3a76653ff6dc2b5e..64fc488912e5a82f 100644\n> --- a/src/libcamera/pipeline/uvcvideo.cpp\n> +++ b/src/libcamera/pipeline/uvcvideo.cpp\n> @@ -31,12 +31,13 @@ class UVCCameraData : public CameraData\n>  {\n>  public:\n>  \tUVCCameraData(PipelineHandler *pipe)\n> -\t\t: CameraData(pipe), video_(nullptr)\n> +\t\t: CameraData(pipe), video_(nullptr), stream_(nullptr)\n>  \t{\n>  \t}\n>  \n>  \t~UVCCameraData()\n>  \t{\n> +\t\tdelete stream_;\n>  \t\tdelete video_;\n>  \t}\n>  \n> @@ -44,7 +45,7 @@ public:\n>  \tvoid bufferReady(Buffer *buffer);\n>  \n>  \tV4L2VideoDevice *video_;\n> -\tStream stream_;\n> +\tStream *stream_;\n>  };\n>  \n>  class UVCCameraConfiguration : public CameraConfiguration\n> @@ -187,7 +188,7 @@ int PipelineHandlerUVC::configure(Camera *camera, CameraConfiguration *config)\n>  \t    format.fourcc != data->video_->toV4L2Fourcc(cfg.pixelFormat))\n>  \t\treturn -EINVAL;\n>  \n> -\tcfg.setStream(&data->stream_);\n> +\tcfg.setStream(data->stream_);\n>  \n>  \treturn 0;\n>  }\n> @@ -265,7 +266,7 @@ int PipelineHandlerUVC::processControls(UVCCameraData *data, Request *request)\n>  int PipelineHandlerUVC::queueRequestHardware(Camera *camera, Request *request)\n>  {\n>  \tUVCCameraData *data = cameraData(camera);\n> -\tBuffer *buffer = request->findBuffer(&data->stream_);\n> +\tBuffer *buffer = request->findBuffer(data->stream_);\n>  \tif (!buffer) {\n>  \t\tLOG(UVC, Error)\n>  \t\t\t<< \"Attempt to queue request with invalid stream\";\n> @@ -310,7 +311,7 @@ bool PipelineHandlerUVC::match(DeviceEnumerator *enumerator)\n>  \t}\n>  \n>  \t/* Create and register the camera. */\n> -\tstd::set<Stream *> streams{ &data->stream_ };\n> +\tstd::set<Stream *> streams{ data->stream_ };\n>  \tstd::shared_ptr<Camera> camera = Camera::create(this, media->model(), streams);\n>  \tregisterCamera(std::move(camera), std::move(data));\n>  \n> @@ -330,6 +331,8 @@ int UVCCameraData::init(MediaEntity *entity)\n>  \tif (ret)\n>  \t\treturn ret;\n>  \n> +\tstream_ = new Stream();\n> +\n>  \tvideo_->bufferReady.connect(this, &UVCCameraData::bufferReady);\n>  \n>  \t/* Initialise the supported controls. */\n> diff --git a/src/libcamera/pipeline/vimc.cpp b/src/libcamera/pipeline/vimc.cpp\n> index f5550a1723668106..3f9e92163642f0c2 100644\n> --- a/src/libcamera/pipeline/vimc.cpp\n> +++ b/src/libcamera/pipeline/vimc.cpp\n> @@ -41,7 +41,8 @@ class VimcCameraData : public CameraData\n>  public:\n>  \tVimcCameraData(PipelineHandler *pipe)\n>  \t\t: CameraData(pipe), sensor_(nullptr), debayer_(nullptr),\n> -\t\t  scaler_(nullptr), video_(nullptr), raw_(nullptr)\n> +\t\t  scaler_(nullptr), video_(nullptr), raw_(nullptr),\n> +\t\t  stream_(nullptr)\n>  \t{\n>  \t}\n>  \n> @@ -52,6 +53,7 @@ public:\n>  \t\tdelete scaler_;\n>  \t\tdelete video_;\n>  \t\tdelete raw_;\n> +\t\tdelete stream_;\n\nI would delete the stream before the video device that it is constructed\nwith, otherwise you'll have an invalid pointer that may become an issue\nlater.\n\nReviewed-by: Laurent Pinchart <laurent.pinchart@ideasonboard.com>\n\n>  \t}\n>  \n>  \tint init(MediaDevice *media);\n> @@ -62,7 +64,7 @@ public:\n>  \tV4L2Subdevice *scaler_;\n>  \tV4L2VideoDevice *video_;\n>  \tV4L2VideoDevice *raw_;\n> -\tStream stream_;\n> +\tStream *stream_;\n>  };\n>  \n>  class VimcCameraConfiguration : public CameraConfiguration\n> @@ -254,7 +256,7 @@ int PipelineHandlerVimc::configure(Camera *camera, CameraConfiguration *config)\n>  \tif (ret)\n>  \t\treturn ret;\n>  \n> -\tcfg.setStream(&data->stream_);\n> +\tcfg.setStream(data->stream_);\n>  \n>  \treturn 0;\n>  }\n> @@ -326,7 +328,7 @@ int PipelineHandlerVimc::processControls(VimcCameraData *data, Request *request)\n>  int PipelineHandlerVimc::queueRequestHardware(Camera *camera, Request *request)\n>  {\n>  \tVimcCameraData *data = cameraData(camera);\n> -\tBuffer *buffer = request->findBuffer(&data->stream_);\n> +\tBuffer *buffer = request->findBuffer(data->stream_);\n>  \tif (!buffer) {\n>  \t\tLOG(VIMC, Error)\n>  \t\t\t<< \"Attempt to queue request with invalid stream\";\n> @@ -376,7 +378,7 @@ bool PipelineHandlerVimc::match(DeviceEnumerator *enumerator)\n>  \t\treturn false;\n>  \n>  \t/* Create and register the camera. */\n> -\tstd::set<Stream *> streams{ &data->stream_ };\n> +\tstd::set<Stream *> streams{ data->stream_ };\n>  \tstd::shared_ptr<Camera> camera = Camera::create(this, \"VIMC Sensor B\",\n>  \t\t\t\t\t\t\tstreams);\n>  \tregisterCamera(std::move(camera), std::move(data));\n> @@ -418,6 +420,8 @@ int VimcCameraData::init(MediaDevice *media)\n>  \tif (video_->open())\n>  \t\treturn -ENODEV;\n>  \n> +\tstream_ = new Stream();\n> +\n>  \tvideo_->bufferReady.connect(this, &VimcCameraData::bufferReady);\n>  \n>  \traw_ = new V4L2VideoDevice(media->getEntityByName(\"Raw Capture 1\"));","headers":{"Return-Path":"<laurent.pinchart@ideasonboard.com>","Received":["from perceval.ideasonboard.com (perceval.ideasonboard.com\n\t[213.167.242.64])\n\tby lancelot.ideasonboard.com (Postfix) with ESMTPS id 78A34601E5\n\tfor <libcamera-devel@lists.libcamera.org>;\n\tSun, 15 Dec 2019 03:01:57 +0100 (CET)","from pendragon.ideasonboard.com (81-175-216-236.bb.dnainternet.fi\n\t[81.175.216.236])\n\tby perceval.ideasonboard.com (Postfix) with ESMTPSA id D4F222D1;\n\tSun, 15 Dec 2019 03:01:56 +0100 (CET)"],"DKIM-Signature":"v=1; a=rsa-sha256; c=relaxed/simple; d=ideasonboard.com;\n\ts=mail; t=1576375317;\n\tbh=S3nioNh5UOTXRfbd9LavPsOWNlOcXz57WTcVE8+8i5M=;\n\th=Date:From:To:Cc:Subject:References:In-Reply-To:From;\n\tb=UMrjBaliVukkTXqq6ZMmXCf4Bq4vNr5mdbIYXg8zKPQR039YplHAgiCnvkWv42ZiQ\n\tzCKl6vUSmphIFJMg2Pbc9dJAAMXtT/DBnupaABXV8x49oh9HdhHqVBYiG9ASuhrzjz\n\tE/5o2ym7oamvb8e02cY9CK0NFriZjYT/e/ifKIAs=","Date":"Sun, 15 Dec 2019 04:01:47 +0200","From":"Laurent Pinchart <laurent.pinchart@ideasonboard.com>","To":"Niklas =?utf-8?q?S=C3=B6derlund?= <niklas.soderlund@ragnatech.se>","Cc":"libcamera-devel@lists.libcamera.org","Message-ID":"<20191215020147.GH16015@pendragon.ideasonboard.com>","References":"<20191126233620.1695316-1-niklas.soderlund@ragnatech.se>\n\t<20191126233620.1695316-22-niklas.soderlund@ragnatech.se>","MIME-Version":"1.0","Content-Type":"text/plain; charset=utf-8","Content-Disposition":"inline","Content-Transfer-Encoding":"8bit","In-Reply-To":"<20191126233620.1695316-22-niklas.soderlund@ragnatech.se>","User-Agent":"Mutt/1.10.1 (2018-07-13)","Subject":"Re: [libcamera-devel] [PATCH 21/30] libcamera: pipelines:\n\tExplicitly allocate streams","X-BeenThere":"libcamera-devel@lists.libcamera.org","X-Mailman-Version":"2.1.29","Precedence":"list","List-Id":"<libcamera-devel.lists.libcamera.org>","List-Unsubscribe":"<https://lists.libcamera.org/options/libcamera-devel>,\n\t<mailto:libcamera-devel-request@lists.libcamera.org?subject=unsubscribe>","List-Archive":"<https://lists.libcamera.org/pipermail/libcamera-devel/>","List-Post":"<mailto:libcamera-devel@lists.libcamera.org>","List-Help":"<mailto:libcamera-devel-request@lists.libcamera.org?subject=help>","List-Subscribe":"<https://lists.libcamera.org/listinfo/libcamera-devel>,\n\t<mailto:libcamera-devel-request@lists.libcamera.org?subject=subscribe>","X-List-Received-Date":"Sun, 15 Dec 2019 02:01:57 -0000"}}]