From patchwork Mon Jul 25 13:46:35 2022 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Naushir Patuck X-Patchwork-Id: 16766 Return-Path: X-Original-To: parsemail@patchwork.libcamera.org Delivered-To: parsemail@patchwork.libcamera.org Received: from lancelot.ideasonboard.com (lancelot.ideasonboard.com [92.243.16.209]) by patchwork.libcamera.org (Postfix) with ESMTPS id C6531C3275 for ; Mon, 25 Jul 2022 13:47:01 +0000 (UTC) Received: from lancelot.ideasonboard.com (localhost [IPv6:::1]) by lancelot.ideasonboard.com (Postfix) with ESMTP id 8A64B63336; Mon, 25 Jul 2022 15:47:01 +0200 (CEST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=libcamera.org; s=mail; t=1658756821; bh=VH+cMqzAjFMUF6gP4pL8/5xjnYGkbOaAN34B+0gU5vw=; h=To:Date:In-Reply-To:References:Subject:List-Id:List-Unsubscribe: List-Archive:List-Post:List-Help:List-Subscribe:From:Reply-To: From; b=HhGca/pSePOoZ/+1cSWXb943uBytNZml2bgK/qhrnIvMhu7Bp/7341LDKKvrG5Tpe vDTLLw+3WvhEi6gV078kG7pL5mCnD8YlkD2BF3KQvaaDwZzC4SsWPxPebAHUUp8w88 SNHONhxRtThKUI0e42o+zn9xhjjk4xQMewTNfeEnChzsr3o2mJ2iHTuRvcTHq1oYfG RHuv3x1+wMy2BI6dxg7O3JlMgojVxaZuSmWpRI/JH8a9wwSuNC8WkXbSGUMhG7gTLF 49trVrZTgz8oVKCdDA3wBQmxNzJiSrCoQ7YdhvhsnqwEBgMRxjVja/avPK3iyHnR7m bPQe++6ukCGTg== Received: from mail-wm1-x330.google.com (mail-wm1-x330.google.com [IPv6:2a00:1450:4864:20::330]) by lancelot.ideasonboard.com (Postfix) with ESMTPS id 4B9316332C for ; Mon, 25 Jul 2022 15:46:55 +0200 (CEST) Authentication-Results: lancelot.ideasonboard.com; dkim=pass (2048-bit key; unprotected) header.d=raspberrypi.com header.i=@raspberrypi.com header.b="Z+osmpG6"; dkim-atps=neutral Received: by mail-wm1-x330.google.com with SMTP id j29-20020a05600c1c1d00b003a2fdafdefbso6396007wms.2 for ; Mon, 25 Jul 2022 06:46:55 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=raspberrypi.com; s=google; h=from:to:cc:subject:date:message-id:in-reply-to:references :mime-version:content-transfer-encoding; bh=6FjY6WN+r/Z7EK23A4EC/a+TpOYs99V1qqD59adpGqg=; b=Z+osmpG6OLn2v8gpiFT+WJXDjuIQqBMNnIdmPg8YKAzEeo76QZ6KUlH9M3czg1AELn qWE0QapqEJfW1HGaLoy1yPBZU1spc5em2MXdgC3OmBSGO/mBXRzZBV9LpXVUb4iDYIfI vy7PuoeqfdATOU6xjEUWE1qvKNKw5fFsITbFS58vhGFkP2G90IEw3sM1nou7ofzIRMca 1Owb9ROprMtpYMQGl0Ojbz0+XCCH4B4QCXOM3MefmoN8SLzFqL7EA298Oev3u2AHu4DL oLhBB77xcETafdzW8JmH/MHLVcxq0aNwE88SQPxq3WLzBcNzonCeb4OL+/GK87ItfCIl LH/A== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=x-gm-message-state:from:to:cc:subject:date:message-id:in-reply-to :references:mime-version:content-transfer-encoding; bh=6FjY6WN+r/Z7EK23A4EC/a+TpOYs99V1qqD59adpGqg=; b=Qn5M35IRD2sQuUNHIpXS0wXxAQ3IgEpYtzcsRJpLO3L9m7NDxvYfCU6b1ycL6wEb2O 6LbjhayiaLYOVdjHuMEHx886weAEKJ2bm75YhAJQkqc9QtGK+QyFfjoCW66wA13PGobD 12TKkfeTobb2999w7/SrIjD2RUVWh99fkAxSxjL47Io3tIF5lpFGGUFr7E4ZIOm+/DMw A2otpcpvBjZeS9LjKkPE/7XBIu1pnJDMTumnzmYwTFu7xyp6nbISlthKrIIlbhTDsNim 13k9hjFzE4mXSh+27k9LtJ+AvidIHaa8FYuzlEtZjQqZfIiKvwJUvNTGQS7PTIAuIJ0s sFjw== X-Gm-Message-State: AJIora+CGvIe7U+kVdfA712HiA0E5bxj7TLTfsJT5M5tT00fHSkyqecR grDHbD+dvYI+VlFrrcv7lpWL6OwbDAp7lA== X-Google-Smtp-Source: AGRyM1vIuSg0EpXwn0ADGbFZGemzHdGA5os3xLyae3EkMlzPZNsyf4E2a+5wjMMoZDdm1ST33ZjLkg== X-Received: by 2002:a05:600c:2c46:b0:3a3:3248:32a6 with SMTP id r6-20020a05600c2c4600b003a3324832a6mr14437696wmg.179.1658756812113; Mon, 25 Jul 2022 06:46:52 -0700 (PDT) Received: from naush-laptop.localdomain ([93.93.133.154]) by smtp.gmail.com with ESMTPSA id a20-20020a05600c225400b003a32167b8d4sm18054320wmm.13.2022.07.25.06.46.51 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Mon, 25 Jul 2022 06:46:51 -0700 (PDT) To: libcamera-devel@lists.libcamera.org Date: Mon, 25 Jul 2022 14:46:35 +0100 Message-Id: <20220725134639.4572-12-naush@raspberrypi.com> X-Mailer: git-send-email 2.25.1 In-Reply-To: <20220725134639.4572-1-naush@raspberrypi.com> References: <20220725134639.4572-1-naush@raspberrypi.com> MIME-Version: 1.0 Subject: [libcamera-devel] [PATCH 11/15] ipa: raspberrypi: Change to C style code comments X-BeenThere: libcamera-devel@lists.libcamera.org X-Mailman-Version: 2.1.29 Precedence: list List-Id: List-Unsubscribe: , List-Archive: List-Post: List-Help: List-Subscribe: , X-Patchwork-Original-From: Naushir Patuck via libcamera-devel From: Naushir Patuck Reply-To: Naushir Patuck Errors-To: libcamera-devel-bounces@lists.libcamera.org Sender: "libcamera-devel" As part of the on-going refactor efforts for the source files in src/ipa/raspberrypi/, switch all C++ style comments to C style comments. Signed-off-by: Naushir Patuck Reviewed-by: Laurent Pinchart --- src/ipa/raspberrypi/cam_helper.hpp | 98 ++++--- .../raspberrypi/controller/agc_algorithm.hpp | 4 +- src/ipa/raspberrypi/controller/agc_status.h | 18 +- src/ipa/raspberrypi/controller/algorithm.cpp | 2 +- src/ipa/raspberrypi/controller/algorithm.hpp | 16 +- src/ipa/raspberrypi/controller/alsc_status.h | 6 +- .../raspberrypi/controller/awb_algorithm.hpp | 4 +- src/ipa/raspberrypi/controller/awb_status.h | 6 +- .../controller/black_level_status.h | 4 +- src/ipa/raspberrypi/controller/camera_mode.h | 30 +- .../raspberrypi/controller/ccm_algorithm.hpp | 4 +- src/ipa/raspberrypi/controller/ccm_status.h | 2 +- .../controller/contrast_algorithm.hpp | 4 +- .../raspberrypi/controller/contrast_status.h | 6 +- src/ipa/raspberrypi/controller/controller.cpp | 6 +- src/ipa/raspberrypi/controller/controller.hpp | 20 +- .../controller/denoise_algorithm.hpp | 4 +- .../raspberrypi/controller/denoise_status.h | 2 +- src/ipa/raspberrypi/controller/dpc_status.h | 4 +- src/ipa/raspberrypi/controller/focus_status.h | 8 +- src/ipa/raspberrypi/controller/geq_status.h | 2 +- src/ipa/raspberrypi/controller/histogram.cpp | 8 +- src/ipa/raspberrypi/controller/histogram.hpp | 18 +- src/ipa/raspberrypi/controller/lux_status.h | 18 +- src/ipa/raspberrypi/controller/metadata.hpp | 20 +- src/ipa/raspberrypi/controller/noise_status.h | 2 +- src/ipa/raspberrypi/controller/pwl.cpp | 40 ++- src/ipa/raspberrypi/controller/pwl.hpp | 60 ++-- src/ipa/raspberrypi/controller/rpi/agc.cpp | 269 +++++++++++------- src/ipa/raspberrypi/controller/rpi/agc.hpp | 24 +- src/ipa/raspberrypi/controller/rpi/alsc.cpp | 180 +++++++----- src/ipa/raspberrypi/controller/rpi/alsc.hpp | 50 ++-- src/ipa/raspberrypi/controller/rpi/awb.cpp | 192 ++++++++----- src/ipa/raspberrypi/controller/rpi/awb.hpp | 112 ++++---- .../controller/rpi/black_level.cpp | 10 +- .../controller/rpi/black_level.hpp | 4 +- src/ipa/raspberrypi/controller/rpi/ccm.cpp | 20 +- src/ipa/raspberrypi/controller/rpi/ccm.hpp | 4 +- .../raspberrypi/controller/rpi/contrast.cpp | 74 +++-- .../raspberrypi/controller/rpi/contrast.hpp | 8 +- src/ipa/raspberrypi/controller/rpi/dpc.cpp | 10 +- src/ipa/raspberrypi/controller/rpi/dpc.hpp | 4 +- src/ipa/raspberrypi/controller/rpi/geq.cpp | 10 +- src/ipa/raspberrypi/controller/rpi/geq.hpp | 6 +- src/ipa/raspberrypi/controller/rpi/lux.cpp | 16 +- src/ipa/raspberrypi/controller/rpi/lux.hpp | 14 +- src/ipa/raspberrypi/controller/rpi/noise.cpp | 24 +- src/ipa/raspberrypi/controller/rpi/noise.hpp | 6 +- src/ipa/raspberrypi/controller/rpi/sdn.cpp | 12 +- src/ipa/raspberrypi/controller/rpi/sdn.hpp | 4 +- .../raspberrypi/controller/rpi/sharpen.cpp | 32 ++- .../raspberrypi/controller/rpi/sharpen.hpp | 4 +- .../controller/sharpen_algorithm.hpp | 4 +- .../raspberrypi/controller/sharpen_status.h | 10 +- src/ipa/raspberrypi/md_parser.hpp | 2 +- 55 files changed, 890 insertions(+), 631 deletions(-) diff --git a/src/ipa/raspberrypi/cam_helper.hpp b/src/ipa/raspberrypi/cam_helper.hpp index 0cd718c4bc4e..2408fa154d3d 100644 --- a/src/ipa/raspberrypi/cam_helper.hpp +++ b/src/ipa/raspberrypi/cam_helper.hpp @@ -21,50 +21,52 @@ namespace RPiController { -// The CamHelper class provides a number of facilities that anyone trying -// to drive a camera will need to know, but which are not provided by the -// standard driver framework. Specifically, it provides: -// -// A "CameraMode" structure to describe extra information about the chosen -// mode of the driver. For example, how it is cropped from the full sensor -// area, how it is scaled, whether pixels are averaged compared to the full -// resolution. -// -// The ability to convert between number of lines of exposure and actual -// exposure time, and to convert between the sensor's gain codes and actual -// gains. -// -// A function to return the number of frames of delay between updating exposure, -// analogue gain and vblanking, and for the changes to take effect. For many -// sensors these take the values 2, 1 and 2 respectively, but sensors that are -// different will need to over-ride the default function provided. -// -// A function to query if the sensor outputs embedded data that can be parsed. -// -// A function to return the sensitivity of a given camera mode. -// -// A parser to parse the embedded data buffers provided by some sensors (for -// example, the imx219 does; the ov5647 doesn't). This allows us to know for -// sure the exposure and gain of the frame we're looking at. CamHelper -// provides functions for converting analogue gains to and from the sensor's -// native gain codes. -// -// Finally, a set of functions that determine how to handle the vagaries of -// different camera modules on start-up or when switching modes. Some -// modules may produce one or more frames that are not yet correctly exposed, -// or where the metadata may be suspect. We have the following functions: -// HideFramesStartup(): Tell the pipeline handler not to return this many -// frames at start-up. This can also be used to hide initial frames -// while the AGC and other algorithms are sorting themselves out. -// HideFramesModeSwitch(): Tell the pipeline handler not to return this -// many frames after a mode switch (other than start-up). Some sensors -// may produce innvalid frames after a mode switch; others may not. -// MistrustFramesStartup(): At start-up a sensor may return frames for -// which we should not run any control algorithms (for example, metadata -// may be invalid). -// MistrustFramesModeSwitch(): The number of frames, after a mode switch -// (other than start-up), for which control algorithms should not run -// (for example, metadata may be unreliable). +/* + * The CamHelper class provides a number of facilities that anyone trying + * to drive a camera will need to know, but which are not provided by the + * standard driver framework. Specifically, it provides: + * + * A "CameraMode" structure to describe extra information about the chosen + * mode of the driver. For example, how it is cropped from the full sensor + * area, how it is scaled, whether pixels are averaged compared to the full + * resolution. + * + * The ability to convert between number of lines of exposure and actual + * exposure time, and to convert between the sensor's gain codes and actual + * gains. + * + * A function to return the number of frames of delay between updating exposure, + * analogue gain and vblanking, and for the changes to take effect. For many + * sensors these take the values 2, 1 and 2 respectively, but sensors that are + * different will need to over-ride the default function provided. + * + * A function to query if the sensor outputs embedded data that can be parsed. + * + * A function to return the sensitivity of a given camera mode. + * + * A parser to parse the embedded data buffers provided by some sensors (for + * example, the imx219 does; the ov5647 doesn't). This allows us to know for + * sure the exposure and gain of the frame we're looking at. CamHelper + * provides functions for converting analogue gains to and from the sensor's + * native gain codes. + * + * Finally, a set of functions that determine how to handle the vagaries of + * different camera modules on start-up or when switching modes. Some + * modules may produce one or more frames that are not yet correctly exposed, + * or where the metadata may be suspect. We have the following functions: + * HideFramesStartup(): Tell the pipeline handler not to return this many + * frames at start-up. This can also be used to hide initial frames + * while the AGC and other algorithms are sorting themselves out. + * HideFramesModeSwitch(): Tell the pipeline handler not to return this + * many frames after a mode switch (other than start-up). Some sensors + * may produce innvalid frames after a mode switch; others may not. + * MistrustFramesStartup(): At start-up a sensor may return frames for + * which we should not run any control algorithms (for example, metadata + * may be invalid). + * MistrustFramesModeSwitch(): The number of frames, after a mode switch + * (other than start-up), for which control algorithms should not run + * (for example, metadata may be unreliable). + */ class CamHelper { @@ -110,8 +112,10 @@ private: unsigned int frameIntegrationDiff_; }; -// This is for registering camera helpers with the system, so that the -// CamHelper::Create function picks them up automatically. +/* + * This is for registering camera helpers with the system, so that the + * CamHelper::Create function picks them up automatically. + */ typedef CamHelper *(*CamHelperCreateFunc)(); struct RegisterCamHelper @@ -120,4 +124,4 @@ struct RegisterCamHelper CamHelperCreateFunc createFunc); }; -} // namespace RPi +} /* namespace RPi */ diff --git a/src/ipa/raspberrypi/controller/agc_algorithm.hpp b/src/ipa/raspberrypi/controller/agc_algorithm.hpp index 51900b687778..b718e595193b 100644 --- a/src/ipa/raspberrypi/controller/agc_algorithm.hpp +++ b/src/ipa/raspberrypi/controller/agc_algorithm.hpp @@ -16,7 +16,7 @@ class AgcAlgorithm : public Algorithm { public: AgcAlgorithm(Controller *controller) : Algorithm(controller) {} - // An AGC algorithm must provide the following: + /* An AGC algorithm must provide the following: */ virtual unsigned int getConvergenceFrames() const = 0; virtual void setEv(double ev) = 0; virtual void setFlickerPeriod(libcamera::utils::Duration flickerPeriod) = 0; @@ -28,4 +28,4 @@ public: virtual void setConstraintMode(std::string const &contraintModeName) = 0; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/agc_status.h b/src/ipa/raspberrypi/controller/agc_status.h index 5d04c61d04bd..6723bc9e8200 100644 --- a/src/ipa/raspberrypi/controller/agc_status.h +++ b/src/ipa/raspberrypi/controller/agc_status.h @@ -8,20 +8,24 @@ #include -// The AGC algorithm should post the following structure into the image's -// "agc.status" metadata. +/* + * The AGC algorithm should post the following structure into the image's + * "agc.status" metadata. + */ #ifdef __cplusplus extern "C" { #endif -// Note: total_exposure_value will be reported as zero until the algorithm has -// seen statistics and calculated meaningful values. The contents should be -// ignored until then. +/* + * Note: total_exposure_value will be reported as zero until the algorithm has + * seen statistics and calculated meaningful values. The contents should be + * ignored until then. + */ struct AgcStatus { - libcamera::utils::Duration totalExposureValue; // value for all exposure and gain for this image - libcamera::utils::Duration targetExposureValue; // (unfiltered) target total exposure AGC is aiming for + libcamera::utils::Duration totalExposureValue; /* value for all exposure and gain for this image */ + libcamera::utils::Duration targetExposureValue; /* (unfiltered) target total exposure AGC is aiming for */ libcamera::utils::Duration shutterTime; double analogueGain; char exposureMode[32]; diff --git a/src/ipa/raspberrypi/controller/algorithm.cpp b/src/ipa/raspberrypi/controller/algorithm.cpp index cfcd18a96c93..e3afa647bdd2 100644 --- a/src/ipa/raspberrypi/controller/algorithm.cpp +++ b/src/ipa/raspberrypi/controller/algorithm.cpp @@ -31,7 +31,7 @@ void Algorithm::process([[maybe_unused]] StatisticsPtr &stats, { } -// For registering algorithms with the system: +/* For registering algorithms with the system: */ static std::map algorithms; std::map const &RPiController::getAlgorithms() diff --git a/src/ipa/raspberrypi/controller/algorithm.hpp b/src/ipa/raspberrypi/controller/algorithm.hpp index a33b14da2726..cad7c15ba5c8 100644 --- a/src/ipa/raspberrypi/controller/algorithm.hpp +++ b/src/ipa/raspberrypi/controller/algorithm.hpp @@ -6,8 +6,10 @@ */ #pragma once -// All algorithms should be derived from this class and made available to the -// Controller. +/* + * All algorithms should be derived from this class and made available to the + * Controller. + */ #include #include @@ -19,7 +21,7 @@ namespace RPiController { -// This defines the basic interface for all control algorithms. +/* This defines the basic interface for all control algorithms. */ class Algorithm { @@ -48,8 +50,10 @@ private: bool paused_; }; -// This code is for automatic registration of Front End algorithms with the -// system. +/* + * This code is for automatic registration of Front End algorithms with the + * system. + */ typedef Algorithm *(*AlgoCreateFunc)(Controller *controller); struct RegisterAlgorithm { @@ -57,4 +61,4 @@ struct RegisterAlgorithm { }; std::map const &getAlgorithms(); -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/alsc_status.h b/src/ipa/raspberrypi/controller/alsc_status.h index d3f579715594..e074f9359faa 100644 --- a/src/ipa/raspberrypi/controller/alsc_status.h +++ b/src/ipa/raspberrypi/controller/alsc_status.h @@ -6,8 +6,10 @@ */ #pragma once -// The ALSC algorithm should post the following structure into the image's -// "alsc.status" metadata. +/* + * The ALSC algorithm should post the following structure into the image's + * "alsc.status" metadata. + */ #ifdef __cplusplus extern "C" { diff --git a/src/ipa/raspberrypi/controller/awb_algorithm.hpp b/src/ipa/raspberrypi/controller/awb_algorithm.hpp index c5d2ca90263c..0de74fce4269 100644 --- a/src/ipa/raspberrypi/controller/awb_algorithm.hpp +++ b/src/ipa/raspberrypi/controller/awb_algorithm.hpp @@ -14,10 +14,10 @@ class AwbAlgorithm : public Algorithm { public: AwbAlgorithm(Controller *controller) : Algorithm(controller) {} - // An AWB algorithm must provide the following: + /* An AWB algorithm must provide the following: */ virtual unsigned int getConvergenceFrames() const = 0; virtual void setMode(std::string const &modeName) = 0; virtual void setManualGains(double manualR, double manualB) = 0; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/awb_status.h b/src/ipa/raspberrypi/controller/awb_status.h index bc428ed3206a..2f6e88ef6e7f 100644 --- a/src/ipa/raspberrypi/controller/awb_status.h +++ b/src/ipa/raspberrypi/controller/awb_status.h @@ -6,8 +6,10 @@ */ #pragma once -// The AWB algorithm places its results into both the image and global metadata, -// under the tag "awb.status". +/* + * The AWB algorithm places its results into both the image and global metadata, + * under the tag "awb.status". + */ #ifdef __cplusplus extern "C" { diff --git a/src/ipa/raspberrypi/controller/black_level_status.h b/src/ipa/raspberrypi/controller/black_level_status.h index d085f64b27fe..ad83ddad5283 100644 --- a/src/ipa/raspberrypi/controller/black_level_status.h +++ b/src/ipa/raspberrypi/controller/black_level_status.h @@ -6,14 +6,14 @@ */ #pragma once -// The "black level" algorithm stores the black levels to use. +/* The "black level" algorithm stores the black levels to use. */ #ifdef __cplusplus extern "C" { #endif struct BlackLevelStatus { - uint16_t black_level_r; // out of 16 bits + uint16_t black_level_r; /* out of 16 bits */ uint16_t black_level_g; uint16_t black_level_b; }; diff --git a/src/ipa/raspberrypi/controller/camera_mode.h b/src/ipa/raspberrypi/controller/camera_mode.h index 8b81ca9df725..47a0fea424ca 100644 --- a/src/ipa/raspberrypi/controller/camera_mode.h +++ b/src/ipa/raspberrypi/controller/camera_mode.h @@ -10,9 +10,11 @@ #include -// Description of a "camera mode", holding enough information for control -// algorithms to adapt their behaviour to the different modes of the camera, -// including binning, scaling, cropping etc. +/* + * Description of a "camera mode", holding enough information for control + * algorithms to adapt their behaviour to the different modes of the camera, + * including binning, scaling, cropping etc. + */ #ifdef __cplusplus extern "C" { @@ -21,27 +23,27 @@ extern "C" { #define CAMERA_MODE_NAME_LEN 32 struct CameraMode { - // bit depth of the raw camera output + /* bit depth of the raw camera output */ uint32_t bitdepth; - // size in pixels of frames in this mode + /* size in pixels of frames in this mode */ uint16_t width, height; - // size of full resolution uncropped frame ("sensor frame") + /* size of full resolution uncropped frame ("sensor frame") */ uint16_t sensorWidth, sensorHeight; - // binning factor (1 = no binning, 2 = 2-pixel binning etc.) + /* binning factor (1 = no binning, 2 = 2-pixel binning etc.) */ uint8_t binX, binY; - // location of top left pixel in the sensor frame + /* location of top left pixel in the sensor frame */ uint16_t cropX, cropY; - // scaling factor (so if uncropped, width*scaleX is sensorWidth) + /* scaling factor (so if uncropped, width*scaleX is sensorWidth) */ double scaleX, scaleY; - // scaling of the noise compared to the native sensor mode + /* scaling of the noise compared to the native sensor mode */ double noiseFactor; - // line time + /* line time */ libcamera::utils::Duration lineLength; - // any camera transform *not* reflected already in the camera tuning + /* any camera transform *not* reflected already in the camera tuning */ libcamera::Transform transform; - // minimum and maximum fame lengths in units of lines + /* minimum and maximum fame lengths in units of lines */ uint32_t minFrameLength, maxFrameLength; - // sensitivity of this mode + /* sensitivity of this mode */ double sensitivity; }; diff --git a/src/ipa/raspberrypi/controller/ccm_algorithm.hpp b/src/ipa/raspberrypi/controller/ccm_algorithm.hpp index b8b5879ba99c..9c7172f5782d 100644 --- a/src/ipa/raspberrypi/controller/ccm_algorithm.hpp +++ b/src/ipa/raspberrypi/controller/ccm_algorithm.hpp @@ -14,8 +14,8 @@ class CcmAlgorithm : public Algorithm { public: CcmAlgorithm(Controller *controller) : Algorithm(controller) {} - // A CCM algorithm must provide the following: + /* A CCM algorithm must provide the following: */ virtual void setSaturation(double saturation) = 0; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/ccm_status.h b/src/ipa/raspberrypi/controller/ccm_status.h index 7e41dd1ff3c0..4cdd8bed0311 100644 --- a/src/ipa/raspberrypi/controller/ccm_status.h +++ b/src/ipa/raspberrypi/controller/ccm_status.h @@ -6,7 +6,7 @@ */ #pragma once -// The "ccm" algorithm generates an appropriate colour matrix. +/* The "ccm" algorithm generates an appropriate colour matrix. */ #ifdef __cplusplus extern "C" { diff --git a/src/ipa/raspberrypi/controller/contrast_algorithm.hpp b/src/ipa/raspberrypi/controller/contrast_algorithm.hpp index c76f3cd759ba..1c0562e1c4a2 100644 --- a/src/ipa/raspberrypi/controller/contrast_algorithm.hpp +++ b/src/ipa/raspberrypi/controller/contrast_algorithm.hpp @@ -14,9 +14,9 @@ class ContrastAlgorithm : public Algorithm { public: ContrastAlgorithm(Controller *controller) : Algorithm(controller) {} - // A contrast algorithm must provide the following: + /* A contrast algorithm must provide the following: */ virtual void setBrightness(double brightness) = 0; virtual void setContrast(double contrast) = 0; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/contrast_status.h b/src/ipa/raspberrypi/controller/contrast_status.h index d7edd4e9990d..5eb084f78e71 100644 --- a/src/ipa/raspberrypi/controller/contrast_status.h +++ b/src/ipa/raspberrypi/controller/contrast_status.h @@ -6,8 +6,10 @@ */ #pragma once -// The "contrast" algorithm creates a gamma curve, optionally doing a little bit -// of contrast stretching based on the AGC histogram. +/* + * The "contrast" algorithm creates a gamma curve, optionally doing a little bit + * of contrast stretching based on the AGC histogram. + */ #ifdef __cplusplus extern "C" { diff --git a/src/ipa/raspberrypi/controller/controller.cpp b/src/ipa/raspberrypi/controller/controller.cpp index e0b152c74384..6d95fa55d1e4 100644 --- a/src/ipa/raspberrypi/controller/controller.cpp +++ b/src/ipa/raspberrypi/controller/controller.cpp @@ -89,8 +89,10 @@ Metadata &Controller::getGlobalMetadata() Algorithm *Controller::getAlgorithm(std::string const &name) const { - // The passed name must be the entire algorithm name, or must match the - // last part of it with a period (.) just before. + /* + * The passed name must be the entire algorithm name, or must match the + * last part of it with a period (.) just before. + */ size_t nameLen = name.length(); for (auto &algo : algorithms_) { char const *algoName = algo->name(); diff --git a/src/ipa/raspberrypi/controller/controller.hpp b/src/ipa/raspberrypi/controller/controller.hpp index a5e1eb38ab9d..29b2e8f34826 100644 --- a/src/ipa/raspberrypi/controller/controller.hpp +++ b/src/ipa/raspberrypi/controller/controller.hpp @@ -6,9 +6,11 @@ */ #pragma once -// The Controller is simply a container for a collecting together a number of -// "control algorithms" (such as AWB etc.) and for running them all in a -// convenient manner. +/* + * The Controller is simply a container for a collecting together a number of + * "control algorithms" (such as AWB etc.) and for running them all in a + * convenient manner. + */ #include #include @@ -25,10 +27,12 @@ class Algorithm; typedef std::unique_ptr AlgorithmPtr; typedef std::shared_ptr StatisticsPtr; -// The Controller holds a pointer to some global_metadata, which is how -// different controllers and control algorithms within them can exchange -// information. The Prepare function returns a pointer to metadata for this -// specific image, and which should be passed on to the Process function. +/* + * The Controller holds a pointer to some global_metadata, which is how + * different controllers and control algorithms within them can exchange + * information. The Prepare function returns a pointer to metadata for this + * specific image, and which should be passed on to the Process function. + */ class Controller { @@ -51,4 +55,4 @@ protected: bool switchModeCalled_; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/denoise_algorithm.hpp b/src/ipa/raspberrypi/controller/denoise_algorithm.hpp index 48de542ac4f3..7004fe55b41f 100644 --- a/src/ipa/raspberrypi/controller/denoise_algorithm.hpp +++ b/src/ipa/raspberrypi/controller/denoise_algorithm.hpp @@ -16,8 +16,8 @@ class DenoiseAlgorithm : public Algorithm { public: DenoiseAlgorithm(Controller *controller) : Algorithm(controller) {} - // A Denoise algorithm must provide the following: + /* A Denoise algorithm must provide the following: */ virtual void setMode(DenoiseMode mode) = 0; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/denoise_status.h b/src/ipa/raspberrypi/controller/denoise_status.h index 67a3c361387e..a41e7e89c428 100644 --- a/src/ipa/raspberrypi/controller/denoise_status.h +++ b/src/ipa/raspberrypi/controller/denoise_status.h @@ -6,7 +6,7 @@ */ #pragma once -// This stores the parameters required for Denoise. +/* This stores the parameters required for Denoise. */ #ifdef __cplusplus extern "C" { diff --git a/src/ipa/raspberrypi/controller/dpc_status.h b/src/ipa/raspberrypi/controller/dpc_status.h index a3ec2762573b..c99ad8c500a6 100644 --- a/src/ipa/raspberrypi/controller/dpc_status.h +++ b/src/ipa/raspberrypi/controller/dpc_status.h @@ -6,14 +6,14 @@ */ #pragma once -// The "DPC" algorithm sets defective pixel correction strength. +/* The "DPC" algorithm sets defective pixel correction strength. */ #ifdef __cplusplus extern "C" { #endif struct DpcStatus { - int strength; // 0 = "off", 1 = "normal", 2 = "strong" + int strength; /* 0 = "off", 1 = "normal", 2 = "strong" */ }; #ifdef __cplusplus diff --git a/src/ipa/raspberrypi/controller/focus_status.h b/src/ipa/raspberrypi/controller/focus_status.h index 656455100b45..c75795dc0621 100644 --- a/src/ipa/raspberrypi/controller/focus_status.h +++ b/src/ipa/raspberrypi/controller/focus_status.h @@ -8,9 +8,11 @@ #include -// The focus algorithm should post the following structure into the image's -// "focus.status" metadata. Recall that it's only reporting focus (contrast) -// measurements, it's not driving any kind of auto-focus algorithm! +/* + * The focus algorithm should post the following structure into the image's + * "focus.status" metadata. Recall that it's only reporting focus (contrast) + * measurements, it's not driving any kind of auto-focus algorithm! + */ #ifdef __cplusplus extern "C" { diff --git a/src/ipa/raspberrypi/controller/geq_status.h b/src/ipa/raspberrypi/controller/geq_status.h index 07fd5f0347ef..0ebb7ce71d5b 100644 --- a/src/ipa/raspberrypi/controller/geq_status.h +++ b/src/ipa/raspberrypi/controller/geq_status.h @@ -6,7 +6,7 @@ */ #pragma once -// The "GEQ" algorithm calculates the green equalisation thresholds +/* The "GEQ" algorithm calculates the green equalisation thresholds */ #ifdef __cplusplus extern "C" { diff --git a/src/ipa/raspberrypi/controller/histogram.cpp b/src/ipa/raspberrypi/controller/histogram.cpp index e865bef0057b..91a759b53d34 100644 --- a/src/ipa/raspberrypi/controller/histogram.cpp +++ b/src/ipa/raspberrypi/controller/histogram.cpp @@ -30,13 +30,13 @@ double Histogram::quantile(double q, int first, int last) const last = cumulative_.size() - 2; assert(first <= last); uint64_t items = q * total(); - while (first < last) // binary search to find the right bin + while (first < last) /* binary search to find the right bin */ { int middle = (first + last) / 2; if (cumulative_[middle + 1] > items) - last = middle; // between first and middle + last = middle; /* between first and middle */ else - first = middle + 1; // after middle + first = middle + 1; /* after middle */ } assert(items >= cumulative_[first] && items <= cumulative_[last + 1]); double frac = cumulative_[first + 1] == cumulative_[first] ? 0 @@ -59,6 +59,6 @@ double Histogram::interQuantileMean(double qLo, double qHi) const sumBinFreq += bin * freq; cumulFreq += freq; } - // add 0.5 to give an average for bin mid-points + /* add 0.5 to give an average for bin mid-points */ return sumBinFreq / cumulFreq + 0.5; } diff --git a/src/ipa/raspberrypi/controller/histogram.hpp b/src/ipa/raspberrypi/controller/histogram.hpp index 4ff5a56b0243..2ed8d9713764 100644 --- a/src/ipa/raspberrypi/controller/histogram.hpp +++ b/src/ipa/raspberrypi/controller/histogram.hpp @@ -10,8 +10,10 @@ #include #include -// A simple histogram class, for use in particular to find "quantiles" and -// averages between "quantiles". +/* + * A simple histogram class, for use in particular to find "quantiles" and + * averages between "quantiles". + */ namespace RPiController { @@ -29,16 +31,18 @@ public: } uint32_t bins() const { return cumulative_.size() - 1; } uint64_t total() const { return cumulative_[cumulative_.size() - 1]; } - // Cumulative frequency up to a (fractional) point in a bin. + /* Cumulative frequency up to a (fractional) point in a bin. */ uint64_t cumulativeFreq(double bin) const; - // Return the (fractional) bin of the point q (0 <= q <= 1) through the - // histogram. Optionally provide limits to help. + /* + * Return the (fractional) bin of the point q (0 <= q <= 1) through the + * histogram. Optionally provide limits to help. + */ double quantile(double q, int first = -1, int last = -1) const; - // Return the average histogram bin value between the two quantiles. + /* Return the average histogram bin value between the two quantiles. */ double interQuantileMean(double qLo, double qHi) const; private: std::vector cumulative_; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/lux_status.h b/src/ipa/raspberrypi/controller/lux_status.h index 8ccfd933829b..c1134bec3694 100644 --- a/src/ipa/raspberrypi/controller/lux_status.h +++ b/src/ipa/raspberrypi/controller/lux_status.h @@ -6,14 +6,16 @@ */ #pragma once -// The "lux" algorithm looks at the (AGC) histogram statistics of the frame and -// estimates the current lux level of the scene. It does this by a simple ratio -// calculation comparing to a reference image that was taken in known conditions -// with known statistics and a properly measured lux level. There is a slight -// problem with aperture, in that it may be variable without the system knowing -// or being aware of it. In this case an external application may set a -// "current_aperture" value if it wishes, which would be used in place of the -// (presumably meaningless) value in the image metadata. +/* + * The "lux" algorithm looks at the (AGC) histogram statistics of the frame and + * estimates the current lux level of the scene. It does this by a simple ratio + * calculation comparing to a reference image that was taken in known conditions + * with known statistics and a properly measured lux level. There is a slight + * problem with aperture, in that it may be variable without the system knowing + * or being aware of it. In this case an external application may set a + * "current_aperture" value if it wishes, which would be used in place of the + * (presumably meaningless) value in the image metadata. + */ #ifdef __cplusplus extern "C" { diff --git a/src/ipa/raspberrypi/controller/metadata.hpp b/src/ipa/raspberrypi/controller/metadata.hpp index a79a67d42cce..9f73e61ef91f 100644 --- a/src/ipa/raspberrypi/controller/metadata.hpp +++ b/src/ipa/raspberrypi/controller/metadata.hpp @@ -6,7 +6,7 @@ */ #pragma once -// A simple class for carrying arbitrary metadata, for example about an image. +/* A simple class for carrying arbitrary metadata, for example about an image. */ #include #include @@ -81,8 +81,10 @@ public: template T *getLocked(std::string const &tag) { - // This allows in-place access to the Metadata contents, - // for which you should be holding the lock. + /* + * This allows in-place access to the Metadata contents, + * for which you should be holding the lock. + */ auto it = data_.find(tag); if (it == data_.end()) return nullptr; @@ -92,13 +94,15 @@ public: template void setLocked(std::string const &tag, T const &value) { - // Use this only if you're holding the lock yourself. + /* Use this only if you're holding the lock yourself. */ data_[tag] = value; } - // Note: use of (lowercase) lock and unlock means you can create scoped - // locks with the standard lock classes. - // e.g. std::lock_guard lock(metadata) + /* + * Note: use of (lowercase) lock and unlock means you can create scoped + * locks with the standard lock classes. + * e.g. std::lock_guard lock(metadata) + */ void lock() { mutex_.lock(); } void unlock() { mutex_.unlock(); } @@ -107,4 +111,4 @@ private: std::map data_; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/noise_status.h b/src/ipa/raspberrypi/controller/noise_status.h index 8439a40213aa..60b995f4fa4f 100644 --- a/src/ipa/raspberrypi/controller/noise_status.h +++ b/src/ipa/raspberrypi/controller/noise_status.h @@ -6,7 +6,7 @@ */ #pragma once -// The "noise" algorithm stores an estimate of the noise profile for this image. +/* The "noise" algorithm stores an estimate of the noise profile for this image. */ #ifdef __cplusplus extern "C" { diff --git a/src/ipa/raspberrypi/controller/pwl.cpp b/src/ipa/raspberrypi/controller/pwl.cpp index 24ff3ea34f5f..d93cd2016dcf 100644 --- a/src/ipa/raspberrypi/controller/pwl.cpp +++ b/src/ipa/raspberrypi/controller/pwl.cpp @@ -66,11 +66,15 @@ double Pwl::eval(double x, int *spanPtr, bool updateSpan) const int Pwl::findSpan(double x, int span) const { - // Pwls are generally small, so linear search may well be faster than - // binary, though could review this if large PWls start turning up. + /* + * Pwls are generally small, so linear search may well be faster than + * binary, though could review this if large PWls start turning up. + */ int lastSpan = points_.size() - 2; - // some algorithms may call us with span pointing directly at the last - // control point + /* + * some algorithms may call us with span pointing directly at the last + * control point + */ span = std::max(0, std::min(lastSpan, span)); while (span < lastSpan && x >= points_[span + 1].x) span++; @@ -87,7 +91,7 @@ Pwl::PerpType Pwl::invert(Point const &xy, Point &perp, int &span, for (span = span + 1; span < (int)points_.size() - 1; span++) { Point spanVec = points_[span + 1] - points_[span]; double t = ((xy - points_[span]) % spanVec) / spanVec.len2(); - if (t < -eps) // off the start of this span + if (t < -eps) /* off the start of this span */ { if (span == 0) { perp = points_[span]; @@ -96,14 +100,14 @@ Pwl::PerpType Pwl::invert(Point const &xy, Point &perp, int &span, perp = points_[span]; return PerpType::Vertex; } - } else if (t > 1 + eps) // off the end of this span + } else if (t > 1 + eps) /* off the end of this span */ { if (span == (int)points_.size() - 2) { perp = points_[span + 1]; return PerpType::End; } prevOffEnd = true; - } else // a true perpendicular + } else /* a true perpendicular */ { perp = points_[span] + spanVec * t; return PerpType::Perpendicular; @@ -133,9 +137,11 @@ Pwl Pwl::inverse(bool *trueInverse, const double eps) const neither = true; } - // This is not a proper inverse if we found ourselves putting points - // onto both ends of the inverse, or if there were points that couldn't - // go on either. + /* + * This is not a proper inverse if we found ourselves putting points + * onto both ends of the inverse, or if there were points that couldn't + * go on either. + */ if (trueInverse) *trueInverse = !(neither || (appended && prepended)); @@ -154,8 +160,10 @@ Pwl Pwl::compose(Pwl const &other, const double eps) const otherSpan + 1 < (int)other.points_.size() && points_[thisSpan + 1].y >= other.points_[otherSpan + 1].x + eps) { - // next control point in result will be where this - // function's y reaches the next span in other + /* + * next control point in result will be where this + * function's y reaches the next span in other + */ thisX = points_[thisSpan].x + (other.points_[otherSpan + 1].x - points_[thisSpan].y) * @@ -164,15 +172,17 @@ Pwl Pwl::compose(Pwl const &other, const double eps) const } else if (abs(dy) > eps && otherSpan > 0 && points_[thisSpan + 1].y <= other.points_[otherSpan - 1].x - eps) { - // next control point in result will be where this - // function's y reaches the previous span in other + /* + * next control point in result will be where this + * function's y reaches the previous span in other + */ thisX = points_[thisSpan].x + (other.points_[otherSpan + 1].x - points_[thisSpan].y) * dx / dy; thisY = other.points_[--otherSpan].x; } else { - // we stay in the same span in other + /* we stay in the same span in other */ thisSpan++; thisX = points_[thisSpan].x, thisY = points_[thisSpan].y; diff --git a/src/ipa/raspberrypi/controller/pwl.hpp b/src/ipa/raspberrypi/controller/pwl.hpp index 4a38d1df5a33..e409c966baa0 100644 --- a/src/ipa/raspberrypi/controller/pwl.hpp +++ b/src/ipa/raspberrypi/controller/pwl.hpp @@ -63,44 +63,56 @@ public: Interval domain() const; Interval range() const; bool empty() const; - // Evaluate Pwl, optionally supplying an initial guess for the - // "span". The "span" may be optionally be updated. If you want to know - // the "span" value but don't have an initial guess you can set it to - // -1. + /* + * Evaluate Pwl, optionally supplying an initial guess for the + * "span". The "span" may be optionally be updated. If you want to know + * the "span" value but don't have an initial guess you can set it to + * -1. + */ double eval(double x, int *spanPtr = nullptr, bool updateSpan = true) const; - // Find perpendicular closest to xy, starting from span+1 so you can - // call it repeatedly to check for multiple closest points (set span to - // -1 on the first call). Also returns "pseudo" perpendiculars; see - // PerpType enum. + /* + * Find perpendicular closest to xy, starting from span+1 so you can + * call it repeatedly to check for multiple closest points (set span to + * -1 on the first call). Also returns "pseudo" perpendiculars; see + * PerpType enum. + */ enum class PerpType { - None, // no perpendicular found - Start, // start of Pwl is closest point - End, // end of Pwl is closest point - Vertex, // vertex of Pwl is closest point - Perpendicular // true perpendicular found + None, /* no perpendicular found */ + Start, /* start of Pwl is closest point */ + End, /* end of Pwl is closest point */ + Vertex, /* vertex of Pwl is closest point */ + Perpendicular /* true perpendicular found */ }; PerpType invert(Point const &xy, Point &perp, int &span, const double eps = 1e-6) const; - // Compute the inverse function. Indicate if it is a proper (true) - // inverse, or only a best effort (e.g. input was non-monotonic). + /* + * Compute the inverse function. Indicate if it is a proper (true) + * inverse, or only a best effort (e.g. input was non-monotonic). + */ Pwl inverse(bool *trueInverse = nullptr, const double eps = 1e-6) const; - // Compose two Pwls together, doing "this" first and "other" after. + /* Compose two Pwls together, doing "this" first and "other" after. */ Pwl compose(Pwl const &other, const double eps = 1e-6) const; - // Apply function to (x,y) values at every control point. + /* Apply function to (x,y) values at every control point. */ void map(std::function f) const; - // Apply function to (x, y0, y1) values wherever either Pwl has a - // control point. + /* + * Apply function to (x, y0, y1) values wherever either Pwl has a + * control point. + */ static void map2(Pwl const &pwl0, Pwl const &pwl1, std::function f); - // Combine two Pwls, meaning we create a new Pwl where the y values are - // given by running f wherever either has a knot. + /* + * Combine two Pwls, meaning we create a new Pwl where the y values are + * given by running f wherever either has a knot. + */ static Pwl combine(Pwl const &pwl0, Pwl const &pwl1, std::function f, const double eps = 1e-6); - // Make "this" match (at least) the given domain. Any extension my be - // clipped or linear. + /* + * Make "this" match (at least) the given domain. Any extension my be + * clipped or linear. + */ void matchDomain(Interval const &domain, bool clip = true, const double eps = 1e-6); Pwl &operator*=(double d); @@ -111,4 +123,4 @@ private: std::vector points_; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/rpi/agc.cpp b/src/ipa/raspberrypi/controller/rpi/agc.cpp index 738cf56c6be0..ec737ea13332 100644 --- a/src/ipa/raspberrypi/controller/rpi/agc.cpp +++ b/src/ipa/raspberrypi/controller/rpi/agc.cpp @@ -28,7 +28,7 @@ LOG_DEFINE_CATEGORY(RPiAgc) #define NAME "rpi.agc" -#define PIPELINE_BITS 13 // seems to be a 13-bit pipeline +#define PIPELINE_BITS 13 /* seems to be a 13-bit pipeline */ void AgcMeteringMode::read(boost::property_tree::ptree const ¶ms) { @@ -150,7 +150,7 @@ void AgcConfig::read(boost::property_tree::ptree const ¶ms) convergenceFrames = params.get("convergence_frames", 6); fastReduceThreshold = params.get("fast_reduce_threshold", 0.4); baseEv = params.get("base_ev", 1.0); - // Start with quite a low value as ramping up is easier than ramping down. + /* Start with quite a low value as ramping up is easier than ramping down. */ defaultExposureTime = params.get("default_exposure_time", 1000) * 1us; defaultAnalogueGain = params.get("default_analogueGain", 1.0); } @@ -170,8 +170,10 @@ Agc::Agc(Controller *controller) maxShutter_(0s), fixedShutter_(0s), fixedAnalogueGain_(0.0) { memset(&awb_, 0, sizeof(awb_)); - // Setting status_.totalExposureValue_ to zero initially tells us - // it's not been calculated yet (i.e. Process hasn't yet run). + /* + * Setting status_.totalExposureValue_ to zero initially tells us + * it's not been calculated yet (i.e. Process hasn't yet run). + */ memset(&status_, 0, sizeof(status_)); status_.ev = ev_; } @@ -185,16 +187,18 @@ void Agc::read(boost::property_tree::ptree const ¶ms) { LOG(RPiAgc, Debug) << "Agc"; config_.read(params); - // Set the config's defaults (which are the first ones it read) as our - // current modes, until someone changes them. (they're all known to - // exist at this point) + /* + * Set the config's defaults (which are the first ones it read) as our + * current modes, until someone changes them. (they're all known to + * exist at this point) + */ meteringModeName_ = config_.defaultMeteringMode; meteringMode_ = &config_.meteringModes[meteringModeName_]; exposureModeName_ = config_.defaultExposureMode; exposureMode_ = &config_.exposureModes[exposureModeName_]; constraintModeName_ = config_.defaultConstraintMode; constraintMode_ = &config_.constraintModes[constraintModeName_]; - // Set up the "last shutter/gain" values, in case AGC starts "disabled". + /* Set up the "last shutter/gain" values, in case AGC starts "disabled". */ status_.shutterTime = config_.defaultExposureTime; status_.analogueGain = config_.defaultAnalogueGain; } @@ -218,8 +222,10 @@ void Agc::resume() unsigned int Agc::getConvergenceFrames() const { - // If shutter and gain have been explicitly set, there is no - // convergence to happen, so no need to drop any frames - return zero. + /* + * If shutter and gain have been explicitly set, there is no + * convergence to happen, so no need to drop any frames - return zero. + */ if (fixedShutter_ && fixedAnalogueGain_) return 0; else @@ -244,14 +250,14 @@ void Agc::setMaxShutter(Duration maxShutter) void Agc::setFixedShutter(Duration fixedShutter) { fixedShutter_ = fixedShutter; - // Set this in case someone calls Pause() straight after. + /* Set this in case someone calls Pause() straight after. */ status_.shutterTime = clipShutter(fixedShutter_); } void Agc::setFixedAnalogueGain(double fixedAnalogueGain) { fixedAnalogueGain_ = fixedAnalogueGain_; - // Set this in case someone calls Pause() straight after. + /* Set this in case someone calls Pause() straight after. */ status_.analogueGain = fixedAnalogueGain; } @@ -280,30 +286,32 @@ void Agc::switchMode(CameraMode const &cameraMode, Duration fixedShutter = clipShutter(fixedShutter_); if (fixedShutter && fixedAnalogueGain_) { - // We're going to reset the algorithm here with these fixed values. + /* We're going to reset the algorithm here with these fixed values. */ fetchAwbStatus(metadata); double minColourGain = std::min({ awb_.gainR, awb_.gainG, awb_.gainB, 1.0 }); ASSERT(minColourGain != 0.0); - // This is the equivalent of computeTargetExposure and applyDigitalGain. + /* This is the equivalent of computeTargetExposure and applyDigitalGain. */ target_.totalExposureNoDG = fixedShutter_ * fixedAnalogueGain_; target_.totalExposure = target_.totalExposureNoDG / minColourGain; - // Equivalent of filterExposure. This resets any "history". + /* Equivalent of filterExposure. This resets any "history". */ filtered_ = target_; - // Equivalent of divideUpExposure. + /* Equivalent of divideUpExposure. */ filtered_.shutter = fixedShutter; filtered_.analogueGain = fixedAnalogueGain_; } else if (status_.totalExposureValue) { - // On a mode switch, various things could happen: - // - the exposure profile might change - // - a fixed exposure or gain might be set - // - the new mode's sensitivity might be different - // We cope with the last of these by scaling the target values. After - // that we just need to re-divide the exposure/gain according to the - // current exposure profile, which takes care of everything else. + /* + * On a mode switch, various things could happen: + * - the exposure profile might change + * - a fixed exposure or gain might be set + * - the new mode's sensitivity might be different + * We cope with the last of these by scaling the target values. After + * that we just need to re-divide the exposure/gain according to the + * current exposure profile, which takes care of everything else. + */ double ratio = lastSensitivity_ / cameraMode.sensitivity; target_.totalExposureNoDG *= ratio; @@ -313,29 +321,31 @@ void Agc::switchMode(CameraMode const &cameraMode, divideUpExposure(); } else { - // We come through here on startup, when at least one of the shutter - // or gain has not been fixed. We must still write those values out so - // that they will be applied immediately. We supply some arbitrary defaults - // for any that weren't set. - - // Equivalent of divideUpExposure. + /* + * We come through here on startup, when at least one of the shutter + * or gain has not been fixed. We must still write those values out so + * that they will be applied immediately. We supply some arbitrary defaults + * for any that weren't set. + */ + + /* Equivalent of divideUpExposure. */ filtered_.shutter = fixedShutter ? fixedShutter : config_.defaultExposureTime; filtered_.analogueGain = fixedAnalogueGain_ ? fixedAnalogueGain_ : config_.defaultAnalogueGain; } writeAndFinish(metadata, false); - // We must remember the sensitivity of this mode for the next SwitchMode. + /* We must remember the sensitivity of this mode for the next SwitchMode. */ lastSensitivity_ = cameraMode.sensitivity; } void Agc::prepare(Metadata *imageMetadata) { status_.digitalGain = 1.0; - fetchAwbStatus(imageMetadata); // always fetch it so that Process knows it's been done + fetchAwbStatus(imageMetadata); /* always fetch it so that Process knows it's been done */ if (status_.totalExposureValue) { - // Process has run, so we have meaningful values. + /* Process has run, so we have meaningful values. */ DeviceStatus deviceStatus; if (imageMetadata->get("device.status", deviceStatus) == 0) { Duration actualExposure = deviceStatus.shutterSpeed * @@ -343,14 +353,16 @@ void Agc::prepare(Metadata *imageMetadata) if (actualExposure) { status_.digitalGain = status_.totalExposureValue / actualExposure; LOG(RPiAgc, Debug) << "Want total exposure " << status_.totalExposureValue; - // Never ask for a gain < 1.0, and also impose - // some upper limit. Make it customisable? + /* + * Never ask for a gain < 1.0, and also impose + * some upper limit. Make it customisable? + */ status_.digitalGain = std::max(1.0, std::min(status_.digitalGain, 4.0)); LOG(RPiAgc, Debug) << "Actual exposure " << actualExposure; LOG(RPiAgc, Debug) << "Use digital_gain " << status_.digitalGain; LOG(RPiAgc, Debug) << "Effective exposure " << actualExposure * status_.digitalGain; - // Decide whether AEC/AGC has converged. + /* Decide whether AEC/AGC has converged. */ updateLockStatus(deviceStatus); } } else @@ -362,44 +374,52 @@ void Agc::prepare(Metadata *imageMetadata) void Agc::process(StatisticsPtr &stats, Metadata *imageMetadata) { frameCount_++; - // First a little bit of housekeeping, fetching up-to-date settings and - // configuration, that kind of thing. + /* + * First a little bit of housekeeping, fetching up-to-date settings and + * configuration, that kind of thing. + */ housekeepConfig(); - // Get the current exposure values for the frame that's just arrived. + /* Get the current exposure values for the frame that's just arrived. */ fetchCurrentExposure(imageMetadata); - // Compute the total gain we require relative to the current exposure. + /* Compute the total gain we require relative to the current exposure. */ double gain, targetY; computeGain(stats.get(), imageMetadata, gain, targetY); - // Now compute the target (final) exposure which we think we want. + /* Now compute the target (final) exposure which we think we want. */ computeTargetExposure(gain); - // Some of the exposure has to be applied as digital gain, so work out - // what that is. This function also tells us whether it's decided to - // "desaturate" the image more quickly. + /* + * Some of the exposure has to be applied as digital gain, so work out + * what that is. This function also tells us whether it's decided to + * "desaturate" the image more quickly. + */ bool desaturate = applyDigitalGain(gain, targetY); - // The results have to be filtered so as not to change too rapidly. + /* The results have to be filtered so as not to change too rapidly. */ filterExposure(desaturate); - // The last thing is to divide up the exposure value into a shutter time - // and analogue gain, according to the current exposure mode. + /* + * The last thing is to divide up the exposure value into a shutter time + * and analogue gain, according to the current exposure mode. + */ divideUpExposure(); - // Finally advertise what we've done. + /* Finally advertise what we've done. */ writeAndFinish(imageMetadata, desaturate); } void Agc::updateLockStatus(DeviceStatus const &deviceStatus) { - const double errorFactor = 0.10; // make these customisable? + const double errorFactor = 0.10; /* make these customisable? */ const int maxLockCount = 5; - // Reset "lock count" when we exceed this multiple of errorFactor + /* Reset "lock count" when we exceed this multiple of errorFactor */ const double resetMargin = 1.5; - // Add 200us to the exposure time error to allow for line quantisation. + /* Add 200us to the exposure time error to allow for line quantisation. */ Duration exposureError = lastDeviceStatus_.shutterSpeed * errorFactor + 200us; double gainError = lastDeviceStatus_.analogueGain * errorFactor; Duration targetError = lastTargetExposure_ * errorFactor; - // Note that we don't know the exposure/gain limits of the sensor, so - // the values we keep requesting may be unachievable. For this reason - // we only insist that we're close to values in the past few frames. + /* + * Note that we don't know the exposure/gain limits of the sensor, so + * the values we keep requesting may be unachievable. For this reason + * we only insist that we're close to values in the past few frames. + */ if (deviceStatus.shutterSpeed > lastDeviceStatus_.shutterSpeed - exposureError && deviceStatus.shutterSpeed < lastDeviceStatus_.shutterSpeed + exposureError && deviceStatus.analogueGain > lastDeviceStatus_.analogueGain - gainError && @@ -430,7 +450,7 @@ static void copyString(std::string const &s, char *d, size_t size) void Agc::housekeepConfig() { - // First fetch all the up-to-date settings, so no one else has to do it. + /* First fetch all the up-to-date settings, so no one else has to do it. */ status_.ev = ev_; status_.fixedShutter = clipShutter(fixedShutter_); status_.fixedAnalogueGain = fixedAnalogueGain_; @@ -438,8 +458,10 @@ void Agc::housekeepConfig() LOG(RPiAgc, Debug) << "ev " << status_.ev << " fixedShutter " << status_.fixedShutter << " fixedAnalogueGain " << status_.fixedAnalogueGain; - // Make sure the "mode" pointers point to the up-to-date things, if - // they've changed. + /* + * Make sure the "mode" pointers point to the up-to-date things, if + * they've changed. + */ if (strcmp(meteringModeName_.c_str(), status_.meteringMode)) { auto it = config_.meteringModes.find(meteringModeName_); if (it == config_.meteringModes.end()) @@ -491,7 +513,7 @@ void Agc::fetchCurrentExposure(Metadata *imageMetadata) void Agc::fetchAwbStatus(Metadata *imageMetadata) { - awb_.gainR = 1.0; // in case not found in metadata + awb_.gainR = 1.0; /* in case not found in metadata */ awb_.gainG = 1.0; awb_.gainB = 1.0; if (imageMetadata->get("awb.status", awb_) != 0) @@ -502,8 +524,10 @@ static double computeInitialY(bcm2835_isp_stats *stats, AwbStatus const &awb, double weights[], double gain) { bcm2835_isp_stats_region *regions = stats->agc_stats; - // Note how the calculation below means that equal weights give you - // "average" metering (i.e. all pixels equally important). + /* + * Note how the calculation below means that equal weights give you + * "average" metering (i.e. all pixels equally important). + */ double rSum = 0, gSum = 0, bSum = 0, pixelSum = 0; for (int i = 0; i < AGC_STATS_SIZE; i++) { double counted = regions[i].counted; @@ -525,11 +549,13 @@ static double computeInitialY(bcm2835_isp_stats *stats, AwbStatus const &awb, return ySum / pixelSum / (1 << PIPELINE_BITS); } -// We handle extra gain through EV by adjusting our Y targets. However, you -// simply can't monitor histograms once they get very close to (or beyond!) -// saturation, so we clamp the Y targets to this value. It does mean that EV -// increases don't necessarily do quite what you might expect in certain -// (contrived) cases. +/* + * We handle extra gain through EV by adjusting our Y targets. However, you + * simply can't monitor histograms once they get very close to (or beyond!) + * saturation, so we clamp the Y targets to this value. It does mean that EV + * increases don't necessarily do quite what you might expect in certain + * (contrived) cases. + */ #define EV_GAIN_Y_TARGET_LIMIT 0.9 @@ -546,18 +572,22 @@ void Agc::computeGain(bcm2835_isp_stats *statistics, Metadata *imageMetadata, double &gain, double &targetY) { struct LuxStatus lux = {}; - lux.lux = 400; // default lux level to 400 in case no metadata found + lux.lux = 400; /* default lux level to 400 in case no metadata found */ if (imageMetadata->get("lux.status", lux) != 0) LOG(RPiAgc, Warning) << "Agc: no lux level found"; Histogram h(statistics->hist[0].g_hist, NUM_HISTOGRAM_BINS); double evGain = status_.ev * config_.baseEv; - // The initial gain and target_Y come from some of the regions. After - // that we consider the histogram constraints. + /* + * The initial gain and target_Y come from some of the regions. After + * that we consider the histogram constraints. + */ targetY = config_.yTarget.eval(config_.yTarget.domain().clip(lux.lux)); targetY = std::min(EV_GAIN_Y_TARGET_LIMIT, targetY * evGain); - // Do this calculation a few times as brightness increase can be - // non-linear when there are saturated regions. + /* + * Do this calculation a few times as brightness increase can be + * non-linear when there are saturated regions. + */ gain = 1.0; for (int i = 0; i < 8; i++) { double initialY = computeInitialY(statistics, awb_, meteringMode_->weights, gain); @@ -565,7 +595,7 @@ void Agc::computeGain(bcm2835_isp_stats *statistics, Metadata *imageMetadata, gain *= extraGain; LOG(RPiAgc, Debug) << "Initial Y " << initialY << " target " << targetY << " gives gain " << gain; - if (extraGain < 1.01) // close enough + if (extraGain < 1.01) /* close enough */ break; } @@ -592,20 +622,23 @@ void Agc::computeGain(bcm2835_isp_stats *statistics, Metadata *imageMetadata, void Agc::computeTargetExposure(double gain) { if (status_.fixedShutter && status_.fixedAnalogueGain) { - // When ag and shutter are both fixed, we need to drive the - // total exposure so that we end up with a digital gain of at least - // 1/min_colour_gain. Otherwise we'd desaturate channels causing - // white to go cyan or magenta. + /* + * When ag and shutter are both fixed, we need to drive the + * total exposure so that we end up with a digital gain of at least + * 1/min_colour_gain. Otherwise we'd desaturate channels causing + * white to go cyan or magenta. + */ double minColourGain = std::min({ awb_.gainR, awb_.gainG, awb_.gainB, 1.0 }); ASSERT(minColourGain != 0.0); target_.totalExposure = status_.fixedShutter * status_.fixedAnalogueGain / minColourGain; } else { - // The statistics reflect the image without digital gain, so the final - // total exposure we're aiming for is: + /* + * The statistics reflect the image without digital gain, so the final + * total exposure we're aiming for is: + */ target_.totalExposure = current_.totalExposureNoDG * gain; - // The final target exposure is also limited to what the exposure - // mode allows. + /* The final target exposure is also limited to what the exposure mode allows. */ Duration maxShutter = status_.fixedShutter ? status_.fixedShutter : exposureMode_->shutter.back(); @@ -625,17 +658,21 @@ bool Agc::applyDigitalGain(double gain, double targetY) double minColourGain = std::min({ awb_.gainR, awb_.gainG, awb_.gainB, 1.0 }); ASSERT(minColourGain != 0.0); double dg = 1.0 / minColourGain; - // I think this pipeline subtracts black level and rescales before we - // get the stats, so no need to worry about it. + /* + * I think this pipeline subtracts black level and rescales before we + * get the stats, so no need to worry about it. + */ LOG(RPiAgc, Debug) << "after AWB, target dg " << dg << " gain " << gain << " target_Y " << targetY; - // Finally, if we're trying to reduce exposure but the target_Y is - // "close" to 1.0, then the gain computed for that constraint will be - // only slightly less than one, because the measured Y can never be - // larger than 1.0. When this happens, demand a large digital gain so - // that the exposure can be reduced, de-saturating the image much more - // quickly (and we then approach the correct value more quickly from - // below). + /* + * Finally, if we're trying to reduce exposure but the target_Y is + * "close" to 1.0, then the gain computed for that constraint will be + * only slightly less than one, because the measured Y can never be + * larger than 1.0. When this happens, demand a large digital gain so + * that the exposure can be reduced, de-saturating the image much more + * quickly (and we then approach the correct value more quickly from + * below). + */ bool desaturate = targetY > config_.fastReduceThreshold && gain < sqrt(targetY); if (desaturate) @@ -649,8 +686,10 @@ bool Agc::applyDigitalGain(double gain, double targetY) void Agc::filterExposure(bool desaturate) { double speed = config_.speed; - // AGC adapts instantly if both shutter and gain are directly specified - // or we're in the startup phase. + /* + * AGC adapts instantly if both shutter and gain are directly specified + * or we're in the startup phase. + */ if ((status_.fixedShutter && status_.fixedAnalogueGain) || frameCount_ <= config_.startupFrames) speed = 1.0; @@ -658,15 +697,19 @@ void Agc::filterExposure(bool desaturate) filtered_.totalExposure = target_.totalExposure; filtered_.totalExposureNoDG = target_.totalExposureNoDG; } else { - // If close to the result go faster, to save making so many - // micro-adjustments on the way. (Make this customisable?) + /* + * If close to the result go faster, to save making so many + * micro-adjustments on the way. (Make this customisable?) + */ if (filtered_.totalExposure < 1.2 * target_.totalExposure && filtered_.totalExposure > 0.8 * target_.totalExposure) speed = sqrt(speed); filtered_.totalExposure = speed * target_.totalExposure + filtered_.totalExposure * (1.0 - speed); - // When desaturing, take a big jump down in exposure_no_dg, - // which we'll hide with digital gain. + /* + * When desaturing, take a big jump down in exposure_no_dg, + * which we'll hide with digital gain. + */ if (desaturate) filtered_.totalExposureNoDG = target_.totalExposureNoDG; @@ -675,9 +718,11 @@ void Agc::filterExposure(bool desaturate) speed * target_.totalExposureNoDG + filtered_.totalExposureNoDG * (1.0 - speed); } - // We can't let the no_dg exposure deviate too far below the - // total exposure, as there might not be enough digital gain available - // in the ISP to hide it (which will cause nasty oscillation). + /* + * We can't let the no_dg exposure deviate too far below the + * total exposure, as there might not be enough digital gain available + * in the ISP to hide it (which will cause nasty oscillation). + */ if (filtered_.totalExposureNoDG < filtered_.totalExposure * config_.fastReduceThreshold) filtered_.totalExposureNoDG = filtered_.totalExposure * config_.fastReduceThreshold; @@ -687,9 +732,11 @@ void Agc::filterExposure(bool desaturate) void Agc::divideUpExposure() { - // Sending the fixed shutter/gain cases through the same code may seem - // unnecessary, but it will make more sense when extend this to cover - // variable aperture. + /* + * Sending the fixed shutter/gain cases through the same code may seem + * unnecessary, but it will make more sense when extend this to cover + * variable aperture. + */ Duration exposureValue = filtered_.totalExposureNoDG; Duration shutterTime; double analogueGain; @@ -721,18 +768,22 @@ void Agc::divideUpExposure() } LOG(RPiAgc, Debug) << "Divided up shutter and gain are " << shutterTime << " and " << analogueGain; - // Finally adjust shutter time for flicker avoidance (require both - // shutter and gain not to be fixed). + /* + * Finally adjust shutter time for flicker avoidance (require both + * shutter and gain not to be fixed). + */ if (!status_.fixedShutter && !status_.fixedAnalogueGain && status_.flickerPeriod) { int flickerPeriods = shutterTime / status_.flickerPeriod; if (flickerPeriods) { Duration newShutterTime = flickerPeriods * status_.flickerPeriod; analogueGain *= shutterTime / newShutterTime; - // We should still not allow the ag to go over the - // largest value in the exposure mode. Note that this - // may force more of the total exposure into the digital - // gain as a side-effect. + /* + * We should still not allow the ag to go over the + * largest value in the exposure mode. Note that this + * may force more of the total exposure into the digital + * gain as a side-effect. + */ analogueGain = std::min(analogueGain, exposureMode_->gain.back()); shutterTime = newShutterTime; } @@ -749,8 +800,10 @@ void Agc::writeAndFinish(Metadata *imageMetadata, bool desaturate) status_.targetExposureValue = desaturate ? 0s : target_.totalExposureNoDG; status_.shutterTime = filtered_.shutter; status_.analogueGain = filtered_.analogueGain; - // Write to metadata as well, in case anyone wants to update the camera - // immediately. + /* + * Write to metadata as well, in case anyone wants to update the camera + * immediately. + */ imageMetadata->set("agc.status", status_); LOG(RPiAgc, Debug) << "Output written, total exposure requested is " << filtered_.totalExposure; @@ -765,7 +818,7 @@ Duration Agc::clipShutter(Duration shutter) return shutter; } -// Register algorithm with the system. +/* Register algorithm with the system. */ static Algorithm *create(Controller *controller) { return (Algorithm *)new Agc(controller); diff --git a/src/ipa/raspberrypi/controller/rpi/agc.hpp b/src/ipa/raspberrypi/controller/rpi/agc.hpp index 4ed7293bce97..c2d68b60f15e 100644 --- a/src/ipa/raspberrypi/controller/rpi/agc.hpp +++ b/src/ipa/raspberrypi/controller/rpi/agc.hpp @@ -15,10 +15,12 @@ #include "../agc_status.h" #include "../pwl.hpp" -// This is our implementation of AGC. +/* This is our implementation of AGC. */ -// This is the number actually set up by the firmware, not the maximum possible -// number (which is 16). +/* + * This is the number actually set up by the firmware, not the maximum possible + * number (which is 16). + */ #define AGC_STATS_SIZE 15 @@ -73,7 +75,7 @@ public: Agc(Controller *controller); char const *name() const override; void read(boost::property_tree::ptree const ¶ms) override; - // AGC handles "pausing" for itself. + /* AGC handles "pausing" for itself. */ bool isPaused() const override; void pause() override; void resume() override; @@ -115,17 +117,17 @@ private: libcamera::utils::Duration shutter; double analogueGain; libcamera::utils::Duration totalExposure; - libcamera::utils::Duration totalExposureNoDG; // without digital gain + libcamera::utils::Duration totalExposureNoDG; /* without digital gain */ }; - ExposureValues current_; // values for the current frame - ExposureValues target_; // calculate the values we want here - ExposureValues filtered_; // these values are filtered towards target + ExposureValues current_; /* values for the current frame */ + ExposureValues target_; /* calculate the values we want here */ + ExposureValues filtered_; /* these values are filtered towards target */ AgcStatus status_; int lockCount_; DeviceStatus lastDeviceStatus_; libcamera::utils::Duration lastTargetExposure_; - double lastSensitivity_; // sensitivity of the previous camera mode - // Below here the "settings" that applications can change. + double lastSensitivity_; /* sensitivity of the previous camera mode */ + /* Below here the "settings" that applications can change. */ std::string meteringModeName_; std::string exposureModeName_; std::string constraintModeName_; @@ -136,4 +138,4 @@ private: double fixedAnalogueGain_; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/rpi/alsc.cpp b/src/ipa/raspberrypi/controller/rpi/alsc.cpp index 4929abc5b360..c9e1b9dc9f7d 100644 --- a/src/ipa/raspberrypi/controller/rpi/alsc.cpp +++ b/src/ipa/raspberrypi/controller/rpi/alsc.cpp @@ -14,7 +14,7 @@ #include "../awb_status.h" #include "alsc.hpp" -// Raspberry Pi ALSC (Auto Lens Shading Correction) algorithm. +/* Raspberry Pi ALSC (Auto Lens Shading Correction) algorithm. */ using namespace RPiController; using namespace libcamera; @@ -68,7 +68,7 @@ static void generateLut(double *lut, boost::property_tree::ptree const ¶ms) double r2 = (dx * dx + dy * dy) / R2; lut[num++] = (f1 * r2 + f2) * (f1 * r2 + f2) / - (f2 * f2); // this reproduces the cos^4 rule + (f2 * f2); /* this reproduces the cos^4 rule */ } } } @@ -171,7 +171,7 @@ void Alsc::initialise() frameCount2_ = frameCount_ = framePhase_ = 0; firstTime_ = true; ct_ = config_.defaultCt; - // The lambdas are initialised in the SwitchMode. + /* The lambdas are initialised in the SwitchMode. */ } void Alsc::waitForAysncThread() @@ -188,8 +188,10 @@ void Alsc::waitForAysncThread() static bool compareModes(CameraMode const &cm0, CameraMode const &cm1) { - // Return true if the modes crop from the sensor significantly differently, - // or if the user transform has changed. + /* + * Return true if the modes crop from the sensor significantly differently, + * or if the user transform has changed. + */ if (cm0.transform != cm1.transform) return true; int leftDiff = abs(cm0.cropX - cm1.cropX); @@ -198,9 +200,11 @@ static bool compareModes(CameraMode const &cm0, CameraMode const &cm1) cm1.cropX - cm1.scaleX * cm1.width); int bottomDiff = fabs(cm0.cropY + cm0.scaleY * cm0.height - cm1.cropY - cm1.scaleY * cm1.height); - // These thresholds are a rather arbitrary amount chosen to trigger - // when carrying on with the previously calculated tables might be - // worse than regenerating them (but without the adaptive algorithm). + /* + * These thresholds are a rather arbitrary amount chosen to trigger + * when carrying on with the previously calculated tables might be + * worse than regenerating them (but without the adaptive algorithm). + */ int thresholdX = cm0.sensorWidth >> 4; int thresholdY = cm0.sensorHeight >> 4; return leftDiff > thresholdX || rightDiff > thresholdX || @@ -210,28 +214,34 @@ static bool compareModes(CameraMode const &cm0, CameraMode const &cm1) void Alsc::switchMode(CameraMode const &cameraMode, [[maybe_unused]] Metadata *metadata) { - // We're going to start over with the tables if there's any "significant" - // change. + /* + * We're going to start over with the tables if there's any "significant" + * change. + */ bool resetTables = firstTime_ || compareModes(cameraMode_, cameraMode); - // Believe the colour temperature from the AWB, if there is one. + /* Believe the colour temperature from the AWB, if there is one. */ ct_ = getCt(metadata, ct_); - // Ensure the other thread isn't running while we do this. + /* Ensure the other thread isn't running while we do this. */ waitForAysncThread(); cameraMode_ = cameraMode; - // We must resample the luminance table like we do the others, but it's - // fixed so we can simply do it up front here. + /* + * We must resample the luminance table like we do the others, but it's + * fixed so we can simply do it up front here. + */ resampleCalTable(config_.luminanceLut, cameraMode_, luminanceTable_); if (resetTables) { - // Upon every "table reset", arrange for something sensible to be - // generated. Construct the tables for the previous recorded colour - // temperature. In order to start over from scratch we initialise - // the lambdas, but the rest of this code then echoes the code in - // doAlsc, without the adaptive algorithm. + /* + * Upon every "table reset", arrange for something sensible to be + * generated. Construct the tables for the previous recorded colour + * temperature. In order to start over from scratch we initialise + * the lambdas, but the rest of this code then echoes the code in + * doAlsc, without the adaptive algorithm. + */ for (int i = 0; i < XY; i++) lambdaR_[i] = lambdaB_[i] = 1.0; double calTableR[XY], calTableB[XY], calTableTmp[XY]; @@ -244,7 +254,7 @@ void Alsc::switchMode(CameraMode const &cameraMode, addLuminanceToTables(syncResults_, asyncLambdaR_, 1.0, asyncLambdaB_, luminanceTable_, config_.luminanceStrength); memcpy(prevSyncResults_, syncResults_, sizeof(prevSyncResults_)); - framePhase_ = config_.framePeriod; // run the algo again asap + framePhase_ = config_.framePeriod; /* run the algo again asap */ firstTime_ = false; } } @@ -260,7 +270,7 @@ void Alsc::fetchAsyncResults() double getCt(Metadata *metadata, double defaultCt) { AwbStatus awbStatus; - awbStatus.temperatureK = defaultCt; // in case nothing found + awbStatus.temperatureK = defaultCt; /* in case nothing found */ if (metadata->get("awb.status", awbStatus) != 0) LOG(RPiAlsc, Debug) << "no AWB results found, using " << awbStatus.temperatureK; @@ -282,18 +292,22 @@ static void copyStats(bcm2835_isp_stats_region regions[XY], StatisticsPtr &stats regions[i].g_sum = inputRegions[i].g_sum / gTable[i]; regions[i].b_sum = inputRegions[i].b_sum / bTable[i]; regions[i].counted = inputRegions[i].counted; - // (don't care about the uncounted value) + /* (don't care about the uncounted value) */ } } void Alsc::restartAsync(StatisticsPtr &stats, Metadata *imageMetadata) { LOG(RPiAlsc, Debug) << "Starting ALSC calculation"; - // Get the current colour temperature. It's all we need from the - // metadata. Default to the last CT value (which could be the default). + /* + * Get the current colour temperature. It's all we need from the + * metadata. Default to the last CT value (which could be the default). + */ ct_ = getCt(imageMetadata, ct_); - // We have to copy the statistics here, dividing out our best guess of - // the LSC table that the pipeline applied to them. + /* + * We have to copy the statistics here, dividing out our best guess of + * the LSC table that the pipeline applied to them. + */ AlscStatus alscStatus; if (imageMetadata->get("alsc.status", alscStatus) != 0) { LOG(RPiAlsc, Warning) @@ -317,8 +331,10 @@ void Alsc::restartAsync(StatisticsPtr &stats, Metadata *imageMetadata) void Alsc::prepare(Metadata *imageMetadata) { - // Count frames since we started, and since we last poked the async - // thread. + /* + * Count frames since we started, and since we last poked the async + * thread. + */ if (frameCount_ < (int)config_.startupFrames) frameCount_++; double speed = frameCount_ < (int)config_.startupFrames @@ -331,12 +347,12 @@ void Alsc::prepare(Metadata *imageMetadata) if (asyncStarted_ && asyncFinished_) fetchAsyncResults(); } - // Apply IIR filter to results and program into the pipeline. + /* Apply IIR filter to results and program into the pipeline. */ double *ptr = (double *)syncResults_, *pptr = (double *)prevSyncResults_; for (unsigned int i = 0; i < sizeof(syncResults_) / sizeof(double); i++) pptr[i] = speed * ptr[i] + (1.0 - speed) * pptr[i]; - // Put output values into status metadata. + /* Put output values into status metadata. */ AlscStatus status; memcpy(status.r, prevSyncResults_[0], sizeof(status.r)); memcpy(status.g, prevSyncResults_[1], sizeof(status.g)); @@ -346,8 +362,10 @@ void Alsc::prepare(Metadata *imageMetadata) void Alsc::process(StatisticsPtr &stats, Metadata *imageMetadata) { - // Count frames since we started, and since we last poked the async - // thread. + /* + * Count frames since we started, and since we last poked the async + * thread. + */ if (framePhase_ < (int)config_.framePeriod) framePhase_++; if (frameCount2_ < (int)config_.startupFrames) @@ -415,8 +433,10 @@ void getCalTable(double ct, std::vector const &calibrations, void resampleCalTable(double const calTableIn[XY], CameraMode const &cameraMode, double calTableOut[XY]) { - // Precalculate and cache the x sampling locations and phases to save - // recomputing them on every row. + /* + * Precalculate and cache the x sampling locations and phases to save + * recomputing them on every row. + */ int xLo[X], xHi[X]; double xf[X]; double scaleX = cameraMode.sensorWidth / @@ -434,7 +454,7 @@ void resampleCalTable(double const calTableIn[XY], xHi[i] = X - 1 - xHi[i]; } } - // Now march over the output table generating the new values. + /* Now march over the output table generating the new values. */ double scaleY = cameraMode.sensorHeight / (cameraMode.height * cameraMode.scaleY); double yOff = cameraMode.cropY / (double)cameraMode.sensorHeight; @@ -461,7 +481,7 @@ void resampleCalTable(double const calTableIn[XY], } } -// Calculate chrominance statistics (R/G and B/G) for each region. +/* Calculate chrominance statistics (R/G and B/G) for each region. */ static_assert(XY == AWB_REGIONS, "ALSC/AWB statistics region mismatch"); static void calculateCrCb(bcm2835_isp_stats_region *awbRegion, double cr[XY], double cb[XY], uint32_t minCount, uint16_t minG) @@ -512,8 +532,10 @@ void compensateLambdasForCal(double const calTable[XY], printf("]\n"); } -// Compute weight out of 1.0 which reflects how similar we wish to make the -// colours of these two regions. +/* + * Compute weight out of 1.0 which reflects how similar we wish to make the + * colours of these two regions. + */ static double computeWeight(double Ci, double Cj, double sigma) { if (Ci == InsufficientData || Cj == InsufficientData) @@ -522,11 +544,11 @@ static double computeWeight(double Ci, double Cj, double sigma) return exp(-diff * diff / 2); } -// Compute all weights. +/* Compute all weights. */ static void computeW(double const C[XY], double sigma, double W[XY][4]) { for (int i = 0; i < XY; i++) { - // Start with neighbour above and go clockwise. + /* Start with neighbour above and go clockwise. */ W[i][0] = i >= X ? computeWeight(C[i], C[i - X], sigma) : 0; W[i][1] = i % X < X - 1 ? computeWeight(C[i], C[i + 1], sigma) : 0; W[i][2] = i < XY - X ? computeWeight(C[i], C[i + X], sigma) : 0; @@ -534,17 +556,19 @@ static void computeW(double const C[XY], double sigma, double W[XY][4]) } } -// Compute M, the large but sparse matrix such that M * lambdas = 0. +/* Compute M, the large but sparse matrix such that M * lambdas = 0. */ static void constructM(double const C[XY], double const W[XY][4], double M[XY][4]) { double epsilon = 0.001; for (int i = 0; i < XY; i++) { - // Note how, if C[i] == INSUFFICIENT_DATA, the weights will all - // be zero so the equation is still set up correctly. + /* + * Note how, if C[i] == INSUFFICIENT_DATA, the weights will all + * be zero so the equation is still set up correctly. + */ int m = !!(i >= X) + !!(i % X < X - 1) + !!(i < XY - X) + - !!(i % X); // total number of neighbours - // we'll divide the diagonal out straight away + !!(i % X); /* total number of neighbours */ + /* we'll divide the diagonal out straight away */ double diagonal = (epsilon + W[i][0] + W[i][1] + W[i][2] + W[i][3]) * C[i]; M[i][0] = i >= X ? (W[i][0] * C[i - X] + epsilon / m * C[i]) / diagonal : 0; M[i][1] = i % X < X - 1 ? (W[i][1] * C[i + 1] + epsilon / m * C[i]) / diagonal : 0; @@ -553,9 +577,11 @@ static void constructM(double const C[XY], double const W[XY][4], } } -// In the compute_lambda_ functions, note that the matrix coefficients for the -// left/right neighbours are zero down the left/right edges, so we don't need -// need to test the i value to exclude them. +/* + * In the compute_lambda_ functions, note that the matrix coefficients for the + * left/right neighbours are zero down the left/right edges, so we don't need + * need to test the i value to exclude them. + */ static double computeLambdaBottom(int i, double const M[XY][4], double lambda[XY]) { @@ -585,7 +611,7 @@ static double computeLambdaTopEnd(int i, double const M[XY][4], return M[i][0] * lambda[i - X] + M[i][3] * lambda[i - 1]; } -// Gauss-Seidel iteration with over-relaxation. +/* Gauss-Seidel iteration with over-relaxation. */ static double gaussSeidel2Sor(double const M[XY][4], double omega, double lambda[XY], double lambdaBound) { @@ -610,8 +636,10 @@ static double gaussSeidel2Sor(double const M[XY][4], double omega, } lambda[i] = computeLambdaTopEnd(i, M, lambda); lambda[i] = std::clamp(lambda[i], min, max); - // Also solve the system from bottom to top, to help spread the updates - // better. + /* + * Also solve the system from bottom to top, to help spread the updates + * better. + */ lambda[i] = computeLambdaTopEnd(i, M, lambda); lambda[i] = std::clamp(lambda[i], min, max); for (i = XY - 2; i >= XY - X; i--) { @@ -637,7 +665,7 @@ static double gaussSeidel2Sor(double const M[XY][4], double omega, return maxDiff; } -// Normalise the values so that the smallest value is 1. +/* Normalise the values so that the smallest value is 1. */ static void normalise(double *ptr, size_t n) { double minval = ptr[0]; @@ -647,7 +675,7 @@ static void normalise(double *ptr, size_t n) ptr[i] /= minval; } -// Rescale the values so that the average value is 1. +/* Rescale the values so that the average value is 1. */ static void reaverage(Span data) { double sum = std::accumulate(data.begin(), data.end(), 0.0); @@ -670,15 +698,17 @@ static void runMatrixIterations(double const C[XY], double lambda[XY], << "Stop after " << i + 1 << " iterations"; break; } - // this happens very occasionally (so make a note), though - // doesn't seem to matter + /* + * this happens very occasionally (so make a note), though + * doesn't seem to matter + */ if (maxDiff > lastMaxDiff) LOG(RPiAlsc, Debug) << "Iteration " << i << ": max_diff gone up " << lastMaxDiff << " to " << maxDiff; lastMaxDiff = maxDiff; } - // We're going to normalise the lambdas so the total average is 1. + /* We're going to normalise the lambdas so the total average is 1. */ reaverage({ lambda, XY }); } @@ -712,41 +742,49 @@ void addLuminanceToTables(double results[3][Y][X], double const lambdaR[XY], void Alsc::doAlsc() { double cr[XY], cb[XY], wr[XY][4], wb[XY][4], calTableR[XY], calTableB[XY], calTableTmp[XY]; - // Calculate our R/B ("Cr"/"Cb") colour statistics, and assess which are - // usable. + /* + * Calculate our R/B ("Cr"/"Cb") colour statistics, and assess which are + * usable. + */ calculateCrCb(statistics_, cr, cb, config_.minCount, config_.minG); - // Fetch the new calibrations (if any) for this CT. Resample them in - // case the camera mode is not full-frame. + /* + * Fetch the new calibrations (if any) for this CT. Resample them in + * case the camera mode is not full-frame. + */ getCalTable(ct_, config_.calibrationsCr, calTableTmp); resampleCalTable(calTableTmp, cameraMode_, calTableR); getCalTable(ct_, config_.calibrationsCb, calTableTmp); resampleCalTable(calTableTmp, cameraMode_, calTableB); - // You could print out the cal tables for this image here, if you're - // tuning the algorithm... - // Apply any calibration to the statistics, so the adaptive algorithm - // makes only the extra adjustments. + /* + * You could print out the cal tables for this image here, if you're + * tuning the algorithm... + * Apply any calibration to the statistics, so the adaptive algorithm + * makes only the extra adjustments. + */ applyCalTable(calTableR, cr); applyCalTable(calTableB, cb); - // Compute weights between zones. + /* Compute weights between zones. */ computeW(cr, config_.sigmaCr, wr); computeW(cb, config_.sigmaCb, wb); - // Run Gauss-Seidel iterations over the resulting matrix, for R and B. + /* Run Gauss-Seidel iterations over the resulting matrix, for R and B. */ runMatrixIterations(cr, lambdaR_, wr, config_.omega, config_.nIter, config_.threshold, config_.lambdaBound); runMatrixIterations(cb, lambdaB_, wb, config_.omega, config_.nIter, config_.threshold, config_.lambdaBound); - // Fold the calibrated gains into our final lambda values. (Note that on - // the next run, we re-start with the lambda values that don't have the - // calibration gains included.) + /* + * Fold the calibrated gains into our final lambda values. (Note that on + * the next run, we re-start with the lambda values that don't have the + * calibration gains included.) + */ compensateLambdasForCal(calTableR, lambdaR_, asyncLambdaR_); compensateLambdasForCal(calTableB, lambdaB_, asyncLambdaB_); - // Fold in the luminance table at the appropriate strength. + /* Fold in the luminance table at the appropriate strength. */ addLuminanceToTables(asyncResults_, asyncLambdaR_, 1.0, asyncLambdaB_, luminanceTable_, config_.luminanceStrength); } -// Register algorithm with the system. +/* Register algorithm with the system. */ static Algorithm *create(Controller *controller) { return (Algorithm *)new Alsc(controller); diff --git a/src/ipa/raspberrypi/controller/rpi/alsc.hpp b/src/ipa/raspberrypi/controller/rpi/alsc.hpp index 7a0949d1ccc5..3ffc175d78b6 100644 --- a/src/ipa/raspberrypi/controller/rpi/alsc.hpp +++ b/src/ipa/raspberrypi/controller/rpi/alsc.hpp @@ -15,7 +15,7 @@ namespace RPiController { -// Algorithm to generate automagic LSC (Lens Shading Correction) tables. +/* Algorithm to generate automagic LSC (Lens Shading Correction) tables. */ struct AlscCalibration { double ct; @@ -23,11 +23,11 @@ struct AlscCalibration { }; struct AlscConfig { - // Only repeat the ALSC calculation every "this many" frames + /* Only repeat the ALSC calculation every "this many" frames */ uint16_t framePeriod; - // number of initial frames for which speed taken as 1.0 (maximum) + /* number of initial frames for which speed taken as 1.0 (maximum) */ uint16_t startupFrames; - // IIR filter speed applied to algorithm results + /* IIR filter speed applied to algorithm results */ double speed; double sigmaCr; double sigmaCb; @@ -39,9 +39,9 @@ struct AlscConfig { double luminanceStrength; std::vector calibrationsCr; std::vector calibrationsCb; - double defaultCt; // colour temperature if no metadata found - double threshold; // iteration termination threshold - double lambdaBound; // upper/lower bound for lambda from a value of 1 + double defaultCt; /* colour temperature if no metadata found */ + double threshold; /* iteration termination threshold */ + double lambdaBound; /* upper/lower bound for lambda from a value of 1 */ }; class Alsc : public Algorithm @@ -57,41 +57,45 @@ public: void process(StatisticsPtr &stats, Metadata *imageMetadata) override; private: - // configuration is read-only, and available to both threads + /* configuration is read-only, and available to both threads */ AlscConfig config_; bool firstTime_; CameraMode cameraMode_; double luminanceTable_[ALSC_CELLS_X * ALSC_CELLS_Y]; std::thread asyncThread_; - void asyncFunc(); // asynchronous thread function + void asyncFunc(); /* asynchronous thread function */ std::mutex mutex_; - // condvar for async thread to wait on + /* condvar for async thread to wait on */ std::condition_variable asyncSignal_; - // condvar for synchronous thread to wait on + /* condvar for synchronous thread to wait on */ std::condition_variable syncSignal_; - // for sync thread to check if async thread finished (requires mutex) + /* for sync thread to check if async thread finished (requires mutex) */ bool asyncFinished_; - // for async thread to check if it's been told to run (requires mutex) + /* for async thread to check if it's been told to run (requires mutex) */ bool asyncStart_; - // for async thread to check if it's been told to quit (requires mutex) + /* for async thread to check if it's been told to quit (requires mutex) */ bool asyncAbort_; - // The following are only for the synchronous thread to use: - // for sync thread to note its has asked async thread to run + /* + * The following are only for the synchronous thread to use: + * for sync thread to note its has asked async thread to run + */ bool asyncStarted_; - // counts up to framePeriod before restarting the async thread + /* counts up to framePeriod before restarting the async thread */ int framePhase_; - // counts up to startupFrames + /* counts up to startupFrames */ int frameCount_; - // counts up to startupFrames for Process function + /* counts up to startupFrames for Process function */ int frameCount2_; double syncResults_[3][ALSC_CELLS_Y][ALSC_CELLS_X]; double prevSyncResults_[3][ALSC_CELLS_Y][ALSC_CELLS_X]; void waitForAysncThread(); - // The following are for the asynchronous thread to use, though the main - // thread can set/reset them if the async thread is known to be idle: + /* + * The following are for the asynchronous thread to use, though the main + * thread can set/reset them if the async thread is known to be idle: + */ void restartAsync(StatisticsPtr &stats, Metadata *imageMetadata); - // copy out the results from the async thread so that it can be restarted + /* copy out the results from the async thread so that it can be restarted */ void fetchAsyncResults(); double ct_; bcm2835_isp_stats_region statistics_[ALSC_CELLS_Y * ALSC_CELLS_X]; @@ -103,4 +107,4 @@ private: double lambdaB_[ALSC_CELLS_X * ALSC_CELLS_Y]; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/rpi/awb.cpp b/src/ipa/raspberrypi/controller/rpi/awb.cpp index 74449c8c7591..d6f79f3a8e14 100644 --- a/src/ipa/raspberrypi/controller/rpi/awb.cpp +++ b/src/ipa/raspberrypi/controller/rpi/awb.cpp @@ -21,8 +21,10 @@ LOG_DEFINE_CATEGORY(RPiAwb) #define AWB_STATS_SIZE_X DEFAULT_AWB_REGIONS_X #define AWB_STATS_SIZE_Y DEFAULT_AWB_REGIONS_Y -// todo - the locking in this algorithm needs some tidying up as has been done -// elsewhere (ALSC and AGC). +/* + * todo - the locking in this algorithm needs some tidying up as has been done + * elsewhere (ALSC and AGC). + */ void AwbMode::read(boost::property_tree::ptree const ¶ms) { @@ -107,11 +109,11 @@ void AwbConfig::read(boost::property_tree::ptree const ¶ms) bayes = false; } } - fast = params.get("fast", bayes); // default to fast for Bayesian, otherwise slow + fast = params.get("fast", bayes); /* default to fast for Bayesian, otherwise slow */ whitepointR = params.get("whitepoint_r", 0.0); whitepointB = params.get("whitepoint_b", 0.0); if (bayes == false) - sensitivityR = sensitivityB = 1.0; // nor do sensitivities make any sense + sensitivityR = sensitivityB = 1.0; /* nor do sensitivities make any sense */ } Awb::Awb(Controller *controller) @@ -147,16 +149,18 @@ void Awb::read(boost::property_tree::ptree const ¶ms) void Awb::initialise() { frameCount_ = framePhase_ = 0; - // Put something sane into the status that we are filtering towards, - // just in case the first few frames don't have anything meaningful in - // them. + /* + * Put something sane into the status that we are filtering towards, + * just in case the first few frames don't have anything meaningful in + * them. + */ if (!config_.ctR.empty() && !config_.ctB.empty()) { syncResults_.temperatureK = config_.ctR.domain().clip(4000); syncResults_.gainR = 1.0 / config_.ctR.eval(syncResults_.temperatureK); syncResults_.gainG = 1.0; syncResults_.gainB = 1.0 / config_.ctB.eval(syncResults_.temperatureK); } else { - // random values just to stop the world blowing up + /* random values just to stop the world blowing up */ syncResults_.temperatureK = 4500; syncResults_.gainR = syncResults_.gainG = syncResults_.gainB = 1.0; } @@ -171,7 +175,7 @@ bool Awb::isPaused() const void Awb::pause() { - // "Pause" by fixing everything to the most recent values. + /* "Pause" by fixing everything to the most recent values. */ manualR_ = syncResults_.gainR = prevSyncResults_.gainR; manualB_ = syncResults_.gainB = prevSyncResults_.gainB; syncResults_.gainG = prevSyncResults_.gainG; @@ -186,8 +190,10 @@ void Awb::resume() unsigned int Awb::getConvergenceFrames() const { - // If not in auto mode, there is no convergence - // to happen, so no need to drop any frames - return zero. + /* + * If not in auto mode, there is no convergence + * to happen, so no need to drop any frames - return zero. + */ if (!isAutoEnabled()) return 0; else @@ -201,11 +207,13 @@ void Awb::setMode(std::string const &modeName) void Awb::setManualGains(double manualR, double manualB) { - // If any of these are 0.0, we swich back to auto. + /* If any of these are 0.0, we swich back to auto. */ manualR_ = manualR; manualB_ = manualB; - // If not in auto mode, set these values into the sync_results which - // means that Prepare() will adopt them immediately. + /* + * If not in auto mode, set these values into the sync_results which + * means that Prepare() will adopt them immediately. + */ if (!isAutoEnabled()) { syncResults_.gainR = prevSyncResults_.gainR = manualR_; syncResults_.gainG = prevSyncResults_.gainG = 1.0; @@ -216,8 +224,10 @@ void Awb::setManualGains(double manualR, double manualB) void Awb::switchMode([[maybe_unused]] CameraMode const &cameraMode, Metadata *metadata) { - // On the first mode switch we'll have no meaningful colour - // temperature, so try to dead reckon one if in manual mode. + /* + * On the first mode switch we'll have no meaningful colour + * temperature, so try to dead reckon one if in manual mode. + */ if (!isAutoEnabled() && firstSwitchMode_ && config_.bayes) { Pwl ctRInverse = config_.ctR.inverse(); Pwl ctBInverse = config_.ctB.inverse(); @@ -226,7 +236,7 @@ void Awb::switchMode([[maybe_unused]] CameraMode const &cameraMode, prevSyncResults_.temperatureK = (ctR + ctB) / 2; syncResults_.temperatureK = prevSyncResults_.temperatureK; } - // Let other algorithms know the current white balance values. + /* Let other algorithms know the current white balance values. */ metadata->set("awb.status", prevSyncResults_); firstSwitchMode_ = false; } @@ -241,8 +251,10 @@ void Awb::fetchAsyncResults() LOG(RPiAwb, Debug) << "Fetch AWB results"; asyncFinished_ = false; asyncStarted_ = false; - // It's possible manual gains could be set even while the async - // thread was running, so only copy the results if still in auto mode. + /* + * It's possible manual gains could be set even while the async + * thread was running, so only copy the results if still in auto mode. + */ if (isAutoEnabled()) syncResults_ = asyncResults_; } @@ -250,9 +262,9 @@ void Awb::fetchAsyncResults() void Awb::restartAsync(StatisticsPtr &stats, double lux) { LOG(RPiAwb, Debug) << "Starting AWB calculation"; - // this makes a new reference which belongs to the asynchronous thread + /* this makes a new reference which belongs to the asynchronous thread */ statistics_ = stats; - // store the mode as it could technically change + /* store the mode as it could technically change */ auto m = config_.modes.find(modeName_); mode_ = m != config_.modes.end() ? &m->second @@ -284,7 +296,7 @@ void Awb::prepare(Metadata *imageMetadata) if (asyncStarted_ && asyncFinished_) fetchAsyncResults(); } - // Finally apply IIR filter to results and put into metadata. + /* Finally apply IIR filter to results and put into metadata. */ memcpy(prevSyncResults_.mode, syncResults_.mode, sizeof(prevSyncResults_.mode)); prevSyncResults_.temperatureK = speed * syncResults_.temperatureK + @@ -304,17 +316,17 @@ void Awb::prepare(Metadata *imageMetadata) void Awb::process(StatisticsPtr &stats, Metadata *imageMetadata) { - // Count frames since we last poked the async thread. + /* Count frames since we last poked the async thread. */ if (framePhase_ < (int)config_.framePeriod) framePhase_++; LOG(RPiAwb, Debug) << "frame_phase " << framePhase_; - // We do not restart the async thread if we're not in auto mode. + /* We do not restart the async thread if we're not in auto mode. */ if (isAutoEnabled() && (framePhase_ >= (int)config_.framePeriod || frameCount_ < (int)config_.startupFrames)) { - // Update any settings and any image metadata that we need. + /* Update any settings and any image metadata that we need. */ struct LuxStatus luxStatus = {}; - luxStatus.lux = 400; // in case no metadata + luxStatus.lux = 400; /* in case no metadata */ if (imageMetadata->get("lux.status", luxStatus) != 0) LOG(RPiAwb, Debug) << "No lux metadata found"; LOG(RPiAwb, Debug) << "Awb lux value is " << luxStatus.lux; @@ -366,15 +378,21 @@ static void generateStats(std::vector &zones, void Awb::prepareStats() { zones_.clear(); - // LSC has already been applied to the stats in this pipeline, so stop - // any LSC compensation. We also ignore config_.fast in this version. + /* + * LSC has already been applied to the stats in this pipeline, so stop + * any LSC compensation. We also ignore config_.fast in this version. + */ generateStats(zones_, statistics_->awb_stats, config_.minPixels, config_.minG); - // we're done with these; we may as well relinquish our hold on the - // pointer. + /* + * we're done with these; we may as well relinquish our hold on the + * pointer. + */ statistics_.reset(); - // apply sensitivities, so values appear to come from our "canonical" - // sensor. + /* + * apply sensitivities, so values appear to come from our "canonical" + * sensor. + */ for (auto &zone : zones_) { zone.R *= config_.sensitivityR; zone.B *= config_.sensitivityB; @@ -383,14 +401,16 @@ void Awb::prepareStats() double Awb::computeDelta2Sum(double gainR, double gainB) { - // Compute the sum of the squared colour error (non-greyness) as it - // appears in the log likelihood equation. + /* + * Compute the sum of the squared colour error (non-greyness) as it + * appears in the log likelihood equation. + */ double delta2Sum = 0; for (auto &z : zones_) { double deltaR = gainR * z.R - 1 - config_.whitepointR; double deltaB = gainB * z.B - 1 - config_.whitepointB; double delta2 = deltaR * deltaR + deltaB * deltaB; - //LOG(RPiAwb, Debug) << "delta_r " << delta_r << " delta_b " << delta_b << " delta2 " << delta2; + /*LOG(RPiAwb, Debug) << "delta_r " << delta_r << " delta_b " << delta_b << " delta2 " << delta2; */ delta2 = std::min(delta2, config_.deltaLimit); delta2Sum += delta2; } @@ -399,15 +419,17 @@ double Awb::computeDelta2Sum(double gainR, double gainB) Pwl Awb::interpolatePrior() { - // Interpolate the prior log likelihood function for our current lux - // value. + /* + * Interpolate the prior log likelihood function for our current lux + * value. + */ if (lux_ <= config_.priors.front().lux) return config_.priors.front().prior; else if (lux_ >= config_.priors.back().lux) return config_.priors.back().prior; else { int idx = 0; - // find which two we lie between + /* find which two we lie between */ while (config_.priors[idx + 1].lux < lux_) idx++; double lux0 = config_.priors[idx].lux, @@ -424,8 +446,10 @@ Pwl Awb::interpolatePrior() static double interpolateQuadatric(Pwl::Point const &a, Pwl::Point const &b, Pwl::Point const &c) { - // Given 3 points on a curve, find the extremum of the function in that - // interval by fitting a quadratic. + /* + * Given 3 points on a curve, find the extremum of the function in that + * interval by fitting a quadratic. + */ const double eps = 1e-3; Pwl::Point ca = c - a, ba = b - a; double denominator = 2 * (ba.y * ca.x - ca.y * ba.x); @@ -434,17 +458,17 @@ static double interpolateQuadatric(Pwl::Point const &a, Pwl::Point const &b, double result = numerator / denominator + a.x; return std::max(a.x, std::min(c.x, result)); } - // has degenerated to straight line segment + /* has degenerated to straight line segment */ return a.y < c.y - eps ? a.x : (c.y < a.y - eps ? c.x : b.x); } double Awb::coarseSearch(Pwl const &prior) { - points_.clear(); // assume doesn't deallocate memory + points_.clear(); /* assume doesn't deallocate memory */ size_t bestPoint = 0; double t = mode_->ctLo; int spanR = 0, spanB = 0; - // Step down the CT curve evaluating log likelihood. + /* Step down the CT curve evaluating log likelihood. */ while (true) { double r = config_.ctR.eval(t, &spanR); double b = config_.ctB.eval(t, &spanB); @@ -462,13 +486,15 @@ double Awb::coarseSearch(Pwl const &prior) bestPoint = points_.size() - 1; if (t == mode_->ctHi) break; - // for even steps along the r/b curve scale them by the current t + /* for even steps along the r/b curve scale them by the current t */ t = std::min(t + t / 10 * config_.coarseStep, mode_->ctHi); } t = points_[bestPoint].x; LOG(RPiAwb, Debug) << "Coarse search found CT " << t; - // We have the best point of the search, but refine it with a quadratic - // interpolation around its neighbours. + /* + * We have the best point of the search, but refine it with a quadratic + * interpolation around its neighbours. + */ if (points_.size() > 2) { unsigned long bp = std::min(bestPoint, points_.size() - 2); bestPoint = std::max(1UL, bp); @@ -496,17 +522,21 @@ void Awb::fineSearch(double &t, double &r, double &b, Pwl const &prior) Pwl::Point transverse(bDiff, -rDiff); if (transverse.len2() < 1e-6) return; - // unit vector orthogonal to the b vs. r function (pointing outwards - // with r and b increasing) + /* + * unit vector orthogonal to the b vs. r function (pointing outwards + * with r and b increasing) + */ transverse = transverse / transverse.len(); double bestLogLikelihood = 0, bestT = 0, bestR = 0, bestB = 0; double transverseRange = config_.transverseNeg + config_.transversePos; const int maxNumDeltas = 12; - // a transverse step approximately every 0.01 r/b units + /* a transverse step approximately every 0.01 r/b units */ int numDeltas = floor(transverseRange * 100 + 0.5) + 1; numDeltas = numDeltas < 3 ? 3 : (numDeltas > maxNumDeltas ? maxNumDeltas : numDeltas); - // Step down CT curve. March a bit further if the transverse range is - // large. + /* + * Step down CT curve. March a bit further if the transverse range is + * large. + */ nsteps += numDeltas; for (int i = -nsteps; i <= nsteps; i++) { double tTest = t + i * step; @@ -514,10 +544,10 @@ void Awb::fineSearch(double &t, double &r, double &b, Pwl const &prior) prior.eval(prior.domain().clip(tTest)); double rCurve = config_.ctR.eval(tTest, &spanR); double bCurve = config_.ctB.eval(tTest, &spanB); - // x will be distance off the curve, y the log likelihood there + /* x will be distance off the curve, y the log likelihood there */ Pwl::Point points[maxNumDeltas]; int bestPoint = 0; - // Take some measurements transversely *off* the CT curve. + /* Take some measurements transversely *off* the CT curve. */ for (int j = 0; j < numDeltas; j++) { points[j].x = -config_.transverseNeg + (transverseRange * j) / (numDeltas - 1); @@ -533,8 +563,10 @@ void Awb::fineSearch(double &t, double &r, double &b, Pwl const &prior) if (points[j].y < points[bestPoint].y) bestPoint = j; } - // We have NUM_DELTAS points transversely across the CT curve, - // now let's do a quadratic interpolation for the best result. + /* + * We have NUM_DELTAS points transversely across the CT curve, + * now let's do a quadratic interpolation for the best result. + */ bestPoint = std::max(1, std::min(bestPoint, numDeltas - 2)); Pwl::Point rbTest = Pwl::Point(rCurve, bCurve) + transverse * interpolateQuadatric(points[bestPoint - 1], @@ -560,12 +592,16 @@ void Awb::fineSearch(double &t, double &r, double &b, Pwl const &prior) void Awb::awbBayes() { - // May as well divide out G to save computeDelta2Sum from doing it over - // and over. + /* + * May as well divide out G to save computeDelta2Sum from doing it over + * and over. + */ for (auto &z : zones_) z.R = z.R / (z.G + 1), z.B = z.B / (z.G + 1); - // Get the current prior, and scale according to how many zones are - // valid... not entirely sure about this. + /* + * Get the current prior, and scale according to how many zones are + * valid... not entirely sure about this. + */ Pwl prior = interpolatePrior(); prior *= zones_.size() / (double)(AWB_STATS_SIZE_X * AWB_STATS_SIZE_Y); prior.map([](double x, double y) { @@ -577,19 +613,23 @@ void Awb::awbBayes() LOG(RPiAwb, Debug) << "After coarse search: r " << r << " b " << b << " (gains r " << 1 / r << " b " << 1 / b << ")"; - // Not entirely sure how to handle the fine search yet. Mostly the - // estimated CT is already good enough, but the fine search allows us to - // wander transverely off the CT curve. Under some illuminants, where - // there may be more or less green light, this may prove beneficial, - // though I probably need more real datasets before deciding exactly how - // this should be controlled and tuned. + /* + * Not entirely sure how to handle the fine search yet. Mostly the + * estimated CT is already good enough, but the fine search allows us to + * wander transverely off the CT curve. Under some illuminants, where + * there may be more or less green light, this may prove beneficial, + * though I probably need more real datasets before deciding exactly how + * this should be controlled and tuned. + */ fineSearch(t, r, b, prior); LOG(RPiAwb, Debug) << "After fine search: r " << r << " b " << b << " (gains r " << 1 / r << " b " << 1 / b << ")"; - // Write results out for the main thread to pick up. Remember to adjust - // the gains from the ones that the "canonical sensor" would require to - // the ones needed by *this* sensor. + /* + * Write results out for the main thread to pick up. Remember to adjust + * the gains from the ones that the "canonical sensor" would require to + * the ones needed by *this* sensor. + */ asyncResults_.temperatureK = t; asyncResults_.gainR = 1.0 / r * config_.sensitivityR; asyncResults_.gainG = 1.0; @@ -599,10 +639,12 @@ void Awb::awbBayes() void Awb::awbGrey() { LOG(RPiAwb, Debug) << "Grey world AWB"; - // Make a separate list of the derivatives for each of red and blue, so - // that we can sort them to exclude the extreme gains. We could - // consider some variations, such as normalising all the zones first, or - // doing an L2 average etc. + /* + * Make a separate list of the derivatives for each of red and blue, so + * that we can sort them to exclude the extreme gains. We could + * consider some variations, such as normalising all the zones first, or + * doing an L2 average etc. + */ std::vector &derivsR(zones_); std::vector derivsB(derivsR); std::sort(derivsR.begin(), derivsR.end(), @@ -613,7 +655,7 @@ void Awb::awbGrey() [](RGB const &a, RGB const &b) { return a.G * b.B < b.G * a.B; }); - // Average the middle half of the values. + /* Average the middle half of the values. */ int discard = derivsR.size() / 4; RGB sumR(0, 0, 0), sumB(0, 0, 0); for (auto ri = derivsR.begin() + discard, @@ -622,7 +664,7 @@ void Awb::awbGrey() sumR += *ri, sumB += *bi; double gainR = sumR.G / (sumR.R + 1), gainB = sumB.G / (sumB.B + 1); - asyncResults_.temperatureK = 4500; // don't know what it is + asyncResults_.temperatureK = 4500; /* don't know what it is */ asyncResults_.gainR = gainR; asyncResults_.gainG = 1.0; asyncResults_.gainB = gainB; @@ -645,7 +687,7 @@ void Awb::doAwb() } } -// Register algorithm with the system. +/* Register algorithm with the system. */ static Algorithm *create(Controller *controller) { return (Algorithm *)new Awb(controller); diff --git a/src/ipa/raspberrypi/controller/rpi/awb.hpp b/src/ipa/raspberrypi/controller/rpi/awb.hpp index 91251d6be2da..597f3182da44 100644 --- a/src/ipa/raspberrypi/controller/rpi/awb.hpp +++ b/src/ipa/raspberrypi/controller/rpi/awb.hpp @@ -16,63 +16,73 @@ namespace RPiController { -// Control algorithm to perform AWB calculations. +/* Control algorithm to perform AWB calculations. */ struct AwbMode { void read(boost::property_tree::ptree const ¶ms); - double ctLo; // low CT value for search - double ctHi; // high CT value for search + double ctLo; /* low CT value for search */ + double ctHi; /* high CT value for search */ }; struct AwbPrior { void read(boost::property_tree::ptree const ¶ms); - double lux; // lux level - Pwl prior; // maps CT to prior log likelihood for this lux level + double lux; /* lux level */ + Pwl prior; /* maps CT to prior log likelihood for this lux level */ }; struct AwbConfig { AwbConfig() : defaultMode(nullptr) {} void read(boost::property_tree::ptree const ¶ms); - // Only repeat the AWB calculation every "this many" frames + /* Only repeat the AWB calculation every "this many" frames */ uint16_t framePeriod; - // number of initial frames for which speed taken as 1.0 (maximum) + /* number of initial frames for which speed taken as 1.0 (maximum) */ uint16_t startupFrames; - unsigned int convergenceFrames; // approx number of frames to converge - double speed; // IIR filter speed applied to algorithm results - bool fast; // "fast" mode uses a 16x16 rather than 32x32 grid - Pwl ctR; // function maps CT to r (= R/G) - Pwl ctB; // function maps CT to b (= B/G) - // table of illuminant priors at different lux levels + unsigned int convergenceFrames; /* approx number of frames to converge */ + double speed; /* IIR filter speed applied to algorithm results */ + bool fast; /* "fast" mode uses a 16x16 rather than 32x32 grid */ + Pwl ctR; /* function maps CT to r (= R/G) */ + Pwl ctB; /* + Pwl ctB; * function maps CT to b (= B/G) + * table of illuminant priors at different lux levels + */ std::vector priors; - // AWB "modes" (determines the search range) + /* AWB "modes" (determines the search range) */ std::map modes; - AwbMode *defaultMode; // mode used if no mode selected - // minimum proportion of pixels counted within AWB region for it to be - // "useful" + AwbMode *defaultMode; /* mode used if no mode selected */ + /* + * minimum proportion of pixels counted within AWB region for it to be + * "useful" + */ double minPixels; - // minimum G value of those pixels, to be regarded a "useful" + /* minimum G value of those pixels, to be regarded a "useful" */ uint16_t minG; - // number of AWB regions that must be "useful" in order to do the AWB - // calculation + /* + * number of AWB regions that must be "useful" in order to do the AWB + * calculation + */ uint32_t minRegions; - // clamp on colour error term (so as not to penalise non-grey excessively) + /* clamp on colour error term (so as not to penalise non-grey excessively) */ double deltaLimit; - // step size control in coarse search + /* step size control in coarse search */ double coarseStep; - // how far to wander off CT curve towards "more purple" + /* how far to wander off CT curve towards "more purple" */ double transversePos; - // how far to wander off CT curve towards "more green" + /* how far to wander off CT curve towards "more green" */ double transverseNeg; - // red sensitivity ratio (set to canonical sensor's R/G divided by this - // sensor's R/G) + /* + * red sensitivity ratio (set to canonical sensor's R/G divided by this + * sensor's R/G) + */ double sensitivityR; - // blue sensitivity ratio (set to canonical sensor's B/G divided by this - // sensor's B/G) + /* + * blue sensitivity ratio (set to canonical sensor's B/G divided by this + * sensor's B/G) + */ double sensitivityB; - // The whitepoint (which we normally "aim" for) can be moved. + /* The whitepoint (which we normally "aim" for) can be moved. */ double whitepointR; double whitepointB; - bool bayes; // use Bayesian algorithm + bool bayes; /* use Bayesian algorithm */ }; class Awb : public AwbAlgorithm @@ -83,7 +93,7 @@ public: char const *name() const override; void initialise() override; void read(boost::property_tree::ptree const ¶ms) override; - // AWB handles "pausing" for itself. + /* AWB handles "pausing" for itself. */ bool isPaused() const override; void pause() override; void resume() override; @@ -108,35 +118,39 @@ public: private: bool isAutoEnabled() const; - // configuration is read-only, and available to both threads + /* configuration is read-only, and available to both threads */ AwbConfig config_; std::thread asyncThread_; - void asyncFunc(); // asynchronous thread function + void asyncFunc(); /* asynchronous thread function */ std::mutex mutex_; - // condvar for async thread to wait on + /* condvar for async thread to wait on */ std::condition_variable asyncSignal_; - // condvar for synchronous thread to wait on + /* condvar for synchronous thread to wait on */ std::condition_variable syncSignal_; - // for sync thread to check if async thread finished (requires mutex) + /* for sync thread to check if async thread finished (requires mutex) */ bool asyncFinished_; - // for async thread to check if it's been told to run (requires mutex) + /* for async thread to check if it's been told to run (requires mutex) */ bool asyncStart_; - // for async thread to check if it's been told to quit (requires mutex) + /* for async thread to check if it's been told to quit (requires mutex) */ bool asyncAbort_; - // The following are only for the synchronous thread to use: - // for sync thread to note its has asked async thread to run + /* + * The following are only for the synchronous thread to use: + * for sync thread to note its has asked async thread to run + */ bool asyncStarted_; - // counts up to framePeriod before restarting the async thread + /* counts up to framePeriod before restarting the async thread */ int framePhase_; - int frameCount_; // counts up to startup_frames + int frameCount_; /* counts up to startup_frames */ AwbStatus syncResults_; AwbStatus prevSyncResults_; std::string modeName_; - // The following are for the asynchronous thread to use, though the main - // thread can set/reset them if the async thread is known to be idle: + /* + * The following are for the asynchronous thread to use, though the main + * thread can set/reset them if the async thread is known to be idle: + */ void restartAsync(StatisticsPtr &stats, double lux); - // copy out the results from the async thread so that it can be restarted + /* copy out the results from the async thread so that it can be restarted */ void fetchAsyncResults(); StatisticsPtr statistics_; AwbMode *mode_; @@ -152,11 +166,11 @@ private: void fineSearch(double &t, double &r, double &b, Pwl const &prior); std::vector zones_; std::vector points_; - // manual r setting + /* manual r setting */ double manualR_; - // manual b setting + /* manual b setting */ double manualB_; - bool firstSwitchMode_; // is this the first call to SwitchMode? + bool firstSwitchMode_; /* is this the first call to SwitchMode? */ }; static inline Awb::RGB operator+(Awb::RGB const &a, Awb::RGB const &b) @@ -176,4 +190,4 @@ static inline Awb::RGB operator*(Awb::RGB const &rgb, double d) return d * rgb; } -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/rpi/black_level.cpp b/src/ipa/raspberrypi/controller/rpi/black_level.cpp index 695b3129dd93..88fe4538d18d 100644 --- a/src/ipa/raspberrypi/controller/rpi/black_level.cpp +++ b/src/ipa/raspberrypi/controller/rpi/black_level.cpp @@ -34,7 +34,7 @@ char const *BlackLevel::name() const void BlackLevel::read(boost::property_tree::ptree const ¶ms) { uint16_t blackLevel = params.get( - "black_level", 4096); // 64 in 10 bits scaled to 16 bits + "black_level", 4096); /* 64 in 10 bits scaled to 16 bits */ blackLevelR_ = params.get("black_level_r", blackLevel); blackLevelG_ = params.get("black_level_g", blackLevel); blackLevelB_ = params.get("black_level_b", blackLevel); @@ -46,8 +46,10 @@ void BlackLevel::read(boost::property_tree::ptree const ¶ms) void BlackLevel::prepare(Metadata *imageMetadata) { - // Possibly we should think about doing this in a switch_mode or - // something? + /* + * Possibly we should think about doing this in a switch_mode or + * something? + */ struct BlackLevelStatus status; status.black_level_r = blackLevelR_; status.black_level_g = blackLevelG_; @@ -55,7 +57,7 @@ void BlackLevel::prepare(Metadata *imageMetadata) imageMetadata->set("black_level.status", status); } -// Register algorithm with the system. +/* Register algorithm with the system. */ static Algorithm *create(Controller *controller) { return new BlackLevel(controller); diff --git a/src/ipa/raspberrypi/controller/rpi/black_level.hpp b/src/ipa/raspberrypi/controller/rpi/black_level.hpp index 0d74f6a4c49b..f01c55151288 100644 --- a/src/ipa/raspberrypi/controller/rpi/black_level.hpp +++ b/src/ipa/raspberrypi/controller/rpi/black_level.hpp @@ -9,7 +9,7 @@ #include "../algorithm.hpp" #include "../black_level_status.h" -// This is our implementation of the "black level algorithm". +/* This is our implementation of the "black level algorithm". */ namespace RPiController { @@ -27,4 +27,4 @@ private: double blackLevelB_; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/rpi/ccm.cpp b/src/ipa/raspberrypi/controller/rpi/ccm.cpp index 24d8e5bd1fd8..9ad63b6e20d9 100644 --- a/src/ipa/raspberrypi/controller/rpi/ccm.cpp +++ b/src/ipa/raspberrypi/controller/rpi/ccm.cpp @@ -19,11 +19,13 @@ using namespace libcamera; LOG_DEFINE_CATEGORY(RPiCcm) -// This algorithm selects a CCM (Colour Correction Matrix) according to the -// colour temperature estimated by AWB (interpolating between known matricies as -// necessary). Additionally the amount of colour saturation can be controlled -// both according to the current estimated lux level and according to a -// saturation setting that is exposed to applications. +/* + * This algorithm selects a CCM (Colour Correction Matrix) according to the + * colour temperature estimated by AWB (interpolating between known matricies as + * necessary). Additionally the amount of colour saturation can be controlled + * both according to the current estimated lux level and according to a + * saturation setting that is exposed to applications. + */ #define NAME "rpi.ccm" @@ -125,11 +127,11 @@ void Ccm::prepare(Metadata *imageMetadata) { bool awbOk = false, luxOk = false; struct AwbStatus awb = {}; - awb.temperatureK = 4000; // in case no metadata + awb.temperatureK = 4000; /* in case no metadata */ struct LuxStatus lux = {}; - lux.lux = 400; // in case no metadata + lux.lux = 400; /* in case no metadata */ { - // grab mutex just once to get everything + /* grab mutex just once to get everything */ std::lock_guard lock(*imageMetadata); awbOk = getLocked(imageMetadata, "awb.status", awb); luxOk = getLocked(imageMetadata, "lux.status", lux); @@ -162,7 +164,7 @@ void Ccm::prepare(Metadata *imageMetadata) imageMetadata->set("ccm.status", ccmStatus); } -// Register algorithm with the system. +/* Register algorithm with the system. */ static Algorithm *create(Controller *controller) { return (Algorithm *)new Ccm(controller); diff --git a/src/ipa/raspberrypi/controller/rpi/ccm.hpp b/src/ipa/raspberrypi/controller/rpi/ccm.hpp index 4c4807b8a942..7622044ce49c 100644 --- a/src/ipa/raspberrypi/controller/rpi/ccm.hpp +++ b/src/ipa/raspberrypi/controller/rpi/ccm.hpp @@ -13,7 +13,7 @@ namespace RPiController { -// Algorithm to calculate colour matrix. Should be placed after AWB. +/* Algorithm to calculate colour matrix. Should be placed after AWB. */ struct Matrix { Matrix(double m0, double m1, double m2, double m3, double m4, double m5, @@ -72,4 +72,4 @@ private: double saturation_; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/rpi/contrast.cpp b/src/ipa/raspberrypi/controller/rpi/contrast.cpp index 169837576678..f11c834a0192 100644 --- a/src/ipa/raspberrypi/controller/rpi/contrast.cpp +++ b/src/ipa/raspberrypi/controller/rpi/contrast.cpp @@ -18,11 +18,13 @@ using namespace libcamera; LOG_DEFINE_CATEGORY(RPiContrast) -// This is a very simple control algorithm which simply retrieves the results of -// AGC and AWB via their "status" metadata, and applies digital gain to the -// colour channels in accordance with those instructions. We take care never to -// apply less than unity gains, as that would cause fully saturated pixels to go -// off-white. +/* + * This is a very simple control algorithm which simply retrieves the results of + * AGC and AWB via their "status" metadata, and applies digital gain to the + * colour channels in accordance with those instructions. We take care never to + * apply less than unity gains, as that would cause fully saturated pixels to go + * off-white. + */ #define NAME "rpi.contrast" @@ -38,15 +40,15 @@ char const *Contrast::name() const void Contrast::read(boost::property_tree::ptree const ¶ms) { - // enable adaptive enhancement by default + /* enable adaptive enhancement by default */ config_.ceEnable = params.get("ce_enable", 1); - // the point near the bottom of the histogram to move + /* the point near the bottom of the histogram to move */ config_.loHistogram = params.get("lo_histogram", 0.01); - // where in the range to try and move it to + /* where in the range to try and move it to */ config_.loLevel = params.get("lo_level", 0.015); - // but don't move by more than this + /* but don't move by more than this */ config_.loMax = params.get("lo_max", 500); - // equivalent values for the top of the histogram... + /* equivalent values for the top of the histogram... */ config_.hiHistogram = params.get("hi_histogram", 0.95); config_.hiLevel = params.get("hi_level", 0.95); config_.hiMax = params.get("hi_max", 2000); @@ -81,8 +83,10 @@ static void fillInStatus(ContrastStatus &status, double brightness, void Contrast::initialise() { - // Fill in some default values as Prepare will run before Process gets - // called. + /* + * Fill in some default values as Prepare will run before Process gets + * called. + */ fillInStatus(status_, brightness_, contrast_, config_.gammaCurve); } @@ -97,8 +101,10 @@ Pwl computeStretchCurve(Histogram const &histogram, { Pwl enhance; enhance.append(0, 0); - // If the start of the histogram is rather empty, try to pull it down a - // bit. + /* + * If the start of the histogram is rather empty, try to pull it down a + * bit. + */ double histLo = histogram.quantile(config.loHistogram) * (65536 / NUM_HISTOGRAM_BINS); double levelLo = config.loLevel * 65536; @@ -109,13 +115,17 @@ Pwl computeStretchCurve(Histogram const &histogram, LOG(RPiContrast, Debug) << "Final values " << histLo << " -> " << levelLo; enhance.append(histLo, levelLo); - // Keep the mid-point (median) in the same place, though, to limit the - // apparent amount of global brightness shift. + /* + * Keep the mid-point (median) in the same place, though, to limit the + * apparent amount of global brightness shift. + */ double mid = histogram.quantile(0.5) * (65536 / NUM_HISTOGRAM_BINS); enhance.append(mid, mid); - // If the top to the histogram is empty, try to pull the pixel values - // there up. + /* + * If the top to the histogram is empty, try to pull the pixel values + * there up. + */ double histHi = histogram.quantile(config.hiHistogram) * (65536 / NUM_HISTOGRAM_BINS); double levelHi = config.hiLevel * 65536; @@ -149,22 +159,30 @@ void Contrast::process(StatisticsPtr &stats, [[maybe_unused]] Metadata *imageMetadata) { Histogram histogram(stats->hist[0].g_hist, NUM_HISTOGRAM_BINS); - // We look at the histogram and adjust the gamma curve in the following - // ways: 1. Adjust the gamma curve so as to pull the start of the - // histogram down, and possibly push the end up. + /* + * We look at the histogram and adjust the gamma curve in the following + * ways: 1. Adjust the gamma curve so as to pull the start of the + * histogram down, and possibly push the end up. + */ Pwl gammaCurve = config_.gammaCurve; if (config_.ceEnable) { if (config_.loMax != 0 || config_.hiMax != 0) gammaCurve = computeStretchCurve(histogram, config_).compose(gammaCurve); - // We could apply other adjustments (e.g. partial equalisation) - // based on the histogram...? + /* + * We could apply other adjustments (e.g. partial equalisation) + * based on the histogram...? + */ } - // 2. Finally apply any manually selected brightness/contrast - // adjustment. + /* + * 2. Finally apply any manually selected brightness/contrast + * adjustment. + */ if (brightness_ != 0 || contrast_ != 1.0) gammaCurve = applyManualContrast(gammaCurve, brightness_, contrast_); - // And fill in the status for output. Use more points towards the bottom - // of the curve. + /* + * And fill in the status for output. Use more points towards the bottom + * of the curve. + */ ContrastStatus status; fillInStatus(status, brightness_, contrast_, gammaCurve); { @@ -173,7 +191,7 @@ void Contrast::process(StatisticsPtr &stats, } } -// Register algorithm with the system. +/* Register algorithm with the system. */ static Algorithm *create(Controller *controller) { return (Algorithm *)new Contrast(controller); diff --git a/src/ipa/raspberrypi/controller/rpi/contrast.hpp b/src/ipa/raspberrypi/controller/rpi/contrast.hpp index 5a6d530f63fd..4793dedc10ff 100644 --- a/src/ipa/raspberrypi/controller/rpi/contrast.hpp +++ b/src/ipa/raspberrypi/controller/rpi/contrast.hpp @@ -13,8 +13,10 @@ namespace RPiController { -// Back End algorithm to appaly correct digital gain. Should be placed after -// Back End AWB. +/* + * Back End algorithm to appaly correct digital gain. Should be placed after + * Back End AWB. + */ struct ContrastConfig { bool ceEnable; @@ -47,4 +49,4 @@ private: std::mutex mutex_; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/rpi/dpc.cpp b/src/ipa/raspberrypi/controller/rpi/dpc.cpp index 42154cf300b8..68ba5e3e37bb 100644 --- a/src/ipa/raspberrypi/controller/rpi/dpc.cpp +++ b/src/ipa/raspberrypi/controller/rpi/dpc.cpp @@ -14,8 +14,10 @@ using namespace libcamera; LOG_DEFINE_CATEGORY(RPiDpc) -// We use the lux status so that we can apply stronger settings in darkness (if -// necessary). +/* + * We use the lux status so that we can apply stronger settings in darkness (if + * necessary). + */ #define NAME "rpi.dpc" @@ -39,13 +41,13 @@ void Dpc::read(boost::property_tree::ptree const ¶ms) void Dpc::prepare(Metadata *imageMetadata) { DpcStatus dpcStatus = {}; - // Should we vary this with lux level or analogue gain? TBD. + /* Should we vary this with lux level or analogue gain? TBD. */ dpcStatus.strength = config_.strength; LOG(RPiDpc, Debug) << "strength " << dpcStatus.strength; imageMetadata->set("dpc.status", dpcStatus); } -// Register algorithm with the system. +/* Register algorithm with the system. */ static Algorithm *create(Controller *controller) { return (Algorithm *)new Dpc(controller); diff --git a/src/ipa/raspberrypi/controller/rpi/dpc.hpp b/src/ipa/raspberrypi/controller/rpi/dpc.hpp index 039310cc8d05..048fa2b8405e 100644 --- a/src/ipa/raspberrypi/controller/rpi/dpc.hpp +++ b/src/ipa/raspberrypi/controller/rpi/dpc.hpp @@ -11,7 +11,7 @@ namespace RPiController { -// Back End algorithm to apply appropriate GEQ settings. +/* Back End algorithm to apply appropriate GEQ settings. */ struct DpcConfig { int strength; @@ -29,4 +29,4 @@ private: DpcConfig config_; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/rpi/geq.cpp b/src/ipa/raspberrypi/controller/rpi/geq.cpp index 0da5efdf3d3d..14f226cf989c 100644 --- a/src/ipa/raspberrypi/controller/rpi/geq.cpp +++ b/src/ipa/raspberrypi/controller/rpi/geq.cpp @@ -18,8 +18,10 @@ using namespace libcamera; LOG_DEFINE_CATEGORY(RPiGeq) -// We use the lux status so that we can apply stronger settings in darkness (if -// necessary). +/* + * We use the lux status so that we can apply stronger settings in darkness (if + * necessary). + */ #define NAME "rpi.geq" @@ -50,7 +52,7 @@ void Geq::prepare(Metadata *imageMetadata) if (imageMetadata->get("lux.status", luxStatus)) LOG(RPiGeq, Warning) << "no lux data found"; DeviceStatus deviceStatus; - deviceStatus.analogueGain = 1.0; // in case not found + deviceStatus.analogueGain = 1.0; /* in case not found */ if (imageMetadata->get("device.status", deviceStatus)) LOG(RPiGeq, Warning) << "no device metadata - use analogue gain of 1x"; @@ -71,7 +73,7 @@ void Geq::prepare(Metadata *imageMetadata) imageMetadata->set("geq.status", geqStatus); } -// Register algorithm with the system. +/* Register algorithm with the system. */ static Algorithm *create(Controller *controller) { return (Algorithm *)new Geq(controller); diff --git a/src/ipa/raspberrypi/controller/rpi/geq.hpp b/src/ipa/raspberrypi/controller/rpi/geq.hpp index bdbc55b2e2d9..5ea424fc768d 100644 --- a/src/ipa/raspberrypi/controller/rpi/geq.hpp +++ b/src/ipa/raspberrypi/controller/rpi/geq.hpp @@ -11,12 +11,12 @@ namespace RPiController { -// Back End algorithm to apply appropriate GEQ settings. +/* Back End algorithm to apply appropriate GEQ settings. */ struct GeqConfig { uint16_t offset; double slope; - Pwl strength; // lux to strength factor + Pwl strength; /* lux to strength factor */ }; class Geq : public Algorithm @@ -31,4 +31,4 @@ private: GeqConfig config_; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/rpi/lux.cpp b/src/ipa/raspberrypi/controller/rpi/lux.cpp index 10654fbba94a..7f86f17470d8 100644 --- a/src/ipa/raspberrypi/controller/rpi/lux.cpp +++ b/src/ipa/raspberrypi/controller/rpi/lux.cpp @@ -25,8 +25,10 @@ LOG_DEFINE_CATEGORY(RPiLux) Lux::Lux(Controller *controller) : Algorithm(controller) { - // Put in some defaults as there will be no meaningful values until - // Process has run. + /* + * Put in some defaults as there will be no meaningful values until + * Process has run. + */ status_.aperture = 1.0; status_.lux = 400; } @@ -71,7 +73,7 @@ void Lux::process(StatisticsPtr &stats, Metadata *imageMetadata) sizeof(stats->hist[0].g_hist[0]); for (int i = 0; i < numBins; i++) sum += bin[i] * (uint64_t)i, num += bin[i]; - // add .5 to reflect the mid-points of bins + /* add .5 to reflect the mid-points of bins */ double currentY = sum / (double)num + .5; double gainRatio = referenceGain_ / currentGain; double shutterSpeedRatio = @@ -89,14 +91,16 @@ void Lux::process(StatisticsPtr &stats, Metadata *imageMetadata) std::unique_lock lock(mutex_); status_ = status; } - // Overwrite the metadata here as well, so that downstream - // algorithms get the latest value. + /* + * Overwrite the metadata here as well, so that downstream + * algorithms get the latest value. + */ imageMetadata->set("lux.status", status); } else LOG(RPiLux, Warning) << ": no device metadata"; } -// Register algorithm with the system. +/* Register algorithm with the system. */ static Algorithm *create(Controller *controller) { return (Algorithm *)new Lux(controller); diff --git a/src/ipa/raspberrypi/controller/rpi/lux.hpp b/src/ipa/raspberrypi/controller/rpi/lux.hpp index 98cfd0ac8bd0..7cf189363c06 100644 --- a/src/ipa/raspberrypi/controller/rpi/lux.hpp +++ b/src/ipa/raspberrypi/controller/rpi/lux.hpp @@ -13,7 +13,7 @@ #include "../lux_status.h" #include "../algorithm.hpp" -// This is our implementation of the "lux control algorithm". +/* This is our implementation of the "lux control algorithm". */ namespace RPiController { @@ -28,16 +28,18 @@ public: void setCurrentAperture(double aperture); private: - // These values define the conditions of the reference image, against - // which we compare the new image. + /* + * These values define the conditions of the reference image, against + * which we compare the new image. + */ libcamera::utils::Duration referenceshutterSpeed_; double referenceGain_; - double referenceAperture_; // units of 1/f - double referenceY_; // out of 65536 + double referenceAperture_; /* units of 1/f */ + double referenceY_; /* out of 65536 */ double referenceLux_; double currentAperture_; LuxStatus status_; std::mutex mutex_; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/rpi/noise.cpp b/src/ipa/raspberrypi/controller/rpi/noise.cpp index d6e4df4192f2..8117ce3608ed 100644 --- a/src/ipa/raspberrypi/controller/rpi/noise.cpp +++ b/src/ipa/raspberrypi/controller/rpi/noise.cpp @@ -34,8 +34,10 @@ char const *Noise::name() const void Noise::switchMode(CameraMode const &cameraMode, [[maybe_unused]] Metadata *metadata) { - // For example, we would expect a 2x2 binned mode to have a "noise - // factor" of sqrt(2x2) = 2. (can't be less than one, right?) + /* + * For example, we would expect a 2x2 binned mode to have a "noise + * factor" of sqrt(2x2) = 2. (can't be less than one, right?) + */ modeFactor_ = std::max(1.0, cameraMode.noiseFactor); } @@ -48,14 +50,16 @@ void Noise::read(boost::property_tree::ptree const ¶ms) void Noise::prepare(Metadata *imageMetadata) { struct DeviceStatus deviceStatus; - deviceStatus.analogueGain = 1.0; // keep compiler calm + deviceStatus.analogueGain = 1.0; /* keep compiler calm */ if (imageMetadata->get("device.status", deviceStatus) == 0) { - // There is a slight question as to exactly how the noise - // profile, specifically the constant part of it, scales. For - // now we assume it all scales the same, and we'll revisit this - // if it proves substantially wrong. NOTE: we may also want to - // make some adjustments based on the camera mode (such as - // binning), if we knew how to discover it... + /* + * There is a slight question as to exactly how the noise + * profile, specifically the constant part of it, scales. For + * now we assume it all scales the same, and we'll revisit this + * if it proves substantially wrong. NOTE: we may also want to + * make some adjustments based on the camera mode (such as + * binning), if we knew how to discover it... + */ double factor = sqrt(deviceStatus.analogueGain) / modeFactor_; struct NoiseStatus status; status.noise_constant = referenceConstant_ * factor; @@ -68,7 +72,7 @@ void Noise::prepare(Metadata *imageMetadata) LOG(RPiNoise, Warning) << " no metadata"; } -// Register algorithm with the system. +/* Register algorithm with the system. */ static Algorithm *create(Controller *controller) { return new Noise(controller); diff --git a/src/ipa/raspberrypi/controller/rpi/noise.hpp b/src/ipa/raspberrypi/controller/rpi/noise.hpp index ed6ffe910e27..56a4707b5ef2 100644 --- a/src/ipa/raspberrypi/controller/rpi/noise.hpp +++ b/src/ipa/raspberrypi/controller/rpi/noise.hpp @@ -9,7 +9,7 @@ #include "../algorithm.hpp" #include "../noise_status.h" -// This is our implementation of the "noise algorithm". +/* This is our implementation of the "noise algorithm". */ namespace RPiController { @@ -23,10 +23,10 @@ public: void prepare(Metadata *imageMetadata) override; private: - // the noise profile for analogue gain of 1.0 + /* the noise profile for analogue gain of 1.0 */ double referenceConstant_; double referenceSlope_; double modeFactor_; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/rpi/sdn.cpp b/src/ipa/raspberrypi/controller/rpi/sdn.cpp index 8707b6d9cd9e..6459b90fb9d4 100644 --- a/src/ipa/raspberrypi/controller/rpi/sdn.cpp +++ b/src/ipa/raspberrypi/controller/rpi/sdn.cpp @@ -17,8 +17,10 @@ using namespace libcamera; LOG_DEFINE_CATEGORY(RPiSdn) -// Calculate settings for the spatial denoise block using the noise profile in -// the image metadata. +/* + * Calculate settings for the spatial denoise block using the noise profile in + * the image metadata. + */ #define NAME "rpi.sdn" @@ -45,7 +47,7 @@ void Sdn::initialise() void Sdn::prepare(Metadata *imageMetadata) { struct NoiseStatus noiseStatus = {}; - noiseStatus.noise_slope = 3.0; // in case no metadata + noiseStatus.noise_slope = 3.0; /* in case no metadata */ if (imageMetadata->get("noise.status", noiseStatus) != 0) LOG(RPiSdn, Warning) << "no noise profile found"; LOG(RPiSdn, Debug) @@ -65,11 +67,11 @@ void Sdn::prepare(Metadata *imageMetadata) void Sdn::setMode(DenoiseMode mode) { - // We only distinguish between off and all other modes. + /* We only distinguish between off and all other modes. */ mode_ = mode; } -// Register algorithm with the system. +/* Register algorithm with the system. */ static Algorithm *create(Controller *controller) { return (Algorithm *)new Sdn(controller); diff --git a/src/ipa/raspberrypi/controller/rpi/sdn.hpp b/src/ipa/raspberrypi/controller/rpi/sdn.hpp index d9b18f296635..8b6e3db1a548 100644 --- a/src/ipa/raspberrypi/controller/rpi/sdn.hpp +++ b/src/ipa/raspberrypi/controller/rpi/sdn.hpp @@ -11,7 +11,7 @@ namespace RPiController { -// Algorithm to calculate correct spatial denoise (SDN) settings. +/* Algorithm to calculate correct spatial denoise (SDN) settings. */ class Sdn : public DenoiseAlgorithm { @@ -29,4 +29,4 @@ private: DenoiseMode mode_; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/rpi/sharpen.cpp b/src/ipa/raspberrypi/controller/rpi/sharpen.cpp index 775ed0fd2c46..b8f5b1005ac6 100644 --- a/src/ipa/raspberrypi/controller/rpi/sharpen.cpp +++ b/src/ipa/raspberrypi/controller/rpi/sharpen.cpp @@ -33,7 +33,7 @@ char const *Sharpen::name() const void Sharpen::switchMode(CameraMode const &cameraMode, [[maybe_unused]] Metadata *metadata) { - // can't be less than one, right? + /* can't be less than one, right? */ modeFactor_ = std::max(1.0, cameraMode.noiseFactor); } @@ -50,24 +50,30 @@ void Sharpen::read(boost::property_tree::ptree const ¶ms) void Sharpen::setStrength(double strength) { - // Note that this function is how an application sets the overall - // sharpening "strength". We call this the "user strength" field - // as there already is a strength_ field - being an internal gain - // parameter that gets passed to the ISP control code. Negative - // values are not allowed - coerce them to zero (no sharpening). + /* + * Note that this function is how an application sets the overall + * sharpening "strength". We call this the "user strength" field + * as there already is a strength_ field - being an internal gain + * parameter that gets passed to the ISP control code. Negative + * values are not allowed - coerce them to zero (no sharpening). + */ userStrength_ = std::max(0.0, strength); } void Sharpen::prepare(Metadata *imageMetadata) { - // The user_strength_ affects the algorithm's internal gain directly, but - // we adjust the limit and threshold less aggressively. Using a sqrt - // function is an arbitrary but gentle way of accomplishing this. + /* + * The user_strength_ affects the algorithm's internal gain directly, but + * we adjust the limit and threshold less aggressively. Using a sqrt + * function is an arbitrary but gentle way of accomplishing this. + */ double userStrengthSqrt = sqrt(userStrength_); struct SharpenStatus status; - // Binned modes seem to need the sharpening toned down with this - // pipeline, thus we use the mode_factor here. Also avoid - // divide-by-zero with the userStrengthSqrt. + /* + * Binned modes seem to need the sharpening toned down with this + * pipeline, thus we use the mode_factor here. Also avoid + * divide-by-zero with the userStrengthSqrt. + */ status.threshold = threshold_ * modeFactor_ / std::max(0.01, userStrengthSqrt); status.strength = strength_ / modeFactor_ * userStrength_; @@ -77,7 +83,7 @@ void Sharpen::prepare(Metadata *imageMetadata) imageMetadata->set("sharpen.status", status); } -// Register algorithm with the system. +/* Register algorithm with the system. */ static Algorithm *create(Controller *controller) { return new Sharpen(controller); diff --git a/src/ipa/raspberrypi/controller/rpi/sharpen.hpp b/src/ipa/raspberrypi/controller/rpi/sharpen.hpp index ced917f3c42b..18c45fd4e2a7 100644 --- a/src/ipa/raspberrypi/controller/rpi/sharpen.hpp +++ b/src/ipa/raspberrypi/controller/rpi/sharpen.hpp @@ -9,7 +9,7 @@ #include "../sharpen_algorithm.hpp" #include "../sharpen_status.h" -// This is our implementation of the "sharpen algorithm". +/* This is our implementation of the "sharpen algorithm". */ namespace RPiController { @@ -31,4 +31,4 @@ private: double userStrength_; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/sharpen_algorithm.hpp b/src/ipa/raspberrypi/controller/sharpen_algorithm.hpp index 888f4569c56a..22cc6090f8fc 100644 --- a/src/ipa/raspberrypi/controller/sharpen_algorithm.hpp +++ b/src/ipa/raspberrypi/controller/sharpen_algorithm.hpp @@ -14,8 +14,8 @@ class SharpenAlgorithm : public Algorithm { public: SharpenAlgorithm(Controller *controller) : Algorithm(controller) {} - // A sharpness control algorithm must provide the following: + /* A sharpness control algorithm must provide the following: */ virtual void setStrength(double strength) = 0; }; -} // namespace RPiController +} /* namespace RPiController */ diff --git a/src/ipa/raspberrypi/controller/sharpen_status.h b/src/ipa/raspberrypi/controller/sharpen_status.h index 2b0490742fba..5ea21ab23f91 100644 --- a/src/ipa/raspberrypi/controller/sharpen_status.h +++ b/src/ipa/raspberrypi/controller/sharpen_status.h @@ -6,20 +6,20 @@ */ #pragma once -// The "sharpen" algorithm stores the strength to use. +/* The "sharpen" algorithm stores the strength to use. */ #ifdef __cplusplus extern "C" { #endif struct SharpenStatus { - // controls the smallest level of detail (or noise!) that sharpening will pick up + /* controls the smallest level of detail (or noise!) that sharpening will pick up */ double threshold; - // the rate at which the sharpening response ramps once above the threshold + /* the rate at which the sharpening response ramps once above the threshold */ double strength; - // upper limit of the allowed sharpening response + /* upper limit of the allowed sharpening response */ double limit; - // The sharpening strength requested by the user or application. + /* The sharpening strength requested by the user or application. */ double userStrength; }; diff --git a/src/ipa/raspberrypi/md_parser.hpp b/src/ipa/raspberrypi/md_parser.hpp index e505108a7adc..a05ab800b9ae 100644 --- a/src/ipa/raspberrypi/md_parser.hpp +++ b/src/ipa/raspberrypi/md_parser.hpp @@ -152,4 +152,4 @@ private: OffsetMap offsets_; }; -} // namespace RPi +} /* namespace RPi */