# SPDX-License-Identifier: LGPL-2.1-or-later # # Copyright (C) 2019, Google Inc. # %YAML 1.2 --- controls: - Location: type: int32_t description: | Camera mounting location enum: - name: CameraLocationFront value: 0 description: | The camera is mounted on the front side of the device, facing the user - name: CameraLocationBack value: 1 description: | The camera is mounted on the back side of the device, facing away from the user - name: CameraLocationExternal value: 2 description: | The camera is attached to the device in a way that allows it to be moved freely - Rotation: type: int32_t description: | The camera rotation is expressed as the angular difference in degrees between two reference systems, one relative to the camera module, and one defined on the external world scene to be captured when projected on the image sensor pixel array. A camera sensor has a 2-dimensional reference system 'Rc' defined by its pixel array read-out order. The origin is set to the first pixel being read out, the X-axis points along the column read-out direction towards the last columns, and the Y-axis along the row read-out direction towards the last row. A typical example for a sensor with a 2592x1944 pixel array matrix observed from the front is 2591 X-axis 0 <------------------------+ 0 .......... ... ..........! .......... ... ..........! Y-axis ... ! .......... ... ..........! .......... ... ..........! 1943 V The external world scene reference system 'Rs' is a 2-dimensional reference system on the focal plane of the camera module. The origin is placed on the top-left corner of the visible scene, the X-axis points towards the right, and the Y-axis points towards the bottom of the scene. The top, bottom, left and right directions are intentionally not defined and depend on the environment in which the camera is used. A typical example of a (very common) picture of a shark swimming from left to right, as seen from the camera, is 0 X-axis 0 +-------------------------------------> ! ! ! ! |\____)\___ ! ) _____ __`< ! |/ )/ ! ! ! V Y-axis With the reference system 'Rs' placed on the camera focal plane. ¸.·˙! ¸.·˙ ! _ ¸.·˙ ! +-/ \-+¸.·˙ ! | (o) | ! Camera focal plane +-----+˙·.¸ ! ˙·.¸ ! ˙·.¸ ! ˙·.¸! When projected on the sensor's pixel array, the image and the associated reference system 'Rs' are typically (but not always) inverted, due to the camera module's lens optical inversion effect. Assuming the above represented scene of the swimming shark, the lens inversion projects the scene and its reference system onto the sensor pixel array, seen from the front of the camera sensor, as follow Y-axis ^ ! ! ! ! |\_____)\__ ! ) ____ ___.< ! |/ )/ ! ! ! 0 +-------------------------------------> 0 X-axis Note the shark being upside-down. The resulting projected reference system is named 'Rp'. The camera rotation property is then defined as the angular difference in the counter-clockwise direction between the camera reference system 'Rc' and the projected scene reference system 'Rp'. It is expressed in degrees as a number in the range [0, 360[. Examples 0 degrees camera rotation Y-Rp ^ Y-Rc ! ^ ! ! ! ! ! ! ! ! ! ! ! ! ! ! ! ! 0 +-------------------------------------> ! 0 X-Rp 0 +-------------------------------------> 0 X-Rc X-Rc 0 <------------------------------------+ 0 X-Rp 0 ! <------------------------------------+ 0 ! ! ! ! ! ! ! ! ! ! ! ! ! ! ! ! V ! Y-Rc V Y-Rp 90 degrees camera rotation 0 Y-Rc 0 +--------------------> ! Y-Rp ! ^ ! ! ! ! ! ! ! ! ! ! ! ! ! ! ! ! ! ! ! 0 +-------------------------------------> ! 0 X-Rp ! ! ! ! V X-Rc 180 degrees camera rotation 0 <------------------------------------+ 0 X-Rc ! Y-Rp ! ^ ! ! ! ! ! ! ! ! ! ! ! ! ! ! V ! Y-Rc 0 +-------------------------------------> 0 X-Rp 270 degrees camera rotation 0 Y-Rc 0 +--------------------> ! 0 ! <-----------------------------------+ 0 ! X-Rp ! ! ! ! ! ! ! ! ! ! ! ! ! ! ! ! ! ! V ! Y-Rp ! ! ! ! V X-Rc Example one - Webcam A camera module installed on the user facing part of a laptop screen casing used for video calls. The captured images are meant to be displayed in landscape mode (width > height) on the laptop screen. The camera is typically mounted upside-down to compensate the lens optical inversion effect. Y-Rp Y-Rc ^ ^ ! ! ! ! ! |\_____)\__ ! ! ) ____ ___.< ! ! |/ )/ ! ! ! ! ! ! ! 0 +-------------------------------------> ! 0 X-Rp 0 +-------------------------------------> 0 X-Rc The two reference systems are aligned, the resulting camera rotation is 0 degrees, no rotation correction needs to be applied to the resulting image once captured to memory buffers to correctly display it to users. +--------------------------------------+ ! ! ! ! ! ! ! |\____)\___ ! ! ) _____ __`< ! ! |/ )/ ! ! ! ! ! ! ! +--------------------------------------+ If the camera sensor is not mounted upside-down to compensate for the lens optical inversion, the two reference systems will not be aligned, with 'Rp' being rotated 180 degrees relatively to 'Rc'. X-Rc 0 <------------------------------------+ 0 ! Y-Rp ! ^ ! ! ! ! |\_____)\__ ! ! ) ____ ___.< ! ! |/ )/ ! ! ! ! ! ! V ! Y-Rc 0 +-------------------------------------> 0 X-Rp The image once captured to memory will then be rotated by 180 degrees +--------------------------------------+ ! ! ! ! ! ! ! __/(_____/| ! ! >.___ ____ ( ! ! \( \| ! ! ! ! ! ! ! +--------------------------------------+ A software rotation correction of 180 degrees should be applied to correctly display the image. +--------------------------------------+ ! ! ! ! ! ! ! |\____)\___ ! ! ) _____ __`< ! ! |/ )/ ! ! ! ! ! ! ! +--------------------------------------+ Example two - Phone camera A camera installed on the back side of a mobile device facing away from the user. The captured images are meant to be displayed in portrait mode (height > width) to match the device screen orientation and the device usage orientation used when taking the picture. The camera sensor is typically mounted with its pixel array longer side aligned to the device longer side, upside-down mounted to compensate for the lens optical inversion effect. 0 Y-Rc 0 +--------------------> ! Y-Rp ! ^ ! ! ! ! ! ! ! ! |\_____)\__ ! ! ) ____ ___.< ! ! |/ )/ ! ! ! ! ! ! ! 0 +-------------------------------------> ! 0 X-Rp ! ! ! ! V X-Rc The two reference systems are not aligned and the 'Rp' reference system is rotated by 90 degrees in the counter-clockwise direction relatively to the 'Rc' reference system. The image once captured to memory will be rotated. +-------------------------------------+ | _ _ | | \ / | | | | | | | | | | | > | | < | | | | | | | . | | V | +-------------------------------------+ A correction of 90 degrees in counter-clockwise direction has to be applied to correctly display the image in portrait mode on the device screen. +--------------------+ | | | | | | | | | | | | | |\____)\___ | | ) _____ __`< | | |/ )/ | | | | | | | | | | | +--------------------+ - Model: type: string description: | The model name shall to the extent possible describe the sensor. For most devices this is the model name of the sensor. While for some devices the sensor model is unavailable as the sensor or the entire camera is part of a larger unit and exposed as a black-box to the system. In such cases the model name of the smallest device that contains the camera sensor shall be used. The model name is not meant to be a camera name displayed to the end-user, but may be combined with other camera information to create a camera name. The model name is not guaranteed to be unique in the system nor is it guaranteed to be stable or have any other properties required to make it a good candidate to be used as a permanent identifier of a camera. The model name shall describe the camera in a human readable format and shall be encoded in ASCII. Example model names are 'ov5670', 'imx219' or 'Logitech Webcam C930e'. - UnitCellSize: type: Size description: | The pixel unit cell physical size, in nanometers. The UnitCellSize properties defines the horizontal and vertical sizes of a single pixel unit, including its active and non-active parts. In other words, it expresses the horizontal and vertical distance between the top-left corners of adjacent pixels. The property can be used to calculate the physical size of the sensor's pixel array area and for calibration purposes. - PixelArraySize: type: Size description: | The camera sensor pixel array readable area vertical and horizontal sizes, in pixels. The PixelArraySize property defines the size in pixel units of the readable part of full pixel array matrix, including optical black pixels used for calibration, pixels which are not considered valid for capture and active pixels containing valid image data. The property describes the maximum size of the raw data captured by the camera, which might not correspond to the physical size of the sensor pixel array matrix, as some portions of the physical pixel array matrix are not accessible and cannot be transmitted out. For example, let's consider a pixel array matrix assembled as follows +--------------------------------------------------+ |xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx| |xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx| |xxDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDxx| |xxDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDxx| |xxDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDxx| |xxDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDxx| |xxDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDxx| |xxDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDxx| ... ... ... ... ... ... ... ... ... ... |xxDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDxx| |xxDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDxx| |xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx| |xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx| +--------------------------------------------------+ starting with two lines of non-readable pixels (x), followed by N lines of readable data (D) surrounded by two columns of non-readable pixels on each side, and ending with two more lines of non-readable pixels. Only the readable portion is transmitted to the receiving side, defining the sizes of the largest possible buffer of raw data that can be presented to applications. PixelArraySize.width /----------------------------------------------/ +----------------------------------------------+ / |DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD| | |DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD| | |DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD| | |DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD| | |DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD| | |DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD| | PixelArraySize.height ... ... ... ... ... ... ... ... ... ... |DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD| | |DDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDDD| | +----------------------------------------------+ / This defines a rectangle whose top-left corner is placed in position (0, 0) and whose vertical and horizontal sizes are defined by this property. All other rectangles that describe portions of the pixel array, such as the optical black pixels rectangles and active pixel areas, are defined relatively to this rectangle. All the coordinates are expressed relative to the default sensor readout direction, without any transformation (such as horizontal and vertical flipping) applied. When mapping them to the raw pixel buffer, applications shall take any configured transformation into account. \todo Rename this property to Size once we will have property categories (i.e. Properties::PixelArray::Size) - PixelArrayOpticalBlackRectangles: type: Rectangle size: [n] description: | The pixel array region(s) which contain optical black pixels considered valid for calibration purposes. This property describes the position and size of optical black pixel regions in the raw data buffer as stored in memory, which might differ from their actual physical location in the pixel array matrix. It is important to note, in fact, that camera sensors might automatically reorder or skip portions of their pixels array matrix when transmitting data to the receiver. For instance, a sensor may merge the top and bottom optical black rectangles into a single rectangle, transmitted at the beginning of the frame. The pixel array contains several areas with different purposes, interleaved by lines and columns which are said not to be valid for capturing purposes. Invalid lines and columns are defined as invalid as they could be positioned too close to the chip margins or to the optical black shielding placed on top of optical black pixels. PixelArraySize.width /----------------------------------------------/ x1 x2 +--o---------------------------------------o---+ / |IIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIII| | |IIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIII| | y1 oIIOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOII| | |IIOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOII| | |IIOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOII| | y2 oIIOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOII| | |IIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIII| | |IIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIII| | y3 |IIOOPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPOOII| | |IIOOPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPOOII| | PixelArraySize.height |IIOOPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPOOII| | ... ... ... ... ... ... ... ... ... ... y4 |IIOOPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPOOII| | |IIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIII| | |IIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIIII| | +----------------------------------------------+ / The readable pixel array matrix is composed by 2 invalid lines (I) 4 lines of valid optical black pixels (O) 2 invalid lines (I) n lines of valid pixel data (P) 2 invalid lines (I) And the position of the optical black pixel rectangles is defined by PixelArrayOpticalBlackRectangles = { { x1, y1, x2 - x1 + 1, y2 - y1 + 1 }, { x1, y3, 2, y4 - y3 + 1 }, { x2, y3, 2, y4 - y3 + 1 }, }; If the camera, when capturing the full pixel array matrix, automatically skips the invalid lines and columns, producing the following data buffer, when captured to memory PixelArraySize.width /----------------------------------------------/ x1 +--------------------------------------------o-+ / |OOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOO| | |OOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOO| | |OOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOO| | |OOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOOO| | y1 oOOPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPOO| | |OOPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPOO| | |OOPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPOO| | PixelArraySize.height ... ... ... ... ... | ... ... ... ... ... | |OOPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPOO| | |OOPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPPOO| | +----------------------------------------------+ / then the invalid lines and columns should not be reported as part of the PixelArraySize property in first place. In this case, the position of the black pixel rectangles will be PixelArrayOpticalBlackRectangles = { { 0, 0, y1 + 1, PixelArraySize[0] }, { 0, y1, 2, PixelArraySize[1] - y1 + 1 }, { x1, y1, 2, PixelArraySize[1] - y1 + 1 }, }; \todo Rename this property to Size once we will have property categories (i.e. Properties::PixelArray::OpticalBlackRectangles) - PixelArrayActiveAreas: type: Rectangle size: [n] description: | The PixelArrayActiveAreas property defines the (possibly multiple and overlapping) portions of the camera sensor readable pixel matrix which are considered valid for image acquisition purposes. This property describes an arbitrary number of overlapping rectangles, with each rectangle representing the maximum image size that the camera sensor can produce for a particular aspect ratio. They are defined relatively to the PixelArraySize rectangle. When multiple rectangles are reported, they shall be ordered from the tallest to the shortest. Example 1 A camera sensor which only produces images in the 4:3 image resolution will report a single PixelArrayActiveAreas rectangle, from which all other image formats are obtained by either cropping the field-of-view and/or applying pixel sub-sampling techniques such as pixel skipping or binning. PixelArraySize.width /----------------/ x1 x2 (0,0)-> +-o------------o-+ / y1 o +------------+ | | | |////////////| | | | |////////////| | | PixelArraySize.height | |////////////| | | y2 o +------------+ | | +----------------+ / The property reports a single rectangle PixelArrayActiveAreas = (x1, y1, x2 - x1 + 1, y2 - y1 + 1) Example 2 A camera sensor which can produce images in different native resolutions will report several overlapping rectangles, one for each natively supported resolution. PixelArraySize.width /------------------/ x1 x2 x3 x4 (0,0)-> +o---o------o---o+ / y1 o +------+ | | | |//////| | | y2 o+---+------+---+| | ||///|//////|///|| | PixelArraySize.height y3 o+---+------+---+| | | |//////| | | y4 o +------+ | | +----+------+----+ / The property reports two rectangles PixelArrayActiveAreas = ((x2, y1, x3 - x2 + 1, y4 - y1 + 1), (x1, y2, x4 - x1 + 1, y3 - y2 + 1)) The first rectangle describes the maximum field-of-view of all image formats in the 4:3 resolutions, while the second one describes the maximum field of view for all image formats in the 16:9 resolutions. Multiple rectangles shall only be reported when the sensor can't capture the pixels in the corner regions. If all the pixels in the (x1,y1) - (x4,y4) area can be captured, the PixelArrayActiveAreas property shall contains the single rectangle (x1,y1) - (x4,y4). \todo Rename this property to ActiveAreas once we will have property categories (i.e. Properties::PixelArray::ActiveAreas) - ScalerCropMaximum: type: Rectangle description: | The maximum valid rectangle for the controls::ScalerCrop control. This reflects the minimum mandatory cropping applied in the camera sensor and the rest of the pipeline. Just as the ScalerCrop control, it defines a rectangle taken from the sensor's active pixel array. This property is valid only after the camera has been successfully configured and its value may change whenever a new configuration is applied. \todo Turn this property into a "maximum control value" for the ScalerCrop control once "dynamic" controls have been implemented. # ---------------------------------------------------------------------------- # Draft properties section - ColorFilterArrangement: type: int32_t draft: true description: | The arrangement of color filters on sensor; represents the colors in the top-left 2x2 section of the sensor, in reading order. Currently identical to ANDROID_SENSOR_INFO_COLOR_FILTER_ARRANGEMENT. enum: - name: RGGB value: 0 description: RGGB Bayer pattern - name: GRBG value: 1 description: GRBG Bayer pattern - name: GBRG value: 2 description: GBRG Bayer pattern - name: BGGR value: 3 description: BGGR Bayer pattern - name: RGB value: 4 description: | Sensor is not Bayer; output has 3 16-bit values for each pixel, instead of just 1 16-bit value per pixel. - name: MONO value: 5 description: | Sensor is not Bayer; output consists of a single colour channel. ...