Developing a machine vision or embedded AI systems and scaling them for deployment are already challenging tasks. Converging these into an integrated Embedded AI Vision System can keep scores of developers and engineers busy for quite some time. The selection of relevant sub-components has a significant impact on the engineering challenge and the final solution’s complexity, performance, physical size, and power consumption. Regarding the optical path of an Embedded AI Vision System, making the right lens selection is a critical milestone and should not be seen as an afterthought since the lens’ performance and characteristics impact almost every aspect of the downstream vision chain. The right lens choice can positively impact the required AI algorithm complexity and final system performance, whereas the wrong lens choice can hamstring subsequent development.
If applied AI for embedded vision aims to replicate human sensing and understanding, then the optical stack plays a significant role in achieving a human-like vision for any camera-based application. However, carefully selecting the right lens will pay huge dividends whether the intended application is in automotive, security, medical, robotics, or any other possible implementations when transitioning from the lab to the real world. For systems that will be deployed in an environment with low or changing light or are exposed to a wide temperature range, the challenges to delivering consistent optical performance increase dramatically. Using AI algorithms, a lens must excel along multiple vectors to support real-time or post-processing.
Choosing the right lens is a critical step in system-level optimization and setting a roadmap to achieving desired outcomes. From our experiences across industries, applications, and clients, we found the following 1st order lens parameters to be critically important:
- MTF – Modulation Transfer Function is a standardized way to describe an optical system regarding sharpness and contrast and is a key performance indicator when comparing lenses. While commonly used to judge “how good” a lens is, it is not the only, and sometimes not even the best metric for determining what will work best in a given use-case.
- F/# – Essentially, the amount of light the lens lets in for a given focal length. The physical aperture or “Iris” of an optical system defines the total light throughput and directly impacts the lens’ ability to produce the required contrast and Depth of Field.
- Distortion – The concept of distortion describes how a lens maps a shape in object space to the image plane. Distortion must be referenced to something, which is sometimes called the “projection” or may be referred to as Rectilinear, F-tan, or F-theta distortion.
- Relative Illumination (RI) – This is a normalized %-value that represents the illumination of any field point relative to the point of maximum illumination, which is typically on-axis. A high RI value means “flat” illumination across the image plane, whereas a low RI may introduce dark corners or edges.
- Dynamic Range – Quantifies a system’s ability to adequately image high lights and dark shadows in a scene, IE, a wide range of lighting values or conditions in the same image without being saturated or too dark.
As listed previously, MTF, or the Module Transfer Function, is a standardized way of comparing the optical performance of different imaging systems. For AI algorithm-based systems, selecting lenses that have a fairly consistent MTF across the FOV (Field of View) and show a stable MTF behavior over temperature is generally advisable. The intended wavelength spectrum, which depending on the application, typically include visible (VIS) and/or near-infrared (IR), also has to be taken into consideration when comparing MTF performance. A fully athermalized lens called hyperspectral, RGB-IR, or Day/Night lens is a lens that checks off all these boxes. Such a lens requires not only a deep design experience but also an in-depth understanding of material properties, advanced coating Know-How, and a manufacturing process expertise that has been optimized and advanced over decades.
Not all applications require such an advanced optical system, and it will be for the application domain expert to decide which component and how far to optimize the system. However, everything that can be optimized at the lens level “at the speed of light” reduces processing time and power later, possibly resulting in smaller and more energy-efficient computing systems. It is also much easier to start with good lens performance rather than recreate “missing” data to compensate for poor optical performance since it is almost impossible to process enough to make up for lost data at the lens level (the often-cited concept of garbage in, garbage out applies here as well.)
Other optical performance parameters, such as the F/# and Relative Illumination (RI), also contribute to the consistency of imaging quality that any AI algorithm has to deal with. Unfortunately, there is no “one size fits all.” For example, the system architect has to decide whether to optimize a system for low-light performance with a low F/# or improve the Depth of Field (DoF), which defines the range from the near to the far object distance that is determined to be in focus, with a higher F/#. In this example, a high RI may contribute to a well-balanced system where the edge illumination would tolerate some “stopping down” (increasing the F/#) to increase DoF. Lens selection can quickly become a multi-dimensional optimization with various trade-offs.
Distortion is one of the lens specs, that software developers might often like to simply make go-away since a high and consistent number of pixels per degree (px/deg) across the entire FOV would be preferred. But since distortion is unavoidable in many cases, Sunex has found ways to manipulate a lens’ distortion profile to align best with and support the algorithm-specific requirements. Our Tailored Distortion™ expertise has often been applied to SuperFisheye™ lenses to correct barrel distortion of large FOV lenses, providing more px/deg at the field edges, while conversely, the initial statement of striving to achieve a human-like vision has led to the development of FOVEA distortion lenses that mimic the human eye by increasing the px/deg density in the center while at the same time maintaining a wide field of view peripheral vision.
Not just the 1st order lens design parameters but also a combination of lens design considerations, coating choices, and surface treatment, contribute to the dynamic range of a lens, which can significantly affect the performance in extreme light situations. The dynamic range of a lens or system is defined as the ratio of the largest non-saturating input signal to the smallest detectable input signal. AI systems running on mobile autonomous systems, such as a delivery vehicle driving from broad daylight into a tunnel (or out), is an authentic example that benefits from a lens with an excellent dynamic range to support consistent results. An agriculture harvesting machine that, on the return leg, drives into the setting sun or stationary systems such as exterior smart infrastructure or security cameras that deal with passing light sources such as vehicles or the sun are further applications in need of a lens with high dynamic range. Many customers ask what dynamic range really means for the lens since it is the imager (CMOS, CCD, etc.) detecting the light, not the lens. The answer is that the lens must be very good regarding stray light, glare, and ghosting. These image artifacts can be ignored for a low dynamic range sensor because the sensor will not pick them up. However, in an HDR/WDR sensor, at best, stray light reduces the image’s signal-to-noise level (contrast). At worst, the artifacts become apparent and may be interpreted as another light source, such as an oncoming vehicle, or could obscure real images.
Who knows where the capabilities of embedded AI-vision systems will lead us in the future; can it ignore dust or dirt on the camera, overcome insufficient lighting, or potentially compensate for MTF changes due to thermal shift? While AI may eventually allow for more relaxed lens requirements like a human brain does, there is still no substitute for good image quality. We can clearly see the benefits and potential that an “AI-optimized” lens can deliver to the overall system reliability and consistent image quality, primarily if the AI-based vision system is situated in varying environmental conditions. Needless to say, size constraints and piece price also factor in when designing the right system, and there is no need to “over-engineer” the final solution. However, the previously mentioned factors that drive image quality, when done right, have the potential to simplify the AI system complexity, optimize algorithms, and reduce system latencies and power consumption while enhancing imaging performance.
Sunex already has many lens designs that combine low F/#, high Relative Illumination (RI), high dynamic range (HDR), high MTF across the field, and a broad wavelength spectrum for consistent performance for many industries and applications. Often the initial engagement with our clients is to select possible options based on our existing portfolio of lenses to get real-world feedback on what optical performance is required in the real use case. Based on this feedback, we review opportunities to optimize an adapt and existing lens or may jointly decide to pursue a purpose-built, custom lens design that matches the application and requirements. For many of our clients, we also offer greater vertical integration by designing and manufacturing the entire sensor board, including fully automated active lens/sensor alignment.
With a 25-year track record as a global optics company, we are leveraging best-in-class design expertise and constantly driving innovation and process improvements, leading to millions of lenses delivered worldwide on time. The goal for Sunex, our clients, and all parties involved is to collaborate to develop a system that can meet the expected performance, for an acceptable price, in the time frame needed.
This article was also published in a similar form by Wiley Industries in their Inspect international 02 2022 issue.