Electronic Imaging 2026 (EI2026)

Where Industry and Academia Meet to Advance Imaging
Join us in person for EI 2026 at the Hyatt Regency San Francisco Airport in Burlingame, California!
EI 2026 offers:
- Exciting Symposium Plenaries and Conference Keynotes.
- A rich technical program of oral talks, interactive poster papers, a demonstration session, and multiple opportunities to network.
- An industry exhibit and conference lunches.
- A robust short course program.
- A dynamic environment to interact with colleagues from industry and academia across the globe.
Electronic Imaging 2026 brings together multiple technical conferences covering all aspects of imaging.
Learn More Registration Coming Soon
Imatest is presenting:
- Exhibition
- Short Course: Information Metrics for Machine Vision
- Paper: Image Sensor Noise model for Image System Simulation
- Paper: Information-based Dynamic Range
- Paper: A method for calculating NIR bandpass-adjusted Optical Densities for better matching common standard test chart specifications.
- Paper: Toward Fair and Accurate Camera Testing: Validation of Skin Tone Test Charts with Real Human Data
Exhibition
The Exhibition will be held on Tuesday March 3rd and Wednesday March 5th. Stop by our booth to chat with our imaging scientists.
Short Course: Information Metrics for Machine Vision
Instructor: Norman L. Koren
When: Afternoon of Sunday, March 1st
Course description:
We introduce the basics of information theory, show why the key information metrics (information capacity and SNRi) are superior to traditional metrics such as sharpness (SFR) and noise for characterizing camera performance, then show how to calculate them from test chart images. Next, we describe the camera performance simulator, including
- creating input images (usually test targets),
- simulating lens degradations,
- the image sensor noise model,
- ISP (Image Signal Processing),
- results, including the standard and new information metrics.
“Image Sensor Noise model for Image System Simulation”
Norman L. Koren
Session: TBD
Date: TBD
Time: TBD (Pacific Time)
Duration: 20 minutes
We present an image sensor noise model, which is part of a complete image system simulation that includes image generation, lens degradations, and ISP (Image Signal Processing), and can produce classic measurements (SFR, noise, etc.) as well as the new information metrics (information capacity, SNRi, etc.).
The noise model is derived from a classic Photon Transfer Curve (PTC) obtained from one or at most two raw (undemosaiced) images of a high dynamic range grayscale test chart. Image sensor noise is composed of three factors.
“Information-based Dynamic Range”
Norman L. Koren
Session: TBD
Date: TBD
Time: TBD PM (Pacific Time)
Duration: 20 minutes
We present a new approach to measuring camera dynamic range and low-light performance based on C4 information capacity, which is measured directly from ISO 12233-standard 4:1 contrast slanted edges. Our initial technique involves photographing a test chart that contains 4:1 slanted edges over an extremely wide range of exposures, from ½ or ¼ second (where the brighter side of the edge saturates) to 1/2000 or 1/4000 second, where the image appears nearly black, but a noisy edge is still present. The major advantages of this method are
- Dynamic range limits are based on an actual performance metric (C4) rather than SNR, which is only one of the factors that contributes to camera performance.
- C4 correctly handles performance degradation due to stray light.
We will discuss new techniques, still under development, for facilitating the measurement.
“A method for calculating NIR bandpass-adjusted Optical Densities for better matching common standard test chart specifications.”
Christian Taylor, Amelia Limbocker
Session:
Date: TBD
Time: TBD (Pacific Time)
Duration: 20 minutes
Near-infrared (NIR) imaging is now prevalent in machine vision, automotive, and biomedical applications, but most step-chart definitions were created for visible imaging. Many standards assume visible lighting conditions and only consider IR-blocking, so NIR-sensitive and RGB+NIR cameras are not adequately addressed. This leads to charts whose nominal densities don’t produce results as intended.

Screenshot
“Toward Fair and Accurate Camera Testing: Validation of Skin Tone Test Charts with Real Human Data”
Megan Borek, Amelia Limbocker, Ellis Monk
Session: Skin Tone Capture and Image Quality I
Date: TBD
Time: TBD PM (Pacific Time)
Duration: 20 minutes
Accurate reproduction of diverse skin tones remains a persistent challenge for consumer cameras, with shortcomings in automatic exposure and white balance often leading to biased results. To address this, we are developing skin tone test charts designed to better represent real human skin across the full range of the Monk Skin Tone Scale. This work validates these charts using real human data collected through multiple methods. First, survey responses capture user perspectives on how their skin tones are represented in smartphone images and what they expect from their devices. Second, controlled photographic sessions with participants across the scale provide image data under varied lighting conditions and multiple exposure strategies, including gray card, spot metering, and bracketing, alongside participant-selected preferred
measurements from skin regions supplement the imaging data with objective ground truth. Together, these datasets allow us to evaluate and tune test chart behavior so that color and exposure responses align more closely with real skin. This validation represents a step toward fairer, more accurate camera testing standards, with implication for consumer photography as well as applications in medical, automotive, and security images.
