Conference PaperPDF Available

Recent advances in high dynamic range imaging technology

Authors:

Abstract

Recently, visual representations using high dynamic range (HDR) images become increasingly popular, with advancement of technologies for increasing the dynamic range of image. HDR image is expected to be used in wide-ranging applications such as digital cinema, digital photography and next generation broadcast, because of its high quality and its powerful expression ability. HDR imaging technologies will spread its sphere of influence in imaging industry. In this paper, we review the state-of-the-art studies and the trends of the HDR imaging, in terms of the following three points: (1) HDR imaging sensor and HDR image generation techniques as image acquisition technologies, (2) encode method of HDR images for efficient transmission and storage, (3) human visual system issues associated with reproduction of HDR image.
RECENT ADVANCES IN HIGH DYNAMIC RANGE IMAGING TECHNOLOGY
Yukihiro BANDOH, Guoping QIU, Masahiro OKUDA††, Scott DALY‡‡, Til Aach††† and Oscar C. AU‡‡‡
NTT Cyber Space Laboratories, NTT Corporation
The University of Nottingham, School of Computer Science
†† University of Kitakyushu
‡‡ Sharp Laboratories of America
††† Institute of Imaging and Computer Vision, RWTH Aachen University
‡‡‡ Hong Kong University of Science and Technology
ABSTRACT
Recently, visual representations using high dynamic range
(HDR) images become increasingly popular, with advance-
ment of technologies for increasing the dynamic range of
image. HDR image is expected to be used in wide-ranging
applications such as digital cinema, digital photography and
next generation broadcast, because of its high quality and its
powerful expression ability. HDR imaging technologies will
spread its sphere of influence in imaging industry. In this
paper, we review the state-of-the-art studies and the trends
of the HDR imaging, in terms of the following three points:
(1) HDR imaging sensor and HDR image generation tech-
niques as image acquisition technologies, (2) encode method
of HDR images for efficient transmission and storage, (3)
human visual system issues associated with reproduction of
HDR image.
Index Termshigh dynamic range image, high bit-depth
image, multiple exposure principle, human visual system,
tone mapping
1. INTRODUCTION
Realistic representations using high quality images are be-
coming increasingly popular. The realistic representations
demand the following four elements: high spatial resolu-
tion, high temporal resolution, reproducing accurate color,
and large dynamic range. For example, digital cinema offers
digital images with high-resolution. In order to represent a
smooth movement, an high-speed HDTV camera that can
shoot at 300 [frames/sec] is developed. The advanced ef-
forts to reproduce accurate color are made. High dynamic
range imaging (HDRI) is a new imaging technology that has
emerged in recent years and it has the promise of bringing a
new revolution to digital imaging [1].
The real world scenes humans experience every day have
far higher luminance dynamic ranges. For instance, a scene
showing both shadows and sunlit areas will have a dynamic
range exceeding 100,000:1. Human visual system is capa-
ble of perceiving light intensities over a range of 4 orders
of magnitude, and with adaptation, its sensitivity can stretch
to 10 orders of magnitude. However, conventional computer
monitors and other reproduction media such as printing pa-
pers have limited dynamic ranges, often less than 2 orders
of magnitude. On the image capture side a similar argument
can be made. Most cameras limit their outputs to eight bits
per colour channel. Therefore, it is clear that the dynamic
range of current display, camera and image file formats are
not enough to represent real scenes and only record a fraction
of the contrast that humans are capable of perceiving.
In HDRI, the image files record the actual colour and
dynamic range of the original scene rather than the limited
gamut and dynamic range of the monitor or other reproduc-
tion media. This means that image processing, manipulation,
display and other operations will no longer be limited by the
number of bits used to represent each pixel. Thus, HDRI will
have widespread applications in digital cinema, digital pho-
tography, computer games, etc., and will open up many new
possibilities, including dramatically improving the visual re-
alism of digital photographs and videos, enabling the devel-
opment of more accurate computational vision techniques,
etc. In the near future, the imaging industry will inevitably
move to HDRI which will affect all components of the digital
imaging pipeline including capturing (sensor, camera), stor-
age (compression, coding) and reproduction (rendering, tone-
mapping, printing and display).
In this survey paper, we discuss technical issues, recent
development and future directions of this very promising and
exciting imaging technology. This paper is composed as fol-
lows. Section 2explains the technologies for acquisition and
generation of HDRI. Then, section 3treats with reproducing
technologies of HDRI. Section 4shows HDRI coding which
is needed for efficient storage and transmission.
2. ACQUISITION AND GENERATION OF HDR
IMAGE
2.1. Recent developments on HDRI sensors
The dynamic range of a camera CCD or CMOS sensor is de-
fined as the ratio of the full well capacity (FWC), i.e. the
maximum measurable signal, and the root-mean-square dark
noise, i.e. the lowest signal differentiable from the noise floor
(NF). Sensor manufacturers can increase the dynamic range
by either decreasing the NF or increasing the FWC and thus
the saturation level. An overview of the different solutions to
increase the sensor’s dynamic range is given in [2] and has
later been extended in [3].
Solutions to increase the dynamic range by changing from
a linear to a logarithmic response for the photon-electron con-
version have been suggested [4]. These sensors, however suf-
fer from increased fixed-pattern-noise (FPN) which becomes
most critical for low light conditions. Therefore combined
log-linear sensors [5] adapt the response curve to the lighting
conditions on a per pixel basis. Both methods reduce, how-
ever, the resolution of detectable light changes due to their
logarithmic responses.
Instead of increasing the FWC the sensor could as well
provide the time to saturation (TTS) as the signal correspond-
ing to the incident light flux. Sensors following this TTS ap-
proach have been presented in [6]. Other approaches deal
with pixel specific adaptive integration time (AIP) [7], which
has been combined with the TTS approach to improve the
low-light behaviour [8]. Furthermore it has been investigated
to include the concept of multiple exposures on the chip in
form of a Bayer-pattern like spatially varying neutral density
filter [9] as well as continually varying filters [10].
2.2. HDR image generation based on multiple exposure
principle
In the past decade, it had been widely agreed in the CG com-
munity that the dynamic range of the traditional imaging is
inadequate for Image-based lighting (IBL)[11], [12]. IBL is
the rendering process of illuminating objects with images of
light taken from the real world. IBL can provide realistic ap-
pearances when the image has a high dynamic range and is
radiometrically calibrated. Debevec first introduced a method
to acquire an omnidirectional HDR image called Light Probe
in [13] and applied it to IBL. Since then, the HDR image has
been widely used and is now available in many graphics pro-
cessors.
Several methods have been proposed to improve the dy-
namic range of general photographs based on a multiple ex-
posure principle. Mann et al. first attempted to construct the
HDR [14]. This algorithm is composed of the four steps: (1)
Photographs of still objects are taken off line by a camera at a
fixed location, (2) A camera response curve is estimated from
the multiple exposure set by self calibration, (3) Linearize the
images by applying the inverse of the response curve, and (4)
Merge the linearized images. These four steps is a basic pro-
cedure for the HDR acquisition and many of conventional al-
gorithms follow it. There are several variations on the camera
response curve estimation. While Mann et al. approximates
the response curve by a simple gamma function [14], Debevec
et al. [13] describes the curve by a set of exposure values for
more precise approximation. Mitsunaga et al. express the
curve by a low order polynomial [15], which provides a more
flexible radiometric model.
These techniques based on the multiple exposures have a
disadvantage. It is assumed that a scene is completely still
during taking photographs. Therefore if there is any motion
of objects or camera shake, ghosting artifacts will appear af-
ter combining the images. Motion compensation is a classical
problem and there exist many methods for optical flow esti-
mation and image registration. Ward [16] presented a method
to align the images and applied it to multiple exposure fusion.
Kang et al. [17] used gradient-based optical flow estimation
to remove the ghosting artifacts of the HDR video.
3. REPRODUCTION OF HDR IMAGE
3.1. Perceptual HDR image quality
Systems for HDR can be dissected into image capture, im-
age/video path (including compression) and displays. The
displays are most intimately related to the human visual sys-
tem (HVS) since there are no intervening unknown elements
between the display and the viewer. Since CRTs and film
have been able to achieve approximately 3 log units of dy-
namic range for many decades, that range forms a convenient
distinction between standard dynamic range (SDR) displays
and HDR displays. An early HDR display by combining 2
digital film images was developed by Greg Ward [18]. This
idea was extended to video by projecting a digital image as
an LCD backlight [19]. Using such equipment, several psy-
chophysical studies began assessing the advantages of HDR
displays for preferences, such as comparing SDR and HDR
imagery [20], comparing HDR images against the real-world
[21], and comparing preferred tonescales for mapping SDR
to HDR, [22].
In addition, the advantages in terms of functionality have
been studied, such as for medical images [23]. Some stud-
ies have been done casting doubt on the ability of the visual
system to be able to distinguish HDR from SDR images due
to optical flare, at least for static images [24], [25]. Regard-
ing video processing and tonescale manipulation, algorithms
have been designed to take into the account the extra range af-
forded by specular highlights and tested with observers [26],
or used HVS models in their design, such as spatial frequency
channels [27].
3.2. Tone Mapping
One of the problems of high dynamic range imaging is the
display of high dynamic range radiance maps on conventional
reproduction media such as LCD panels. One solution to this
problem is to compress the dynamic range of the radiance
maps such that the mapped image can be fitted into the dy-
namic range of the display devices. This mapping is called
tone mapping [28]. Several tone mapping methods have ap-
peared in the literature in recent years.
These methods can be divided into two broad categories.
The global tone mapping techniques [29, 30, 31] using a sin-
gle appropriately designed spatially invariant mapping func-
tion for all pixels in the image; the local mapping techniques
[32, 33, 34] adapt the mapping functions to local pixel statis-
tics and local pixel contexts. Global tone mapping is simpler
to implement but tend to lose details. Local tone mapping
is much more computationally intensive and harder to get it
right since there are often a number of parameters in the al-
gorithms which have to be set empirically. Given the nature
of the problem, it is not possible to have one method fits all,
i.e., it may not possible to have one method that will solve the
problem once and for all. What are needed are multiple meth-
ods and depending on the particular requirements of the users,
one method will be better suited than others, or a combination
of methods will be necessary.
4. EFFICIENT CODING METHOD FOR HDR IMAGE
Increase in dynamic range requires increase in bit depth in or-
der to represent natural quality with smooth gradation. Since
it leads to the amount of image data, we need the efficient en-
coding algorithm. As the studies for video coding with high
bit depth over 10 [bits/channel], layered scalable coding ap-
proaches are considered. These approaches feature to offer a
bit-depth scalability that can support multi bit-depth, and the
compatibility of the base layer with JPEG[35], MPEG-4[36]
and AVC/H.264[37]. Mantiuk et al. study a high dynamic
range video encoding that optimizes luminance quantization
based on the contrast threshold perception in the human vi-
sual system [38]. Ito et al. study a encoder design for high bit
depth sequences based on the optimization of tone mapping
curve [39]. This method features to design a tone mapping
curve that can minimize bit-depth transform error.
Recent inter-national standards of image/video coding
also support HDRI through increasing processable bit-depth.
AVC/H.264 includes three profiles (High 4:4:4 predictive,
High 4:4:4 intra and CAVLC 4:4:4 intra) [40] that support bit
depth up to 14 [bits/channel] and 4:4:4 color format, though
conventional video codecs mainly consider sequences with
4:2:0 format of 8 [bits/channel]. JPEG2000 supports bit depth
up to 12 [bits/channel] and 4:4:4 color format. Furthermore,
JPEG-XR can handle bit depth up to 32 [bits/channel] and
4:4:4 color format. AVC/H.264 also offer a SIE message [41]
that transmits information for describing tone mapping oper-
ation which adjusts the bit depth of each pixel values in the
reconstructed images. Using this SIE message, the decoder
of AVC/H.264 can display is given a post-processing func-
tionality that can map a higher bit depth sequences images to
a lower bit depth display.
5. CONCLUSION
Research in high dynamic range imaging has started more
than a decade ago, and it is a relatively young field. Already,
we have seen examples of dramatic visual quality improve-
ments that HDR image can achieve over traditional low dy-
namic range image. Currently, this new technology is still
in its early stage of development. In order to achieve its full
potential, for example, to implement HDR imaging technol-
ogy in consumer level digital camera to enable ordinary users
to take high quality photographs and videos under any light-
ing conditions, there are still many technical hurdles to over-
come, including, image sensing, coding and storage and dis-
play. This paper briefly surveyed what have been achieved in
HDR imaging which also highlighted that much needs to be
done in order to make this promising imaging technology the
mainstay of digital imaging.
6. REFERENCES
[1] G. Ward E. Reinhard, S.Pattanaik and P. Debevec, High Dy-
namic Range Imaging: Acquisition, Display, and Image-Based
Lighting, Morgan Kaufmann Publisher, 2005.
[2] O. Yadid-Pecht, “Wide-dynamic-range sensors, Optical En-
gineering, vol. 38, no. 10, pp. 1650–16602007, 1999.
[3] A. Spivak, A. Belenky, A. Fish, and O. Yadid-Pecht, “Wide-
dynamic-range CMOS image sensors—comparative perfor-
mance analysis,” IEEE Trans. on Electron Devices, vol. 56,
no. 11, pp. 2446–2461, Nov. 2009.
[4] H.-Y. Cheng, B. Choubey, and S. Collins, “A high-dynamic-
range integrating pixel with an adaptive logarithmic response,
IEEE Photonics Technology Letters, vol. 19, no. 15, pp. 1169–
1171, Aug. 2007.
[5] J. Guo and S. Sonkusale, “A high dynamic range CMOS im-
age sensor for scientific imaging applications,” IEEE Sensors
Journal, vol. 9, no. 10, pp. 1209–1218, Oct. 2009.
[6] X. Guo, X. Qi, and J. G. Harris, “A time-to-first-spike CMOS
image sensor, IEEE Sensors Journal, vol. 7, no. 8, pp. 1165–
1175, Aug. 2007.
[7] P. M. Acosta-Serafini, I. Masaki, and C.G.Sodini, “A 1/3”
VGA linear wide dynamic range CMOS image sensor imple-
menting a predictive multiple sampling algorithm with over-
lapping integration intervals, IEEE Journal of Solid-State Cir-
cuits, vol. 39, no. 9, pp. 1487–1496, Sept. 2004.
[8] D. Stoppa, M. Vatteroni, D. Covi, A. Baschirotto, A. Sartori,
and A. Simoni, A 120-dB dynamic range CMOS image sen-
sor with programmable power responsivity,” IEEE Journal of
Solid-State Circuits, vol. 42, no. 7, pp. 1555–1563, July 2007.
[9] S. K. Nayar and T. Mitsunaga, “High dynamic range imaging:
spatially varying pixel exposures, in Proc. IEEE Conf. on
Computer Vision and Pattern Recognition, 2000, vol. 1, pp.
472–479 vol.1.
[10] S. K .Nayar and V. Branzoi, “Adaptive dynamic range imaging:
optical control of pixel exposures over space and time, in
Proc. IEEE Int. Conf. on Computer Vision, Oct. 2003, vol. 1,
pp. 1168–1175.
[11] G. Ward, “Real pixels,” in Graphic Gems II, pp. 80–83. Cam-
bridge University Press, 1991.
[12] P. Debevec, “Image-based lighting,” IEEE Computer Graphics
and Applications, vol. March/April, pp. 26–34, 2002.
[13] P. E. Debevec and J. Malik, “Recovering high dynamic range
radiance maps from photographs,” Proc. of SIGGRAPH 97,
Computer Graphics Proc., pp. 369–378, 1997.
[14] S. Mann and R. Picard, “On being ’undigital’ with digital
cameras: Extending dynamic range by combining differently
exposed pictures,” IS&T’s Annual Conf.; Society for Imaging
Science and Technology, vol. 56, pp. 2446–2461, May 1995.
[15] T. Mitsunaga and S. K. Nayer, “Radiometric self calibration,”
IEEE Conf. on Computer Vision and Pattern Recognition, vol.
1, pp. 374–380, Jun. 1999.
[16] G. Ward, “Fast, robust image registration for compositing high
dynamic range photographs from hand-held exposures, Jour-
nal of Graphics Tools, vol. 8, pp. 17–30, 2003.
[17] S. Winder S. Kang, M. Uyttendaele and R. Szeliski, “High
dynamic range video,” Trans. on Graphics 2003, pp. 319–325,
2003.
[18] G .Ward, “A wide field, high dynamic range, stereographic
viewer,Proc. of PICS 2002, April 2002.
[19] H. Seetzen, W. Heidrich, W. Stuerzlinger G. Ward, L. White-
head, M. Trentacoste, A. Ghosh, and A. Vorozcovs, “High
dynamic range display systems,” ACM SIGGRAPH, 2004.
[20] P. Ledda, A. Chalmers, T. Troscianko, and H. Seetzen, “Evalu-
ation of tone mapping operators using a high dynamic range
display, in Proc. of ACM SIGGRAPH/EUROGRAPHICS
Symp. on Graphics hardware, 2007, vol. 24.
[21] A. Yoshida, V. Blanz, K. Myszkowski, and H. Seidel, “Per-
ceptual evaluation of tone mapping operators with real-world
scenes,” in Proc. of Human Vision and Electronic Imaging X,
IS&T/SPIE Annual Symp. Electronic Imaging, 2005, pp. 192 –
203.
[22] A. Aykuz, R. Fleming, B. Rieche, E. Rheinhard, and
H. Bultoff, “Do HDR displays support LDR content ? A psy-
chophysical evaluation, Trans.of the ACM, 2007.
[23] A. Badano, G. Guarnieri, G. Ramponi, and L. Albani, “Quanti-
zation in medical imaging displays: initial observer results for
a high-luminance range dual-layer LCD,” Proc. SID, vol. 39,
pp. 923–926, 2009.
[24] A. Rizzi and J. McCann, “Glare-limited appearances in HDR
images,” Journal of the Society for Information Display, vol.
17, no. 1, pp. 3–12, 2009.
[25] J. McCann and A. Rizzi, “Retinal HDR images: Intraocular
glare and object size,” Journal of the Society for Information
Display, vol. 17, no. 11, pp. 913–920, 2009.
[26] L. Meylan, S. Daly, and S. S¨
usstrunk, “The reproduction
of specular highlights on high dynamic range displays,” in
IS&T/SID Color Imaging Conf. (CIC), 2006.
[27] R. Mantiuk, S. Daly, and L. Kerofsky, “Display adaptive tone
mapping,” ACM Trans. on Graphics (Proc. of SIGGRAPH),
vol. 27, no. 3, 2008.
[28] J. DiCarlo and B. Wandell, “Rendering high dynamic range
images,” in Proc. SPIE, 2001, vol. Proc. SPIE, pp. 392–401.
[29] G. W. Larson, H. Rushmeier, and C. Piatko, “A visibility
matching tone reproduction operator for high dynamic range
scenes,” IEEE Trans. on Visualization and Computer Graph-
ics, pp. 291–306, 1997.
[30] G. Qiu, J. Guan, J. Duan, and M. Chen, “Tone mapping for
HDR image using optimization - a new closed form solution,
in Int. Conf. on Pattern Recognition, 2006, vol. 1, pp. 996–999.
[31] J. Duan, M. Bressan, C. Dance, and G. Qiu, “Tone-mapping
high dynamic range images by novel histogram adjustment,
Pattern Recognition, vol. 43, no. 5, pp. 1847–1862, May 2010.
[32] F. Durand and J. Dorsey, “Fast bilateral filtering for the display
of high-dynamic- range images,” ACM Trans. on Graphics, pp.
257–266, 2002.
[33] E. Rheinhard, M. Stark, P. Shirley, and J. Ferwerda, “Pho-
tographic tone reproduction for digital images,” ACM SIG-
GRAPH, 2002.
[34] R. Fattal, D. Lischinski, and M. Werman, “Gradient domain
high dynamic range compression,” ACM Trans. on Graphics,
pp. 249–256, 2002.
[35] G. Ward and M. Simmons, “JPEG-HDR: A backwardscom-
patible, high dynamic range extension to JPEG,” in IEEE Int.
Conf. on Image Processing, 2005, pp. 283–290.
[36] R. Mantiuk, A. Efremov, K. Myszkowski, and H. Seidel,
“Backward compatible high dynamic range mpeg video com-
pression,” ACM Trans. Graph., 2006.
[37] M. Okuda and N. Adami, “Two-layer coding algorithm for
high dynamic range images based on luminance compensa-
tion,” Journal of Visual Communication and Image Represen-
tation, vol. 18, no. 5, pp. 377–386, Oct. 2007.
[38] R. Mantiuk, G. Krawczyk, K. Myszkowski, and H. Seidel,
“Perception-motivated high dynamic range video encoding,
ACM Trans. Graph., vol. 23, no. 3, pp. 733–741, 2004.
[39] T. Ito, Y. Bandoh, S.Takamura, and H. Jozawa, “Coding algo-
rithm for high bit depth video using bit depth transform,” in
IEEE Int. Conf. on Image Processing, in this special session,
2010.
[40] ISO/IEC SC 29/WG 11 (MPEG) 14496-10:2004, Coding of
audio visual objects – Part 10: Advanced Video Coding 3rd
Ed., Int. Standard, Nov. 2007.
[41] A. Segall, A. Kerofsky, and S. Lei, “Tone mapping SEI mes-
sage,” JVT-T060. Klagenfurt, Austria, July 2007.
... Conventional imaging sensors, such as CCD or CMOS, pose limitations in capturing the full range of visual intensity information [1]. Typically, these sensors quantize intensity levels to 8 bits per color channel, resulting in a maximum of 256 levels. ...
... Sensor manufacturers face two options to enhance the dynamic range: reducing NF or increasing the FWC. However, most conventional CCD or CMOS sensors have a fixed dynamic range, which limits their ability to capture the full range of light intensities [1]. ...
... image. The dynamic range of a CCD or CMOS sensor is the ratio of the maximum measurable signal and the noise floor [90]. Human visual systems have a very high-luminance dynamic range, thus allowing humans to perceive a great degree of contrast in a real-life scene. ...
Article
Full-text available
Huge advancements have been made over the years in terms of modern image-sensing hardware and visual computing algorithms (e.g., computer vision, image processing, and computational photography). However, to this day, there still exists a current gap between the hardware and software design in an imaging system, which silos one research domain from another. Bridging this gap is the key to unlocking new visual computing capabilities for end applications in commercial photography, industrial inspection, and robotics. In this survey, we explore existing works in the literature that can be leveraged to replace conventional hardware components in an imaging system with software for enhanced reconfigurability. As a result, the user can program the image sensor in a way best suited to the end application. We refer to this as software-defined imaging (SDI), where image sensor behavior can be altered by the system software depending on the user’s needs. The scope of our survey covers imaging systems for single-image capture, multi-image, and burst photography, as well as video. We review works related to the sensor primitives, image signal processor (ISP) pipeline, computer architecture, and operating system elements of the SDI stack. Finally, we outline the infrastructure and resources for SDI systems, and we also discuss possible future research directions for the field.
... The gain and gamma corrections can adjust the overall brightness and darkness uniformly, but they cannot make adjustments when the bright and dark areas are mixed in the same image [26]. Artevo 800® has two integrated 4 K video cameras which are not High Dynamic Range (HDR) video cameras, so the video cameras cannot improve the dynamic range [27,28]. If the brightness is calculated and set by referring to the histogram of luminance values in a small surrounding range for each pixel unit, the dynamic range can be optimized up to all corners of the screen. ...
Article
Full-text available
Abstract Background Image-sharpening algorithms with color adjustments enable real-time processing of the surgical field with a delay of 4 msec for heads-up surgery using digital three-dimensional displays. The aim of this study was to investigate the usefulness of the algorithms with the Artevo 800® digital microscope. Methods Seven vitreoretinal surgeons evaluated the effects of image-sharpening processing on the clarity of the surgical field with the Artevo 800® system that is used for cataract and vitreous surgeries. The scorings were made on a 10-point scale for anterior capsulotomy, phacoemulsification, cortex aspiration, core vitrectomy, and peeling of an epiretinal membrane or an internal limiting membrane. In addition, the images during the internal limiting membrane peeling were processed with or without color adjustments. We also evaluated the skewness (asymmetry in the distribution of the pixels) and kurtosis (sharpness in the distribution of the pixel) of the images to evaluate the contrast with each intensity of image-sharpening. Results Our results showed that the mean visibility score increased significantly from 4.9 ± 0.5 at 0% (original image) to 6.6 ± 0.5 at 25% intensity of the image-sharpening algorithm (P
... HDRI is the technique of capturing full dynamic range of luminance and contrast present in a scene, similar to what is experienced by the human visual system [8,191]. The human eye, continuously adapts itself, through the aperture of the iris, both to high and low exposure regions in the image. ...
Thesis
Dans cette thèse, nous discutons de quatre scénarios d’application différents qui peuvent être largement regroupés dans le cadre plus large de l’analyse et du traitement d’images à haute résolution à l’aide de techniques d’apprentissage approfondi. Les trois premiers chapitres portent sur le traitement des images de télédétection (RS) captées soit par avion, soit par satellite à des centaines de kilomètres de la Terre. Nous commençons par aborder un problème difficile lié à l’amélioration de la classification des scènes aériennes complexes par le biais d’un paradigme d’apprentissage profondément faiblement supervisé. Nous montrons comment en n’utilisant que les étiquettes de niveau d’image, nous pouvons localiser efficacement les régions les plus distinctives dans les scènes complexes et éliminer ainsi les ambiguïtés qui mènent à une meilleure performance de classification dans les scènes aériennes très complexes. Dans le deuxième chapitre, nous traiterons de l’affinement des étiquettes de segmentation des empreintes de pas des bâtiments dans les images aériennes. Pour ce faire, nous détectons d’abord les erreurs dans les masques de segmentation initiaux et corrigeons uniquement les pixels de segmentation où nous trouvons une forte probabilité d’erreurs. Les deux prochains chapitres de la thèse portent sur l’application des Réseaux Adversariatifs Génératifs. Dans le premier, nous construisons un modèle GAN nuageux efficace pour éliminer les couches minces de nuages dans l’imagerie Sentinel-2 en adoptant une perte de consistance cyclique. Ceci utilise une fonction de perte antagoniste pour mapper des images nuageuses avec des images non nuageuses d’une manière totalement non supervisée, où la perte cyclique aide à contraindre le réseau à produire une image sans nuage correspondant a` l’image nuageuse d’entrée et non à aucune image aléatoire dans le domaine cible. Enfin, le dernier chapitre traite d’un ensemble différent d’images `à haute résolution, ne provenant pas du domaine RS mais plutôt de l’application d’imagerie à gamme dynamique élevée (HDRI). Ce sont des images 32 bits qui capturent toute l’étendue de la luminance présente dans la scène. Notre objectif est de les quantifier en images LDR (Low Dynamic Range) de 8 bits afin qu’elles puissent être projetées efficacement sur nos écrans d’affichage normaux tout en conservant un contraste global et une qualité de perception similaires à ceux des images HDR. Nous adoptons un modèle GAN multi-échelle qui met l’accent à la fois sur les informations plus grossières et plus fines nécessaires aux images à haute résolution. Les sorties finales cartographiées par ton ont une haute qualité subjective sans artefacts perçus.
... With the continuous progress of imaging technologies in recent years, one particular demand is to display the acquired images in high quality to resemble the real scenes [1]. The ordinary low dynamic range (LDR) image with 8 bits per channel is usually insufficient to cover all light attributes of a real scene [2]. To derive the brightness information, researchers and practitioners have examined the transformation between the low and high-intensity dynamic extents to obtain the high dynamic range (HDR) images. ...
Article
Full-text available
Extending the dynamic range can present much richer contrasts and physical information from the traditional low dynamic range (LDR) images. To tackle this, we propose a method to generate a high dynamic range image from a single LDR image. In addition, a technique for the matching between the histogram of a high dynamic range (HDR) image and the original image is introduced. To evaluate the results, we utilize the dynamic range for independent image quality assessment. It recognizes the difference in subtle brightness, which is a significant role in the assessment of novel lighting, rendering, and imaging algorithms. The results show that the picture quality is improved, and the contrast is adjusted. The performance comparison with other methods is carried out using the predicted visibility (HDR-VDP-2). Compared to the results obtained from other techniques, our extended HDR images can present a wider dynamic range with a large difference between light and dark areas.
Article
This letter presents a low‐power single‐slope analog‐to‐digital converter (ADC) for column‐parallel architectures. A simple and effective design technique is proposed to solve the input‐dependent power consumption problem of the conventional single‐slope ADCs. A decision‐feedback loop is implemented in the second stage of the comparator. Based on the negative‐feedback path, which is activated after the signal decision of the comparator, the input‐dependent dynamic current path in the amplifier is disabled. Furthermore, an additional low‐power design technique is proposed to save the power consumption of the ADC by optimizing the offset cancelling auto‐zero period. With the combination of the proposed techniques, the power consumption of the single‐slope ADC can be effectively saved while suppressing the dynamic current.
Preprint
High Dynamic Range Imaging Based on Attenuation Microarray Mask has broad application prospects due to its good real-time performance and small size. But at the current level of craftsmanship, it is hard to fabricate a micro-attenuation array mask whose attenuation rate is adjustable. This leads to the fact that the imaging dynamic range cannot adapt to changes in scene brightness in most cases. To this end, this paper proposes a novel imaging system where the dynamic range can be adaptively changed according to the brightness of the scene. The core components are the micro polarization array mask mounted on the CMOS surface and the on-sensor rotatable linear polarizer in front of the lens. By controlling the rotation angle of the polarizer placed before the lens, the CMOS pixel exposure can be precisely controlled. Therefore, the imaging system dynamic range can be adjusted adaptively according to the scene brightness. The experimental results show that the imaging performance remains consistently good even when the dynamic range of the scene is large. By rotating the polarizer in front of the lens to a specific angle, the high dynamic imaging of the scene can be significantly improved.
Article
Full-text available
In this paper, the author shares her experiences in producing time-lapse movies dedicated to teaching introductory level meteorology and micro-meteorology, intended as dynamical phenomena unfolding in time. In this work, the process of producing teaching-grade meteorological time lapses is discussed, along with indications based on literature and the author’s experience to obtain evocative results using easily accessible, low-cost off-the-shelf equipment. Ideas on use of time-lapse meteorological and micro-meteorological movies to build teaching presentations are also given, with practical suggestions on improving students learning experience, to allow them developing their own scientific voice, and hopefully to spark enduring passions on a subject often considered dry and cold.
Article
Full-text available
Within the image reproduction pipeline, display devices limit the maximum dynamic range. The dynamic range of natural scenes can exceed three orders of magnitude. Many films and newer camera sensors can capture this range, and people can discriminate intensities spanning this range. Display devices, however, such as CRTs, LCDs, and print media, are limited to a dynamic range of roughly one to two orders of magnitude. In this paper, we review several algorithms that have been proposed to transform a high dynamic range image into a reduced dynamic range image that matches the general appearance of the original. We organize these algorithms into two categories: tone reproduction curves (TRCs) and tone reproduction operators (TROs). TRCs operate pointwise on the image data, making the algorithms simple and efficient. TROs use the spatial structure of the image data and attempt to preserve local image contrast. The basic properties of algorithms from each class are described using both real and synthetic monochrome images. We find that TRCs have difficulty preserving local image contrast when the intensity distributions of the bright and dark regions of the image overlap. TROs, which are traditionally based on multiresolution decomposition algorithms such as Gaussian decomposition, better measure and preserve local image contrast. However, the decomposition process can lead to unwanted spatial artifacts. We describe an approach for reducing these artifacts using robust operators. Coupled with further analyses of the illumination distribution in natural scenes, we believe robust operators show promise for reducing unwanted artifacts in dynamic range compression algorithms.
Article
Full-text available
— Intraocular glare and simultaneous contrast control appearance in high-dynamic-range (HDR) images. This paper describes unique test targets that simulate real images. These targets change the HDR range by 500 times, without significantly changing the veiling glare on the retina. As well, these targets have nearly constant simultaneous contrast. The range of appearances possible from HDR images with different average luminances were measured. The targets displayed a maximum luminance range of 5.4 log units. Using magnitude estimates (MagEst) of appearances, the relationship between luminance and lightness from white to black was measured. With one exception, only small changes in appearance with large changes in dynamic range were found. It was also found that appearance was scene-dependent. The same dark grays (MagEst = 10) were observed with luminances of 10, 4.2, 1.1, and 0.063, depending on the percentage of white area in the surround. Glare from more white increases the retinal luminance of the test areas. Simultaneous contrast counteracts glare by making the appearance range (white-black) with a much smaller range of luminances. Appearance is controlled by both the optical scattered light and the spatial processing. A single tone-scale function of luminance cannot describe appearance controlled by scatter and spatial processing.
Article
Full-text available
— Starting from measured scene luminances, the retinal images of high-dynamic-range (HDR) test targets were calculated. These test displays contain 40 gray squares with a 50% average surround. In order to approximate a natural scene, the surround area was made up of half-white and half-black squares of different sizes. In this display, the spatial-frequency distribution approximates a 1/f function of energy vs. spatial frequency. Images with 2.7 and 5.4 optical density ranges were compared. Although the target luminances are very different, after computing the retinal image according to the CIE scatter glare formula, it was found that the retinal ranges are very similar. Intraocular glare strongly restricts the range of the retinal image. Furthermore, uniform, equiluminant target patches are spatially transformed to different gradients with unequal retinal luminances. The usable dynamic range of the display correlates with the range on the retina. Observers report that appearances of white and black squares are constant and uniform, despite the fact that the retinal stimuli are variable and non-uniform. Human vision uses complex spatial processing to calculate appearance from retinal arrays. Spatial image processing increases apparent contrast with increased white area in the surround. Post-retinal spatial vision counteracts glare.
Article
Full-text available
The transition from traditional 24-bit RGB to high dynamic range (HDR) images is hindered by excessively large file formats with no backwards compatibility. In this paper, we demonstrate a simple approach to HDR encoding that parallels the evolution of color television from its grayscale beginnings. A tone-mapped version of each HDR original is accompanied by restorative information carried in a subband of a standard output-referred image. This subband contains a compressed ratio image, which when multiplied by the tone-mapped foreground, recovers the HDR original. The tone-mapped image data is also compressed, and the composite is delivered in a standard JPEG wrapper. To naive software, the image looks like any other, and displays as a tone-mapped version of the original. To HDR-enabled software, the foreground image is merely a tone-mapping suggestion, as the original pixel data are available by decoding the information in the subband. Our method further extends the color range to encompass the visible gamut, enabling a new generation of display devices that are just beginning to enter the market.
Article
Full-text available
We studied the effect of image quantization by compar-ing observer detection performance with 8-and 16-bit grayscale presentation. Eight readers evaluated 532 image pairs using a two-alternative forced choice experimental design. The image set consisted of synthetic backgrounds generated using the mammography-like cluster lumpy background (CLB) technique with a dual-layer approach with parameter values that have been shown to replicate the correlation structure found in digital mammography. The image pairs were reviewed in a display device pro-totype with one million pixels capable of processing and displaying 16-bit images (up to 65536 luminance values). These image pairs were presented either as non-quantized (full range) images in a 16-bit presentation scale, or as quantized, 8-bit images, with a perceptual mapping of gray levels to luminance. The difference in reader perfor-mance between reads on quantized image pairs and reads on non-quantized image pairs were derived using fraction of correct decisions. The variance of our measurements was estimated using a multi-reader, multi-case analysis. Average reader performance difference between 16-and 8-bit quantization was 0.065 with an associated standard deviation of 0.048. Our study showed that image quanti-zation is an important factor in visual detection task, that is, a quantization from 16-to 8-bit significantly reduces reader detection performance.. The mention of commercial products herein is not to be construed as either an actual or implied endorsement of such products by the Department of Health and Human Services. This is a contribution of the Food and Drug Administration and is not subject to copyright. Disclosure: This research is funded in part with a Cooperative Research and Development Agreement between FDA and Philips FIMI (Italy).
Chapter
This chapter analyzes the concept of real pixels. It discusses a floating point format that only requires 32 bits per pixel and is completely portable between machine architectures. The idea is using an 8-bit mantissa for each primary and following it with a single 8-bit exponent. In most floating point formats, the mantissa is normalized to lie between .5 and 1. Because this format uses the same exponent for three mantissas, only the largest value is guaranteed this normalization, and the other two may be less than .5. It appears that this format favors the largest primary value at the expense of accuracy in the other two primaries. This is true, but it is also true that the largest value dominates the displayed pixel color so that the other primaries become less noticeable. The 32-bit real pixel format presented in the chapter preserves the bits that are most significant, which is the general goal of any floating point format. The chapter highlights that besides the ability to perform more general image processing without losing accuracy, real pixels are great for radiosity and other lighting simulation programs, because the results can be evaluated numerically well outside the dynamic range of the display.
Conference Paper
Article
We present a high dynamic range viewer based on the 120-degree field-of-view LEEP stereo optics used in the original NASA virtual reality systems. By combining these optics with an intense backlighting system (20 Kcd/m2) and layered transparencies, we are able to reproduce the absolute luminance levels and full dynamic range of almost any visual environment. This technology may enable researchers to conduct controlled experiments in visual contrast, chromatic adaptation, and disability and discomfort glare without the usual limitations of dynamic range and field of view imposed by conventional CRT display systems. In this paper, we describe the basic system and techniques used to produce the transparency layers from a high dynamic range rendering or scene capture. We further present an empirical validation demonstrating device's ability to reproduce visual percepts, and compare this to results obtained using direct viewing and a visibility matching tone reproduction operator presented on a conventional CRT display.
Article
In this paper, we present a fast, robust, and completely automatic method for translational alignment of hand-held photographs. The technique employs percentile threshold bitmaps to accelerate image operations and avoid problems with the varying exposure levels used in high dynamic range (HDR) photography. An image pyramid is constructed from grayscale versions of each exposure, and these are converted to bitmaps which are then aligned horizontally and vertically using inexpensive shift and difference operations over each image. The cost of the algorithm is linear with respect to the number of pixels and effectively independent of the maximum translation. A three million pixel exposure can be aligned in a fraction of a second on a contemporary microprocessor using this technique.
Article
The work done to provide image sensors (CCDs and CMOS) with a wide dynamic range is reviewed. The different classes of solutions, which consist of logarithmic sensors, "clipped" sensors, multimode sensors, frequency-based sensors, and sensors with control over integration time are described. The pros and cons of each solution are discussed, and some new experimental results are shown. Active pixel sensors with a wide dynamic range are analyzed and possible future directions are pointed out. (C) 1999 Society of Photo-Optical Instrumentation Engineers. [S0091-3286(99)01610-4].