Impact Factor 2.380 | CiteScore 4.7
More on impact ›


Front. Comput. Neurosci., 22 December 2015 |

A Retina Inspired Model for Enhancing Visibility of Hazy Images

Xian-Shi Zhang1, Shao-Bing Gao1, Chao-Yi Li1,2 and Yong-Jie Li1*
  • 1Key Laboratory for Neuroinformation of Ministry of Education, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu, China
  • 2Center for Life Sciences, Shanghai Institutes for Biological Sciences, Chinese Academy of Sciences, Shanghai, China

The mammalian retina seems far smarter than scientists have believed so far. Inspired by the visual processing mechanisms in the retina, from the layer of photoreceptors to the layer of retinal ganglion cells (RGCs), we propose a computational model for haze removal from a single input image, which is an important issue in the field of image enhancement. In particular, the bipolar cells serve to roughly remove the low-frequency of haze, and the amacrine cells modulate the output of cone bipolar cells to compensate the loss of details by increasing the image contrast. Then the RGCs with disinhibitory receptive field surround refine the local haze removal as well as the image detail enhancement. Results on a variety of real-world and synthetic hazy images show that the proposed model yields results comparative to or even better than the state-of-the-art methods, having the advantage of simultaneous dehazing and enhancing of single hazy image with simple and straightforward implementation.


The necessary processing of visual information already begins in the eye, the very first stage of the visual system. Increasing evidence suggests that the mammalian retina seems far smarter than scientists have believed so far (Gollisch and Meister, 2010; Lee et al., 2010; Masland, 2012). The retina is a neural circuit composed of at least 50 clearly distinct cell types (Joselevitch, 2008). These cells form various retinal subsystems that serve a diverse set of specific tasks, e.g., light adaptation and image sharpening. Together, these retinal neurons and their coding strategies enable the visual system to perform well by adapting to the complicated environments, e.g., with changing air conditions.

From the point of view of engineering, many computer vision applications expect clear input images with high-contrast details. However, this situation is not always true in practical scenarios. Due to the presence of aerosols such as dust, mist and water droplets in the atmosphere, the reflected light from the object surface has already been scattered before it reaches the camera. As shown in Figure 1, this light-scattering phenomenon consequently results in the contrast reduction and color fading, which eventually cause the definition decrease in the captured images. This image degradation annoys not only computer vision applications but also consumers and commercial photographs, and therefore, image dehazing and visibility enhancing have become more and more important in this digital age.


Figure 1. Hazy images (top) and the images after haze removal by the proposed method (bottom).

Haze removal is a challenging task because the haze condition is always unknown. Most existing methods use additional information to solve this ill-posed problem. For example, Schaul et al. combine the hazy image with its near-infrared version (Schaul et al., 2009). Depth-based methods assume that the depth information of the image is available or is accessible by the three-dimensional (3D) geometrical model of the scene and then use it to enhance the image (Tan and Oakley, 2000; Narasimhan and Nayar, 2003b; Hautière et al., 2007; Kopf et al., 2008). Although these methods can enhance the visibility of hazy images, they cannot be applied when such additional information is unavailable. There are also some methods that use two or more images of a same scene to estimate the haze (Nayar and Narasimhan, 1999; Narasimhan and Nayar, 2000, 2003a; Schechner et al., 2001; Shwartz et al., 2006; Namer et al., 2009; Treibitz and Schechner, 2009). In general, these methods can remove haze well but require multiple images. For a single image, haze removal methods based on some priors have been developed. For example, Tan (2008) assumes that a haze-free image should have higher contrast than the input hazy image, based on which the haze is removed by maximizing the local contrast of the input hazy image. Tarel et al. develop a real-time haze removal method (Tarel and Hautiere, 2009), also based on the information of local contrast, by assuming that the atmospheric veil is smooth at most of the time and taking the percentage between the local standard deviation and the local mean of the whiteness as the air light. Fattal (2008) assumes that the transmission and the surface shading in a single input image are locally uncorrelated, and then eliminates the scattered light by estimating the optical transmission in hazy scenes. Kratz and Nishino (2009) try to estimate both the scene albedo and depth, which contain valuable structural information of haze. Ancuti et al. (2011) assume that the distance between the observer and the scene objects is highly correlated with the contrast degradation and color fading, and then detect haze quickly by a comparison of the hue values in the input image and its “semi-inversed” version. Dark channel prior based methods (He et al., 2009, 2011) assume that the pixels in the dark channel of haze-free images are close to zero, while in hazy images these pixels obtain higher intensity from the airlight, and therefore utilizing these pixels can accurately estimate the haze transmission. Many recent progresses improve the dark channel based methods by replacing matting with other filters (Yu et al., 2010; Gibson et al., 2012) or adding new constraints (Tarel et al., 2012; Caraffa and Tarel, 2013; Meng et al., 2013). More recently, Tang et al. (2014) propose a learning framework to combine different haze-relevant features and provide flexibility for different specific situations.

Although there are remarkable progresses in single image haze removal, the problem is that most of these methods require additional information or prior assumptions. However, in different real-world images, the information may miss, or the assumptions may fail, and then the methods based on them would perform worse than expected. For example, when the scene objects are similar to the atmospheric light, the dark channel of the scene radiance has bright values near such objects, which means that the dark channel prior is invalid, and as a result the haze layer will be overestimated (He et al., 2011).

Along another line, attempts that follow the information processing mechanisms of the human visual system (HVS) seems to be a promising route to address this problem, inspired by the amazing ability of HVS to achieve stable perception under varying natural light environments (Foster, 2011). Models inspired by HVS have succeeded in many fields, such as face recognition (Vu and Caplier, 2009), boundary detection (Yang et al., 2013, 2015), key point descriptor (Alahi et al., 2012), color constancy (Spitzer and Semo, 2002; Spitzer and Barkan, 2005; Gao et al., 2013, 2015), multi-resolution image-fusion (Ghassemian, 2001) and image enhancement (Land and McCann, 1971; Land, 1986; Jobson et al., 1997; Rahman et al., 2004), with robust performance under varying conditions. Representatively, the models based on the famous Retinex theory, which approximates the spectral properties of object surfaces by the ratio of the reflected light in this area to others (Land and McCann, 1971; Land, 1986; Jobson et al., 1997; Rahman et al., 2004), can enhance hazy images with (Xie et al., 2010; Nair et al., 2014) or without priors (Woodell et al., 2005; Rajput and Rahman, 2008; Zhou and Zhou, 2013). Taken inspiration mainly from the color perception behavior of human in psychophysical experiments, Retinex theory has not yet clarified whether the formation of lightness images and their comparison occur in the retina, the cortex, or the both (Foster, 2011).

Different from the Retinex based models, our proposed model enhances hazy images by simulating the underlying mechanisms at the specific level of retina. In particular, the proposed model includes the processing inspired by the physiological findings that the receptive field (RF) surround of retinal ganglion cell (RGC) consists of many inhibitory subunits (or subfields), and the inhibitory interactions among them lead to a disinhibitory effect, which means an adaptive reduction of the surround inhibition to the RF center (Li and He, 1987; Li et al., 1991, 1992; Li and Li, 1994; Qiu and Li, 1995; Li, 1996). In addition, the information processing along the ON and OFF pathways is combined. This proposed model directly enhances the input hazy image, without requiring to first estimate a transmission map as did in most dehazing models. The novelty of the proposed model lies in its ability of simultaneous haze removing and detail enhancing of single image due to the specific mechanisms of different retinal sub-layers. The contribution of this work is not only an efficient way to enhance hazy images for computer vision applications, but also a computational description about the possible retinal mechanisms of image enhancement in biological vision.


General Description

The proposed method follows the color processing mechanisms in the retina (Figure 2). The single hazy image is the input and the enhanced image is the output.


Figure 2. The structure of the proposed retina-based model. The R, G, B, and brightness components of the input color image are respectively sent into the corresponding photoreceptors. Then the outputs of cone photoreceptors are transmitted via cone bipolar cells, the RF of which is a difference of Gaussian function, to the RGCs, the RF of which consists of a small excitatory center and a relatively large inhibitory surround (also named the non-classical receptive field, nCRF). The surround is composed of many inhibitory subunits, which first inhibit each other, and then inhibit the center. The outputs of rod bipolar cells modulate the outputs of cone bipolars. From top to bottom, the image patches listed on left side are respectively the input image, the output of the bipolar cells, the bipolar output modulated by AII amacrine cells, and the final output the ganglion cells. Note that in primates, the densities of M and L cones are typically higher than S cone density by 20-fold (Euler et al., 2014), so the signals sent into the cone type-non-selective diffuse bipolar cells are mainly R and G components, and here we omit the B component sent into the diffuse bipolars for clarity and simplicity. w is a weight controlling the relative contributions of the ON and OFF pathways, and we set w = fL(x, y) in this work (see details in the text).

Retinal information processing begins with the sampling of rod and cone photoreceptors. The red (R), green (G), and blue (B) components of the input color image are responded respectively by long-, medium-, and short-wavelength cone photoreceptors (i.e., L, M, and S cones) of retina, while the brightness of the dim regions in the input color image is sensitively responded by rod photoreceptors (Masland, 2012). The photoreceptor activities are adjusted and transmitted to the RGCs via several retinal sub-layers like horizontal cells, bipolar cells and amacrine cells. In particular, the cone signals are then transmitted by ON- and OFF-type cone bipolar cells, whereas the rod signals are transmitted only by the ON-type rod bipolars, supported by the current biological evidence that unlike the cones, the rod system in the cat, monkey, and human has only ON-type rod bipolars (Schiller, 2010). In the ganglion-cell layer, the output layer of the retina, color signals are processed by the RGCs with single-opponent receptive field (RF), which receives opponent stimuli in its excitatory center and inhibitory surround from two (or more) different cones (Conway et al., 2010). The RF surround of a RGC is physiologically supposed to comprise many inhibitory subunits, each of which is first inhibited by its neighboring subunits (i.e., disinhibit), and then inhibits the neuronal response elicited by the RF center.

Visual Processing in Bipolar Cells

After the light absorption by photoreceptors, bipolar cells transmit the neuronal activities to RGCs. Continuous glutamate release from photoreceptors is suppressed by the light. Thus, bipolar cells that express sign-conserving ionotropic glutamate receptors are depolarized in the dark while bipolar cells that express sign-inverting metabotropic glutamate receptor are depolarized in the light. Differed in terms of response polarity, there are two major classes of bipolar cells: ON bipolar cells, which response to light on-set, and OFF bipolar cells, which response to light off-set (Schiller, 2010). Beside the RF center, an inhibitory surround of bipolar cells has been identified more than 40 years (Werblin and Dowling, 1969b). The RF of most bipolar cells consists of two regions: a smaller excitatory center and a lager inhibitory annular surround (Werblin and Dowling, 1969a; Kaneko and Tachibana, 1983), which is commonly described by the “difference of Gaussian” (DOG) model (Rodieck and Stone, 1965; Enroth-Cugell and Robson, 1966). The neurons with such RF type can transmit high-acuity and chromatic opponent signals. Although the cellular mechanisms and neural circuitry are not totally clear, many evidences support that horizontal cells mainly contribute to this surround antagonism (Thoreson and Mangel, 2012).

Based on the dendritic morphology, primate cone bipolar cell types have been divided into three groups: diffuse bipolar cells, S-cone-selective bipolar cells, and midget bipolar cells. Diffuse bipolar cells non-selectively contact between four and fifteen neighboring cones of various types. Because in primates, the combine density of M- and L-cones is typically higher than S-cone density by 20-fold, the signals of diffuse bipolar cells tend to be chromatically biased toward yellow (Euler et al., 2014). For better understanding, we also call these diffuse bipolar cells the yellow-sensitive bipolar cells in this paper. Combined with S-cone-selective bipolar cells, diffuse bipolar cells contribute to blue-yellow color vision. Near the fovea, a midget bipolar cell receives direct input from just one cone and in turn transfers the signal to one midget ganglion cell. The midget bipolar cells carry chromatic signals and are considered to be the basis for red-green color vision (Euler et al., 2014). The majority of cones in central retina connect with at least one ON bipolar cell and one OFF bipolar cell, which in turn connect respectively with ON and OFF retinal ganglion cells (Schiller, 2010).

There are two pathways in which cone bipolar cells carry chromatic information to RGCs. First, the cone type unselective pathway, in which a midget bipolar cell randomly contacts a single cone and feed into a single midget RGC, carries red-green information. Second, blue-yellow information is carried by the cone type selective pathway. In this way, the fact that blue-ON-yellow-OFF RGCs differentially pool signals from S-cone-selective ON and diffuse OFF bipolar cells is well-known, but where blue-OFF-yellow-ON RGCs inherit the blue-OFF signal from is still controversial. Some former researches report that the blue-OFF signal is from S-cone-selective ON bipolar cells via a sign-inverting small field amacrine cell (Chen and Li, 2012), while recent anatomical and physiological evidences support that it is from S-cone-selective OFF bipolar cells directly (Mills et al., 2014).

Given an input image fc(x, y), c ∈ {R, G, B} normalized within [0, 1] by dividing each channel with the maximum intensity across three channels, the three components are respectively sampled by three cone types of cones, and the luminance component fL(x, y) described by Equation (1) is sampled by the rods

fL(x,y)=(fR(x,y)+fG(x,y)+fB(x,y))/3.    (1)

Then, the signals sampled by the photoreceptors are further sent into the bipolar cells. For clarity, we use fY(x, y) to represent the combined signals of R, G, and B sent into the diffuse yellow (Y)-sensitive bipolar cells. By ignoring the quite less input from B channel based on the physiological evidence (Euler et al., 2014), fY(x, y) can be simply computed as

fY(x,y)=(fR(x,y)+fG(x,y))/2.    (2)

As indicated in Figure 2, the color components sampled by cone photoreceptors are further processed by the R, G, B, and Y-sensitive bipolar cells of both ON and OFF types, and in contrast, the luminance component sampled by rods is processed by the rod bipolar cells of only ON type (Schiller, 2010). In a general form, we denote the outputs of ON bipolar cells as BPc(x, y), c ∈ {R, G, B, Y, L} and the outputs of OFF bipolar cells as BPc(x,y), c{R,G,B,Y}, which are computed as

BPc(x,y)=fc(x,y)(g(x,y;σcen)k·g(x,y;σsur))    (3)
BPc(x,y)=fc(x,y)(g(x,y;σcen)k·g(x,y;σsur))    (4)

where ⊗ is a convolution operator. fc(x,y)=1fc(x,y) is the input sent into the OFF bipolar cells. k represents the sensitivity of the inhibitory annular surround, σcen and σsur are respectively the standard deviations of Gaussian shaped RF center and its surround, which are experimentally set to be 0.5 and 1.0, respectively in this work. g(x, y; σ) is a two-dimensional (2D) Gaussian function written as

g(x,y;σ)=12πσ2exp((x2+y2)/(2σ2)).    (5)

The ON-type rod bipolar cells send their outputs BPL (x, y) to the specific AII type of amacrine cells (Schiller, 2010). The AII amacrine cells modulate the activities of ON cone bipolar cells via sign-conserving gap junctions and OFF cone bipolar cells via inhibitory synapses (Lee et al., 2010). By this way, rod bipolar cells finally excite the ON RGCs and inhibit the OFF RGCs indirectly. Based on this biological fact, the modulated responses of ON and OFF bipolar cells, MBPc(x, y) and MBPc(x,y), are respectively given by

MBPc(x,y)=BPc(x,y)·(ε+BPL(x,y))    (6)
MBPc(x,y)=BPc(x,y)/(ε+BPL(x,y))    (7)

where ε is a constant, which is set to be 0.5 to avoid being divided by zero and also to keep the multiplier and divisor to be around 1.0 for easy control in the later use. Note that BPc(x, y) and BPc(x,y) have been rectified by setting the negative values to zero and then normalized respectively to [0, 1] by dividing each channel with the maximum intensity across three channels.

It is clear that from Equations (6) and (7), with the modulation of AII amacrine cells that respond to the luminance component of scene (i.e., BPL(x, y)), the outputs of most cone ON bipolar cells (i.e., BPc(x, y), c ∈ {R, G, B, Y}) will be amplified in the brighter regions eliciting higher BPL (x, y), and in contrast, the outputs of most cone OFF bipolar cells (i.e., BPc(x,y), c{R,G,B,Y}) will be enhanced in the darker regions eliciting lower BPL(x, y). This will be further demonstrated in the section of Experiments.

Disinhibition in RGCs

Retinal ganglion cells (RGCs) receive multiple cone signals transmitted via bipolar cells (and other cells) and compare them with the color-opponent mechanism. Like bipolar cells, the RF of most RGCs also consists of a smaller excitatory center and a larger inhibitory annular surround, and chromatically single-opponent RGCs receive inputs of different cone types within these two different RF regions (Conway et al., 2010). Although the RF center of a RGC, whose diameter is larger than a single cone because of physiological optics, a point spread function of the eye exceeds the size of a single cone in the fovea, is created by direct inputs from a single bipolar cell which contacts an individual cone, the synaptic pathways that create the opponent RF surround remain controversial (Lee et al., 2010). There are two main hypothesizes, i.e., unselective-connection and selective-connection. The unselective-connection hypothesis predicts that mixed cone inputs from non-selective horizontal cells to the surround lead to the opponency. In this hypothesis, color opponency is a product of interaction created by the horizontal cells and arises by subtracting individual cone forming center with all cones feeding in surround. This hypothesis is supported by some recently researches in the peripheral retina (Field et al., 2010; Crook et al., 2011). However, there are more direct physiological evidences supporting the hypothesis that the opponent surround is the result of cone specific or partially selective connection with bipolar cells (Martin et al., 2001; Reid and Shapley, 2002; Buzás et al., 2006; Sun et al., 2006; Lee et al., 2012). In this work we consider three types of single-opponent ON RGCs with selective connections: L/M, M/L, S/(L+M), which means that the firing rate of a RGC increases with the activation of one cone type (e.g., L or R) and decreases with the activation of a different cone type (e.g., M or G). In the following, we will use R, G, and B for short to denote respectively the above three single-opponent channels. In addition, physiological experiments have observed a secondary gentle rise in the neuronal response of some RGCs when the stimulus was further extended far beyond the RF center, which indicates a disinhibitory effect contributed by the extensive surround (Li and He, 1987; Li et al., 1991, 1992; Li and Li, 1994). The RF surround with disinhibitory effect (also called the non-classical receptive field, nCRF) was presumed to comprise many inhibitory subunits, which first inhibit each other, and then inhibit the RF center (Qiu and Li, 1995; Li, 1996).

In the propose model, Uc(x, y; σu), c ∈ {R, G, B, Y} denotes the response of a subunit centered at (x, y) in the RF surround after being inhibited by other subunits. We compute it according to

Uc(x,y;σu)=MAX[0, MBPc(x,y)Au · MBPc(x,y)                        g(x,y;σu)]    (8)

where Au represents the sensitivity of subunits. MAX is a max operator to keep the neuronal response non-negative.

Then, let Sc(x, y; σs), c ∈ {R, G, B, Y} denote the total responses of all the inhibited subunits in the surround. We compute them as

Sc(x,y;σs)=Uc(x,y;σu)g(x,y;σs).    (9)

Then, the final response of a ON ganglion cell, GCc(x, y), c ∈ {R, G, B}, is the response elicited by the excitatory RF center in one channel (e.g., R) subtracted by the total surround inhibition from its opponent channel (e.g., G), which is written as

GCR(x,y;σcen)=Acen·MAX[0,MBPR(x,y)g(x,y;σcen)                         As·SG(x,y;σs)]GCG(x,y;σcen)=Acen·MAX[0,MBPG(x,y)g(x,y;σcen)                         As·SR(x,y;σs)]GCB(x,y;σcen)=Acen·MAX[0,MBPB(x,y)g(x,y;σcen)                        As·SY(x,y;σs)]    (10)

where Acen represents the sensitivity of the excitatory RF center and As is the sensitivity of inhibitory RF surround of ganglion cell.

In Equations (8)–(10), σu, σs, and σcen are the standard deviations of the 2D Gaussian functions describing the subunit, RF surround and RF center, respectively, and they are set to be one third of the radius of 2D Gaussian shaped regions. In this work, we experimentally set the radius of subunit, RF surround and center as 1, 3, and 1 pixel, respectively. A partial evidence in support of this setting is the neurophysiological finding that the size of RF surround is normally 2–5 times larger (in diameter) than that of RF center (Li and He, 1987; Li et al., 1991, 1992; Li and Li, 1994).

Similarly, with MBPc(x,y) as the modulated output from a OFF bipolar cell, the response of a OFF ganglion cell, GCc(x,y),c{R,G,B}, can be easily computed based on Equations (8)–(10).

Biologically, the ON and OFF cells in the retina form almost separated pathways (Schiller, 2010), the signals along which are integrated in the visual cortexes via lateral connections. Modeling such integration is beyond the scope of this work. Here we simply compute the retinal outputs of ON and OFF pathways, respectively, as

OutON(x,y)=13c{R,G,B}GCc(x,y)OutOFF(x,y)=13c{R,G,B}GCc(x,y).    (11)

Then the final output of the proposed model is computed by combining the outputs of ON and OFF pathways according to

Out(x,y)=w·OutON(x,y)+(1w)·(1OutOFF(x,y))    (12)

where w is a weight controlling the relative contributions of the ON and OFF pathways. As mentioned above, the ON channel emphasizes the perception of bright regions, and the OFF system emphasizes the perception of dark regions. Such notion will be further validated by our experiment (see the following section). This inspires us to set w = fL(x, y). Such setting indicates that the signals from the ON and OFF ganglion cells are adaptively combined based on the local brightness at (x, y).


In this section, we first show the responsive properties of the model ganglion cells and the model bipolar cells involved in the proposed system to demonstrate how they contribute to the haze removal. We also illustrate the different roles of the ON and OFF pathways in enhancing hazy images. Then we conduct the qualitative comparison on real hazy images as well as quantitative comparison on synthetic images with other representative algorithms.

Responsive Properties of the Bipolar Cells

The RF of bipolar cell is described by a DOG model [see Equations (3) and (4)]. Due that the DOG-shaped RF is a typical band-pass filter, which may partially attenuate the low frequency components (e.g., the dispersively distributed haze) of the hazy image when the two Gaussian functions are unbalanced. In addition, there is no doubt that the high-frequency details of the image will be also degraded to certain extent. As shown in Figure 3, from left to right, the higher the sensitivity of the inhibitory annular surround [i.e., the parameter k in Equations (2) and (3)] is, the greater the bipolar cells remove the haze, but more details loss. The difference between the top and bottom images in Figure 3 shows that the modulation from the rod bipolar cells via AII amacrine cells can compensate the loss of details by increasing the image contrast, but degrade the saturation.


Figure 3. Response of the bipolar cells. From left to right, the sensitivity of surround [i.e., the parameter k in Equations (2) and (3)] is set to be 0.5, 0.6, and 0.7. The top and bottom rows are the images before and after being modulated by the brightness-driven AII amacrine cells.

Responsive Properties of the Ganglion Cells

Considering that the RF center and its surround of a red-ON-green-OFF ganglion cell are covered by the red and green patches of equi-luminance, we measured a group of area-response curves (Figure 4) by computing the response of this model ganglion cell with varying subunit sensitivities (Au) while keeping the other parameters fixed. It is clear that for Au > 0, when the stimulus is extended larger than the RF center, the neuronal response is reduced due to the involved surround inhibition, and then gradually enhanced due to the increasing disinhibition effect deduced by the inhibitory interaction among more surround subunits. This observation is quite consistent with that obtained from electrophysiological studies in the cat retinal ganglion cells (Li and He, 1987; Li et al., 1992; Li and Qiu, 1994). This figure clearly shows that higher Au values result in stronger disinhibition effects, and hence weaker surround inhibitions and higher neuronal responses.


Figure 4. Responsive curves as a function of stimulus size (in radius) for different subunit sensitivities. The sizes of RF center, RF surround and subunits are 3, 10, 3 pixels, respectively. The sensitivity of RF surround [i.e., As in Equation (9)] is 3.0, and the subunit sensitivity Au in Equation (8) is set to be 0.8, 0.6, 0.4, 0.2, and 0.0. Relative response represents the final response in channel R (i.e., GCR (x, y)) of a red-ON-green-OFF ganglion cell computed by Equation (10) with Acen = 1.0 certain Au when the RF center and its surround are covered by the red and green patches of equi-luminance.

To illustrate how ganglion cells enhance the hazy images, we designed a series of stimulus patches (Figure 5A), each of which contains a red block (simulating an object surface) covering the full RF center and a patterned white foreground across both the RF center and its black surround. When all parameters are fixed and the white foreground is spatially distributed with systematically increasing dispersion (quantified by a “dispersive angle,” Li and He, 1987; Li et al., 1992; Li and Qiu, 1994) while keeping the total light flux identical, the response of a red-ON ganglion cell to the composited stimuli is consistently increased with the increasing dispersive angle of the white foreground (the blue curve in Figure 5B), because more subunits in the surround are activated and hence stronger disinhibition (and then weaker surround inhibition) is introduced. Note that this increase is not observed when no disinhibition is involved (i.e., Au = 0), because the surround inhibition is always equal due to the identical total light flux (with any degree of dispersion) in the surround.


Figure 5. An example illustrating the effect of various degrees of dispersion of foreground light on the response of a retinal ganglion cell. (A) The stimuli with a size of 5 by 5 pixels, which contains a red block of 3 by 3 pixels covering the full RF center and a patterned bright foreground across both the RF center and its surround. The bright patterns have the same total light flux but different dispersive angles, from 0 to 180°, as indicated on the top of the panels. (B) Response vs. dispersion curves of a ganglion cell. The red curve is with surround inhibition but without subunit disinhibition. The blue curve is for the case that both the subunit disinhibition and surround inhibition work. The sizes of RF center, RF surround and subunits are 3*3, 7*7, 3*3 pixels, respectively, with Acen = 1.0 and Au = 0.5As. R channel response denotes GCR(x, y) computed by Equation (10) for a red-ON ganglion cell to certain stimulus patch listed in (A).

Considering that the white foreground also covers the red block in the RF center, more dispersive the white foreground is, more likely it looks that the foreground is the haze (e.g., the last panel of Figure 5A). In this situation, the hazy appearance of the red block (most likely to be a local object surface) is relatively weakened by the increasing of the R channel response, i.e., the red block becomes redder than the original block covered with gray foreground. This example provides a biological explanation about how the disinhibition among subunits in the RF surround contributes to the hazy image enhancement at the retinal ganglion layer.

Different Roles of ON and OFF Channels

As mentioned above, ON and OFF neurons are excited by the increment and decrement of light intensity in the visual scene, which have been suggested to enable more efficient encoding of sensory stimuli (Joselevitch, 2008). In specific, the functional significance of the separated ON and OFF channels mainly includes the extension of dynamic range, improved SNR, increased spatial resolution, and other information processing benefits (Joselevitch, 2008). For example, ON and OFF bipolar cells would each amplify half of the dynamic range of the photoreceptor signals by responding mainly to the positive and negative contrasts, respectively (Ratliff et al., 2010). Figure 6 shows an example illustrating the different functional roles of the ON and OFF system. For the ON pathway, the quite bright roof of the small tabernacle at the right-top corner of the image can be clearly responded by the ON ganglion cells with appropriate setting (e.g., Au = 0.5 or 0.7), but the details and colors of the sunless regions (e.g., the dominant region with a crowd of people) can not be properly enhanced along the ON pathway. In contrast, though the OFF system outputs an over-saturated roof of the tabernacle in terms of brightness, this pathway can vividly enhance the details of the dim regions. In short, it is clear that the OFF system carries on more dark elements while the ON system transmits and processes more bright elements. By merging the outputs from the ON and OFF systems, the final result combines the benefits from the both sources, as indicated by the first image of the bottom row in Figure 6.


Figure 6. Illustration of the different roles of ON and OFF pathways in haze removal. The first image of the top row is the original hazy image. The images from the second to fourth columns of the top and bottom rows are respectively the outputs of ON and OFF ganglion cells with various subunit sensitivity values (Au). The first image of the bottom row is the image after haze removal by the full model proposed in this work, i.e., by linearly combining the output of the ON ganglion cells with Au = 0.7 and the output of the OFF ganglion cells with Au = 0.5.

Visual Comparison on Real-World Images

For all the input hazy images in the following sections, we experimentally set k = 0.3, Ac = As = 2, and Au = 0.7 and 0.5 for the ON and OFF ganglion cells, respectively.

Figure 7 compares the result of our method with that of several state-of-the-art methods (Fattal, 2008; Kopf et al., 2008; Tan, 2008; He et al., 2009, 2011). Note that the results of Tan, Fattal, Koph, and He are quoted from In specific, Tan's method (Tan, 2008) removes haze by maximizing the local contrast of the input hazy image, which can well-enhance the local contrast, but normally results in oversaturated colors and halo artifacts. Fattal's method (Fattal, 2008) eliminates the scattered light by estimating the optical transmission in hazy scenes and requires sufficient color information and variance. Considering that the color of distant parts of the hazy image in Figure 7 is faint, which does not satisfy the requirement of Fattal's method, and hence in his result, this part is too bright to see the details like the towers. The method of Kopf et al. (2008) removes the haze by 3D models and texture maps of the scene, which needs additional geography information. With visible residual hazes, the result image of this method is not very clear. The method of He et al. (2009, 2011) is based on the dark channel prior. By removing haze and enhancing details, our result is comparable in saturation with He's, but has higher visibility of structures, except a little haze left over the distant region.


Figure 7. Qualitative comparison with several representative methods on a real-world image.

In Figure 8, we also compare our method with He's method on several traffic scenes captured by our group. Note that He's results on these images were calculated by us with the common parameter setting suggested in their paper: ω = 0.9, t0 = 0.1 and a patch size of 15*15 pixels (see the detailed meaning of these parameters in He et al., 2009, 2011). Since He's dark channel prior based method estimates the atmospheric light (the parameter A in his model) as constant, its result is sometimes a little dark and may lose some details (e.g., the electrical wires occurring in the sky region of the bottom image of Figure 8) when the atmospheric light seems to vary spatially. In contrast, in the results recovered by our model, cars are easier to see, and the road markings and traffic signs are more visible and can be easier to read out.


Figure 8. Comparisons with He's method on real-world traffic images.

Quantitative Comparison on Synthetic Images

With the haze-free images and the known disparity maps d(x) (Scharstein and Szeliski, 2003; Hirschmuller and Scharstein, 2007; Scharstein and Pal, 2007), we set the transmission map t(x) = 0.8 × d(x) in the Koschmieder model (Koschmieder, 1925), and then synthesize the hazy images according to

I(x)=J(x)·t(x)+A·(1t(x))    (13)

where J(x) is the input haze-free image, I(x) is the hazy image, and A is the global atmospheric airlight, assumed as pure white for simplicity.

Quantifying by the commonly used mean squared error (MSE) between the original haze-free image and the recovered image, we compared the performance of our method with that of He (He et al., 2009, 2011) in Table 1 on a group of synthetic images. In general, a lower MSE means a better performance in removing haze. This table clearly shows that our results are much more close to the original haze-free images in all eight scenes. Two examples are shown in Figure 9. He's results are a little dark and our results have more vivid colors. The enlarged local patches show that He's results are a little hazier than that by our model.


Table 1. Quantitative comparison of MSE with He's model on synthetic images.


Figure 9. Comparison with He's method on two synthetic hazy images. (A) Teddy image, (B) Art image. For each of (A,B), the top row shows the clear image and the image added with haze, the middle row lists the results by He's and our methods, and the bottom row lists the enlarged local patches. The Teddy and Art images are downloaded from


In this paper, without explicit assumption, we proposed a hazy image enhancement method inspired by the information processing mechanisms of the retinal network in the biological visual system, from the photoreceptors absorption via the bipolar cells to the opponent ganglion cells equipped with subunit-structured RF surround. The proposed model was compared with the state-of-the-art methods qualitatively and quantitatively, and shows competitive results on both real-world and synthetic images. In particular, our model can remove the haze as well as clearly enhance the details, without sacrificing the color fidelity.

This simplified retinal model proposed here also provides valuable suggestions about the role of the RGC surround disinhibition as well as the ON and OFF pathways in image enhancement. As shown in Equation (13), an image with haze loses its contrast and saturation. In our model, the DOG shaped RF of bipolar cells, which partially filters out the low spatial frequency elements, serves to obtain an incomplete but fast reduction in the obvious influence of the airlight components. The modulation from the AII amacrine cells driven by rod bipolars is capable of increasing the contrast of the image, but decreasing the saturation. This undesired decrease of saturation is then compensated by the chromatically single-opponent RGCs. Subunit-deduced surround disinhibition in RGCs contributes a fine and adaptive correction based on the spatial resolution of the disinhibition. Roughly speaking, the proposed model realizes the haze removal by enhancing the contrast of details and recovering the object colors while reducing the low spatial frequency components of scenes.

Though performing well in dehazing and enhancing single haze images, our model also has some limitations. In particular, our method does not perform quite well for the images with quite dense haze. As shown in Figure 10, due to the heterogeneously distrusted thick haze, the travelers, trestle and plants are almost visually undetectable. Though the visibility is improved significantly after haze removal by our model, there is still clear haze left in the top part of the scene. However, it is interesting to point out that the dehazing performance for the scenes with heterogeneously distributed heavy haze as shown in Figure 10 may be improved by introducing the dynamic spatiotemporal structure of receptive field (RF). In the space domain, profound contrast-dependent change in RF size, normally an inverse relationship between stimulus contrast and the RF size, has been well-observed for the retinal ganglion cells, LGN and V1 cells (Sceniak et al., 1999; Nolt et al., 2004; Chen et al., 2013). For example, at the level of retina, Nolt et al. (2004) reported that the size of the receptive field center decreases with an increase in contrast for both LGN cells and RGCs. In particular, the center size was, on average, 1.99 times greater at low contrast than at high contrast for the RGCs. Such adaptive changes in the spatial summation as a function of local contrast may allow the visual system to optimize performance under changing stimulus conditions (Sceniak et al., 1999). For example, for the heavy hazy local regions of low contrast, expansion of spatial summation may produce increased sensitivity and a better detection capability for weak signals (with the sacrificing of spatial resolution) of an image. Inspired by such property of contrast-dependent change in RF size and the results shown in Figure 6 for the analysis of the roles of ON and OFF systems, we expect to improve the performance of our model in the future by automatically selecting the model parameters that are spatially adaptive to the local haze related visual features.


Figure 10. Our result on an image with dense haze. (A) The hazy image. (B) The image after haze removal by the proposed method.

Undoubtedly, another future direction is to introduce the information processing mechanisms of the higher visual cortexes, considering that the scene depth is a quite important cue that is closely related to the thickness of haze. Based on the binocular disparity, the human visual system can easily extract the scene depth (Haefner and Cumming, 2007), which is expected to be especially helpful for the observers to estimate and remove the haze before the objects, though it is very difficult for the visual system to exactly extract the binocular disparity from the scenes with heterogeneously distributed thick haze.

Author Contributions

XZ conducted the experiments and wrote the paper. SG conducted the experiments. CL reviewed and edited the manuscript. YL designed and supervised the study and reviewed and edited the manuscript.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.


This work was supported by the Major State Basic Research Program (#2013CB329401) and the Natural Science Foundations of China (#91420105, #61375115).


Alahi, A., Ortiz, R., and Vandergheynst, P. (2012). “Freak: fast retina keypoint,” in IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Providence, RI), 510–517.

Ancuti, C. O., Ancuti, C., Hermans, C., and Bekaert, P. (2011). “A fast semi-inverse approach to detect and remove the haze from a single image,” in Computer Vision–ACCV (Queenstown), 501–514.

Google Scholar

Buzás, P., Blessing, E. M., Szmajda, B. A., and Martin, P. R. (2006). Specificity of M and L cone inputs to receptive fields in the parvocellular pathway: random wiring with functional bias. J. Neurosci. 26, 11148–11161. doi: 10.1523/JNEUROSCI.3237-06.2006

PubMed Abstract | CrossRef Full Text | Google Scholar

Caraffa, L., and Tarel, J.-P. (2013). “Markov random field model for single image defogging,” in IEEE Intelligent Vehicles Symposium (IV) (Gold Coast, QLD), 994–999.

Google Scholar

Chen, K., Song, X. M., and Li, C. Y. (2013). Contrast-dependent variations in the excitatory classical receptive field and suppressive nonclassical receptive field of cat primary visual cortex. Cereb. Cortex 23, 283–292. doi: 10.1093/cercor/bhs012

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, S., and Li, W. (2012). A color-coding amacrine cell may provide a blue-off signal in a mammalian retina. Nat. Neurosci. 15, 954–956. doi: 10.1038/nn.3128

PubMed Abstract | CrossRef Full Text | Google Scholar

Conway, B. R., Chatterjee, S., Field, G. D., Horwitz, G. D., Johnson, E. N., Koida, K., et al. (2010). Advances in color science: from retina to behavior. J. Neurosci. 30, 14955–14963. doi: 10.1523/JNEUROSCI.4348-10.2010

PubMed Abstract | CrossRef Full Text | Google Scholar

Crook, J. D., Manookin, M. B., Packer, O. S., and Dacey, D. M. (2011). Horizontal cell feedback without cone type-selective inhibition mediates “red–green” color opponency in midget ganglion cells of the primate retina. J. Neurosci. 31, 1762–1772. doi: 10.1523/JNEUROSCI.4385-10.2011

PubMed Abstract | CrossRef Full Text | Google Scholar

Enroth-Cugell, C., and Robson, J. G. (1966). The contrast sensitivity of retinal ganglion cells of the cat. J. Physiol. 187, 517–552. doi: 10.1113/jphysiol.1966.sp008107

PubMed Abstract | CrossRef Full Text | Google Scholar

Euler, T., Haverkamp, S., Schubert, T., and Baden, T. (2014). Retinal bipolar cells: elementary building blocks of vision. Nat. Rev. Neurosci. 15, 507–519. doi: 10.1038/nrn3783

PubMed Abstract | CrossRef Full Text | Google Scholar

Fattal, R. (2008). “Single image dehazing,” in ACM Transactions on Graphics (TOG): ACM (Los Angeles, CA), 72. doi: 10.1145/1399504.1360671

CrossRef Full Text | Google Scholar

Field, G. D., Gauthier, J. L., Sher, A., Greschner, M., Machado, T. A., Jepson, L. H., et al. (2010). Functional connectivity in the retina at the resolution of photoreceptors. Nature 467, 673–677. doi: 10.1038/nature09424

PubMed Abstract | CrossRef Full Text | Google Scholar

Foster, D. H. (2011). Color constancy. Vision Res. 51, 674–700. doi: 10.1016/j.visres.2010.09.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Gao, S. B., Yang, K. F., Li, C. Y., and Li, Y. J. (2013). “A color constancy model with double-opponency mechanisms,” in IEEE International Conference on Computer Vision (ICCV) (Sydney), 929–936.

Google Scholar

Gao, S. B., Yang, K. F., Li, C. Y., and Li, Y. J. (2015). Color constancy using double-opponency. IEEE Trans. Pattern Anal. Mach. Intell. 37, 1973–1985. doi: 10.1109/TPAMI.2015.2396053

PubMed Abstract | CrossRef Full Text | Google Scholar

Ghassemian, H. (2001). “A retina based multi-resolution image-fusion,” in Geoscience and Remote Sensing Symposium, 2001. IGARSS'01. IEEE 2001 International (Sydney), 709–711.

Google Scholar

Gibson, K. B., Võ, D. T., and Nguyen, T. Q. (2012). An investigation of dehazing effects on image and video coding. IEEE Trans. Image Process. 21, 662–673. doi: 10.1109/TIP.2011.2166968

PubMed Abstract | CrossRef Full Text | Google Scholar

Gollisch, T., and Meister, M. (2010). Eye smarter than scientists believed: neural computations in circuits of the retina. Neuron 65, 150–164. doi: 10.1016/j.neuron.2009.12.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Haefner, R. M., and Cumming, B. G. (2007). Adaptation to natural binocular disparities in primate V1 explained by a generalized energy model. Neuron 57, 147–158. doi: 10.1016/j.neuron.2007.10.042

PubMed Abstract | CrossRef Full Text | Google Scholar

Hautière, N., Tarel, J.-P., and Aubert, D. (2007). “Towards fog-free in-vehicle vision systems through contrast restoration,” in IEEE Conference on Computer Vision and Pattern Recognition, CVPR (Minneapolis, MN), 1–8.

Google Scholar

He, K., Sun, J., and Tang, X. (2009). “Single image haze removal using dark channel prior,” in IEEE Conference on Computer Vision and Pattern Recognition, CVPR (Miami, FL), 1956–1963.

Google Scholar

He, K., Sun, J., and Tang, X. (2011). Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 33, 2341–2353. doi: 10.1109/TPAMI.2010.168

PubMed Abstract | CrossRef Full Text | Google Scholar

Hirschmuller, H., and Scharstein, D. (2007). “Evaluation of cost functions for stereo matching,” in IEEE Conference on Computer Vision and Pattern Recognition, CVPR'07 (Minneapolis, MN), 1–8.

Google Scholar

Jobson, D. J., Rahman, Z.-U., and Woodell, G. A. (1997). A multiscale retinex for bridging the gap between color images and the human observation of scenes. IEEE Trans. Image Process. 6, 965–976. doi: 10.1109/83.597272

PubMed Abstract | CrossRef Full Text | Google Scholar

Joselevitch, C. (2008). Human retinal circuitry and physiology. Psychol. Neurosci. 1, 141–165. doi: 10.3922/j.psns.2008.2.008

CrossRef Full Text | Google Scholar

Kaneko, A., and Tachibana, M. (1983). Double color-opponent receptive fields of carp bipolar cells. Vision Res. 23, 381–388. doi: 10.1016/0042-6989(83)90085-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Kopf, J., Neubert, B., Chen, B., Cohen, M., Cohen-Or, D., Deussen, O., et al. (2008). “Deep photo: Model-based photograph enhancement and viewing,” in ACM Transactions on Graphics (TOG) (Providence, RI).

Google Scholar

Koschmieder, H. (1925). Theorie der Horizontalen Sichtweite: Kontrast und Sichtweite. Munich: Keim & Nemnich.

Kratz, L., and Nishino, K. (2009). “Factorizing scene albedo and depth from a single foggy image,” in IEEE 12th International Conference on Computer Vision (Kyoto), 1701–1708. doi: 10.1109/iccv.2009.5459382

CrossRef Full Text | Google Scholar

Land, E. H. (1986). Recent advances in retinex theory. Vision Res. 26, 7–21. doi: 10.1016/0042-6989(86)90067-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Land, E. H., and McCann, J. (1971). Lightness and retinex theory. J. Opt. Soc. Am. 61, 1–11. doi: 10.1364/JOSA.61.000001

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, B. B., Martin, P. R., and Grünert, U. (2010). Retinal connectivity and primate vision. Prog. Retin. Eye Res. 29, 622–639. doi: 10.1016/j.preteyeres.2010.08.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, B. B., Shapley, R. M., Hawken, M. J., and Sun, H. (2012). Spatial distributions of cone inputs to cells of the parvocellular pathway investigated with cone-isolating gratings. J. Opt. Soc. Am. A Opt.Image Sci. Vis. 29, A223–A232. doi: 10.1364/JOSAA.29.00A223

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, C. Y. (1996). Integration fields beyond the classical receptive field: organization and functional properties. News Physiol. Sci. 11, 181–186.

Google Scholar

Li, C. Y., and He, Z. J. (1987). Effects of patterned backgrounds on responses of lateral geniculate neurons in cat. Exp. Brain Res. 67, 16–26. doi: 10.1007/BF00269448

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, C. Y., and Li, W. (1994). Extensive integration field beyond the classical receptive field of cat's striate cortical neurons—classification and tuning properties. Vision Res. 34, 2337–2355. doi: 10.1016/0042-6989(94)90280-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, C. Y., Pei, X., Zhow, Y. X., and von Mitzlaff, H. C. (1991). Role of the extensive area outside the X-cell receptive field in brightness information transmission. Vision Res. 31, 1529–1540. doi: 10.1016/0042-6989(91)90130-W

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, C. Y., and Qiu, F. T. (1994). Simulation of spatial transfer properties of cat retinal ganglion cell. Acta Biophys. Sin. 11, 395–400.

Google Scholar

Li, C. Y., Zhou, Y. X., Pei, X., Qiu, F. T., Tang, C. Q., and Xu, X. Z. (1992). Extensive disinhibitory region beyond the classical receptive field of cat retinal ganglion cells. Vision Res. 32, 219–228. doi: 10.1016/0042-6989(92)90131-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Martin, P. R., Lee, B. B., White, A. J., Solomon, S. G., and Rüttiger, L. (2001). Chromatic sensitivity of ganglion cells in the peripheral primate retina. Nature 410, 933–936. doi: 10.1038/35073587

PubMed Abstract | CrossRef Full Text | Google Scholar

Masland, R. H. (2012). The neuronal organization of the retina. Neuron 76, 266–280. doi: 10.1016/j.neuron.2012.10.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Meng, G., Wang, Y., Duan, J., Xiang, S., and Pan, C. (2013). “Efficient image dehazing with boundary constraint and contextual regularization,” in IEEE International Conference on Computer Vision (ICCV) (Sydney), 617–624.

Google Scholar

Mills, S. L., Tian, L.-M., Hoshi, H., Whitaker, C. M., and Massey, S. C. (2014). Three distinct blue-green color pathways in a mammalian retina. J. Neurosci. 34, 1760–1768. doi: 10.1523/JNEUROSCI.3901-13.2014

PubMed Abstract | CrossRef Full Text | Google Scholar

Nair, D., Kumar, P. A., and Sankaran, P. (2014). “An effective surround filter for image dehazing,” in Proceedings of the 2014 International Conference on Interdisciplinary Advances in Applied Computing (New York, NY).

Google Scholar

Namer, E., Shwartz, S., and Schechner, Y. Y. (2009). Skyless polarimetric calibration and visibility enhancement. Opt. Express 17, 472–493. doi: 10.1364/OE.17.000472

PubMed Abstract | CrossRef Full Text | Google Scholar

Narasimhan, S. G., and Nayar, S. K. (2000). “Chromatic framework for vision in bad weather,” in Proceedings. IEEE Conference on Computer Vision and Pattern Recognition, CVPR (Hilton Head Island, SC), 598–605.

Google Scholar

Narasimhan, S. G., and Nayar, S. K. (2003a). Contrast restoration of weather degraded images. IEEE Trans. Patt. Anal. Mach. Intell. 25, 713–724. doi: 10.1109/TPAMI.2003.1201821

CrossRef Full Text | Google Scholar

Narasimhan, S. G., and Nayar, S. K. (2003b). “Interactive (de) weathering of an image using physical models,” in IEEE Workshop on Color and Photometric Methods in Computer Vision (Nice).

Google Scholar

Nayar, S. K., and Narasimhan, S. G. (1999). “Vision in bad weather,” in The Proceedings of the Seventh IEEE International Conference on Computer Vision (Kerkyra), 820–827.

Google Scholar

Nolt, M. J., Kumbhani, R. D., and Palmer, L. A. (2004). Contrast-dependent spatial summation in the lateral geniculate nucleus and retina of the cat. J. Neurophysiol. 92, 1708–1717. doi: 10.1152/jn.00176.2004

PubMed Abstract | CrossRef Full Text | Google Scholar

Qiu, F. T., and Li, C. Y. (1995). Mathematical simulation of disinhibitory properties of concentric receptive field. Acta Biophys. Sin. 11, 214–220.

Rahman, Z.-U., Jobson, D. J., and Woodell, G. A. (2004). Retinex processing for automatic image enhancement. J. Electron. Imaging 13, 100–110. doi: 10.1117/1.1636183

CrossRef Full Text | Google Scholar

Rajput, G. S., and Rahman, Z. U. (2008). “Hazard detection on runways using image processing techniques,” in SPIE Defense and Security Symposium: International Society for Optics and Photonics), 69570D–69512.

Google Scholar

Ratliff, C. P., Borghuis, B. G., Kao, Y.-H., Sterling, P., and Balasubramanian, V. (2010). Retina is structured to process an excess of darkness in natural scenes. Proc. Natl. Acad. Sci. U.S.A. 107, 17368–17373. doi: 10.1073/pnas.1005846107

PubMed Abstract | CrossRef Full Text | Google Scholar

Reid, R. C., and Shapley, R. M. (2002). Space and time maps of cone photoreceptor signals in macaque lateral geniculate nucleus. J. Neurosci. 22, 6158–6175.

PubMed Abstract | Google Scholar

Rodieck, R. W., and Stone, J. (1965). Response of cat retinal ganglion cells to moving visual patterns. J. Neurophysiol. 28, 819–832.

PubMed Abstract | Google Scholar

Sceniak, M. P., Ringach, D. L., Hawken, M. J., and Shapley, R. (1999). Contrast's effect on spatial summation by macaque V1 neurons. Nat. Neurosci. 2, 733–739. doi: 10.1038/11197

PubMed Abstract | CrossRef Full Text | Google Scholar

Scharstein, D., and Pal, C. (2007). “Learning conditional random fields for stereo,” in IEEE Conference on Computer Vision and Pattern Recognition, CVPR'07 (Minneapolis, MN), 1–8.

Google Scholar

Scharstein, D., and Szeliski, R. (2003). “High-accuracy stereo depth maps using structured light,” in Proceedings. 2003. IEEE Computer Society Conference on Computer Vision and Pattern Recognition (Madison, WI), I-195–I-202.

Google Scholar

Schaul, L., Fredembach, C., and Süsstrunk, S. (2009). “Color image dehazing using the near-infrared,” in IEEE International Conference on Image ProcessingCiteseer (Cairo), 1629–1632.

Google Scholar

Schechner, Y. Y., Narasimhan, S. G., and Nayar, S. K. (2001). “Instant dehazing of images using polarization,” in Proceedings of the 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2001. CVPR 2001 (Kauai), I-325–I-332.

Google Scholar

Schiller, P. H. (2010). Parallel information processing channels created in the retina. Proc. Natl. Acad. Sci. U.S.A. 107, 17087–17094. doi: 10.1073/pnas.1011782107

PubMed Abstract | CrossRef Full Text | Google Scholar

Shwartz, S., Namer, E., and Schechner, Y. Y. (2006). “Blind haze separation,” in Computer Vision and Pattern Recognition, 2006. CVPR 2006. IEEE Computer Society Conference on, 1984–1991. doi: 10.1109/cvpr.2006.71

CrossRef Full Text | Google Scholar

Spitzer, H., and Barkan, Y. (2005). Computational adaptation model and its predictions for color induction of first and second orders. Vision Res. 45, 3323–3342. doi: 10.1016/j.visres.2005.08.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Spitzer, H., and Semo, S. (2002). Color constancy: a biological model and its application for still and video images. Patt. Recogn. 35, 1645–1659. doi: 10.1016/S0031-3203(01)00160-1

CrossRef Full Text | Google Scholar

Sun, H., Smithson, H. E., Zaidi, Q., and Lee, B. B. (2006). Specificity of cone inputs to macaque retinal ganglion cells. J. Neurophysiol. 95, 837–849. doi: 10.1152/jn.00714.2005

PubMed Abstract | CrossRef Full Text | Google Scholar

Tan, K., and Oakley, J. P. (2000). “Enhancement of color images in poor visibility conditions,” in Image Processing, 2000. Proceedings. 2000 International Conference on (Vancouver, BC), 788–791.

Google Scholar

Tan, R. T. (2008). “Visibility in bad weather from a single image,” in Computer Vision and Pattern Recognition, 2008. CVPR 2008. IEEE Conference on (Anchorage, AK), 1–8. doi: 10.1109/cvpr.2008.4587643

CrossRef Full Text | Google Scholar

Tang, K., Yang, J., and Wang, J. (2014). “Investigating haze-relevant features in a learning framework for image dehazing,” in Computer Vision and Pattern Recognition, 2014. CVPR 2014. IEEE Conference on (Columbus, OH), 2995–3002. doi: 10.1109/cvpr.2014.383

CrossRef Full Text | Google Scholar

Tarel, J.-P., and Hautiere, N. (2009). “Fast visibility restoration from a single color or gray level image,” in Computer Vision, 2009 IEEE 12th International Conference on (Kyoto), 2201–2208. doi: 10.1109/iccv.2009.5459251

CrossRef Full Text | Google Scholar

Tarel, J.-P., Hautière, N., Caraffa, L., Cord, A., Halmaoui, H., and Gruyer, D. (2012). Vision enhancement in homogeneous and heterogeneous fog. IEEE Intell. Transp. Syst. Mag. 4, 6–20. doi: 10.1109/MITS.2012.2189969

CrossRef Full Text | Google Scholar

Thoreson, W. B., and Mangel, S. C. (2012). Lateral interactions in the outer retina. Prog. Retin. Eye Res. 31, 407–441. doi: 10.1016/j.preteyeres.2012.04.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Treibitz, T., and Schechner, Y. Y. (2009). “Polarization: beneficial for visibility enhancement?,” in Computer Vision and Pattern Recognition, 2009. CVPR 2009. IEEE Conference on (Miami, FL), 525–532. doi: 10.1109/cvpr.2009.5206551

CrossRef Full Text | Google Scholar

Vu, N.-S., and Caplier, A. (2009). “Illumination-robust face recognition using retina modeling,” in Image Processing (ICIP), 2009 16th IEEE International Conference on (Cairo), 3289–3292.

Google Scholar

Werblin, F. S., and Dowling, J. E. (1969a). Organization of the retina of the mudpuppy, necturus macubsus. II. Intracellular recording. J. Neurophysiol. 32, 339.

PubMed Abstract | Google Scholar

Werblin, F. S., and Dowling, J. E. (1969b). Organization of the retina of the mudpuppy, Necturus maculosus. II. Intracellular recording. J. Neurophysiol. 32, 339–355.

PubMed Abstract | Google Scholar

Woodell, G., Jobson, D. J., Rahman, Z.-u., and Hines, G. (2005). “Enhancement of imagery in poor visibility conditions,” in Defense and Security: International Society for Optics and Photonics (Orlando, FL), 673–683.

Google Scholar

Xie, B., Guo, F., and Cai, Z. (2010). “Improved single image dehazing using dark channel prior and multi-scale Retinex,” in Intelligent System Design and Engineering Application (ISDEA), 2010 International Conference on (Hong Kong), 848–851. doi: 10.1109/isdea.2010.141

CrossRef Full Text | Google Scholar

Yang, K. F., Gao, S. B., Guo, C. F., Li, C. Y., and Li, Y. J. (2015). Boundary detection using double-opponency and spatial sparseness constraint. IEEE Trans. Image Process. 24, 2565–2578. doi: 10.1109/TIP.2015.2425538

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, K. F., Gao, S. B., Li, C. Y., and Li, Y. J. (2013). “Efficient color boundary detection with color-opponent mechanisms,” in Computer Vision and Pattern Recognition (CVPR), 2013 IEEE Conference on (Portland, OR), 2810–2817. doi: 10.1109/cvpr.2013.362

CrossRef Full Text | Google Scholar

Yu, J., Xiao, C., and Li, D. (2010). “Physics-based fast single image fog removal,” in Signal Processing (ICSP), 2010 IEEE 10th International Conference on (Beijing), 1048–1052. doi: 10.1109/icosp.2010.5655901

CrossRef Full Text | Google Scholar

Zhou, J., and Zhou, F. (2013). “Single image dehazing motivated by Retinex theory,” in Instrumentation and Measurement, Sensor Network and Automation (IMSNA), 2013 2nd International Symposium on (Toronto, ON), 243–247. doi: 10.1109/imsna.2013.6743260

CrossRef Full Text | Google Scholar

Keywords: haze removal, retina inspired model, retinal ganglion cell, non-classical receptive field, disinhibitory effect

Citation: Zhang X-S, Gao S-B, Li C-Y and Li Y-J (2015) A Retina Inspired Model for Enhancing Visibility of Hazy Images. Front. Comput. Neurosci. 9:151. doi: 10.3389/fncom.2015.00151

Received: 08 October 2015; Accepted: 03 December 2015;
Published: 22 December 2015.

Edited by:

Pei-Ji Liang, Shanghai Jiao Tong University, China

Reviewed by:

Da-Hui Wang, Beijing Normal University, China
Petia D. Koprinkova-Hristova, Bulgarian Academy of Sciences, Bulgaria

Copyright © 2015 Zhang, Gao, Li and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yong-Jie Li,