# Analog Image Modeling for 3D Single Image Super Resolution and Pansharpening

^{1}Department of Mathematics, Applied Mathematics and Statistics, Case Western Reserve University, Cleveland, OH, United States^{2}German Aerospace Center (DLR), Wessling, Germany^{3}Department of Aerospace and Geodesy, Technical University of Munich, Munich, Germany

Image super-resolution is an image reconstruction technique which attempts to reconstruct a high resolution image from one or more under-sampled low-resolution images of the same scene. High resolution images aid in analysis and inference in a multitude of digital imaging applications. However, due to limited accessibility to high-resolution imaging systems, a need arises for alternative measures to obtain the desired results. We propose a three-dimensional single image model to improve image resolution by estimating the analog image intensity function. In recent literature, it has been shown that image patches can be represented by a linear combination of appropriately chosen basis functions. We assume that the underlying analog image consists of smooth and edge components that can be approximated using a reproducible kernel Hilbert space function and the Heaviside function, respectively. We also extend the proposed method to pansharpening, a technology to fuse a high resolution panchromatic image with a low resolution multi-spectral image for a high resolution multi-spectral image. Various numerical results of the proposed formulation indicate competitive performance when compared to some state-of-the-art algorithms.

## 1. Introduction

During image acquisition, we often loose resolution due to the limited density of the imaging sensors and the blurring of the acquisition lens. Image super resolution, a post process to increase image resolution, is an efficient way to achieve a resolution beyond what the hardware offers. A detailed overview of this area is outlined in Farsiu et al. [1], Yang et al. [2], and Park et al. [3]. Current techniques of solving this ill-posed problem are often insufficient for many real world applications [4, 5]. We propose an analog modeling based three-dimensional (3D) color image super resolution and pansharpening method.

In recent literature, image super-resolution (SR) methods typically fall into three categories: interpolation based methods, learning based methods, and constrained reconstruction methods. Interpolation based methods [6, 7] such as nearest-neighbor, bilinear and bicubic interpolation use analytical priors to recover a continuous-time signal from the discrete pixels. They work very well for smooth regions and are simple to compute. In dictionary learning based models, the missing high frequency detail is recovered from a sparse linear combination of dictionary atoms trained from patches of the given low-resolution (LR) image [8, 9] or from an external database of patches [10–12]. Dictionary learning based methods explore the self-similarities of the image and are robust, albeit rather slow training time of the dictionary atoms. Models based on constrained reconstruction use the underlying nature of the LR image to define prior constraints for the target HR image. Sharper structures can be recovered using these models. Priors common to these techniques include total variation priors [13–15], gradient priors [16, 17] and priors based on edge statistics [18]. The proposed method falls in this category. Another category is deep neural network based approach that uses a large amount of training data to obtain super results [19–22].

Increased accessibility to satellite data has led to a wide range of applications in multispectral (MS) SR. These applications include target detection, scene classification, and change detection. A multitude of algorithms have been presented in recent literature to address this task. However, the performance of many of these algorithms are data dependent with respect to existing quantitative assessment measures. With the increasing advancements in technology, there have been varying modes of image acquisition systems. In remote sensing, satellites carry sensors that acquire images of a scene in multiple channels across the electromagnetic spectrum. This gives rise to what is referred to as *multispectral imaging*. Due to high cost overheads in developing sensors that can faithfully acquire scene data with high spectral and spatial resolution, the spatial resolution in the data acquisition process is often compromised. The low-spatial resolution MS images usually consist of images across 4–8 spectral bands: red, green, blue, as well as near infrared bands (provide Landsat, IKONOS, etc. specifications).

Most satellites also acquire a high-spatial resolution image known as a *panchromatic image*. The technique of using panchromatic (PAN) to compensate for the reduction in spatial-resolution of the MS images is referred to as *pansharpening*. The success of pansharpening relies on the extent one can explore the structural similarity between the PAN and the high-resolution (HR) MS image [23]. In recent years, pansharpening methods fall under three main categories: component substitituion (CS), multiresolution analysis (MRA), and variational methods. Detailed surveys can be found in Loncan et al. [24] and Vivone et al. [25].

The main assumption of CS methods is that the spatial and spectral information of the high spectral resolution image can be separated in some space [24]. Examples of CS methods include principal component analysis (PCA) [26, 27], intensity-hue-saturation (IHS) methods [28–30], Brovey method [31], and Gram-Schmidt (GS) based methods [32]. These methods enhance the spatial resolution of the LR MS image at the expense of some loss in the spectral detail. Compared to the CS methods, MRA methods have better temporal coherence between the PAN and the enhanced MS image. MRA methods yield results with better spectral consistency at the expense of designing complex filters to mimic the reduction in the resolution by the sensors. Spatial details are injected into the MS data after a multi-scale decomposition of the PAN [24]. Examples of MRA methods include high-pass filtering (HPF) [30], smooth filter-based intensity modulation (SFIM) [33], generalized Laplacian pyramid (GLP) methods [34, 35], additive “á trous” wavelet transform (ATWT) methods [23, 36] and additive wavelet luminance proportional (AWLP) [37]. Variational methods are more robust compared to the CS and MRA techniques, albeit rather slow in comparison [24]. A common tool used by these approaches is regularization. This is used to promote some structural similarity between the LR MS, HR MS, and PAN images [24, 38, 39]. Among these methods include models based on the Bayesian framework and matrix factorization methods [24, 38, 39].

In this paper, we extend the key ideas in the SR model in Deng et al. [40] to 3D framework. We then test the performance of the model with single image super-resolution and pansharpening.

### 1.1. Related Work

In Deng et al. [40], a 2D image defined on a continuous domain [0, 1] × [0, 1] is assumed to have a density function that is the sum of two parts, reproducing kernel Hilbert space (RKHS) and Heaviside functions to study single image super-resolution (SISR) whereby the problem is cast as an image intensity function estimation problem. They assume that the smooth components of the image belong to a special Hilbert space called RKHS, and can be spanned by a basis based on an approximation by spline interpolation [41]. The Heaviside function is used to model the edges and discontinuous information in the input image. Using a patch-based approach, the intensity information of the LR input image patches are defined on a coarser grid to estimate the coefficients of the basis functions. To obtain the enhanced resolution result, they utilize the obtained coefficients and an approximation of the basis functions on a finer grid. Furthermore, to recover more high frequency information the authors employ an iterative back projection method [42] after the HR image has been obtained. Color images were tested in Deng et al. [40], however the analog modeling was done the luminance channel only after transforming the color space. This RKHS+Heaviside method is extended to pansharpening in Deng et al. [43]. In the pansharpening framework the different bands share the same basis but with different coefficients patch by patch, and channel by channel. A panchromatic image is used to guide the model and it is assumed to be a linear combination of the multiple bands.

### 1.2. Contribution

Using the same basis as in Deng et al. [40], we present a three-dimensional SR algorithm for true color images based on continuous/analog modeling. We also extend it to pansharpening. Instead of estimating the basis coefficients patch by patch and band by band, we cluster patches and conduct computation cluster by cluster. Similarity of patches within one cluster leads to some natural regularity. We jointly optimize all the coefficients for all the bands of the model rather than optimizing for each independent band. Furthermore, we use clustering techniques to improve the structural coherence of the desired results in the optimization model. Spatially, we divide the images into small overlapping patches and cluster these patches classes using k-means clustering. To improve the spectral coherence of the results, we cluster the image bands into groups based on correlation statistics and the perform the optimization on each of the groups obtained.

In this paper, we use the RKHS approximations to model the smooth components of the image whiles preserving the non-smooth components by the approximated Heaviside function. For the pansharpening experiments, We also preprocess the data using fast state-of-the-art MRA techniques [37] to incorporate some similarity in the spectral channels of the fused image. Thus we develop a continuous 3D modeling framework for multispectral pansharpening.

The experiments that follow show that these contributions yield better enhanced resolution results with faster convergence speeds at all scales.

The paper is organized as follows. In section 2, we review the regular functions used in decomposing the image. The proposed model is described in section 3 with numerical results outlined in section 4.

## 2. Background Review

In this section, we review the mathematical functions used in the formulation of our model. Similar as in Deng et al. [40], we model image intensity function defined on a small image patch as a linear combination of reproducing Kernel Hilbert function (RKHS) and approximated Heaviside function (AHF) which models the smooth and edge components, respectively.

### 2.1. Reproducible Kernel Hilbert Spaces

RKHS can be used to define feature spaces to help compare objects that have complex structures and are hard to distinguish in the original feature space. Wahba [41] proposed polynomial spline models for smoothing problems. Our proposed model is based on this approach using the Taylor series expansion. We review these methods in the following subsections.

#### 2.1.1. Signal Smoothing Using Splines

Let ${G}$ represent a family of functions on [0,1] with continuous derivatives up to order (*m* − 1),

By Taylor's theorem with remainder for $f\in {G}$, we may write

for some *u*, where (*x*)_{+} = *x* for *x* ≥ 0 and (*x*)_{+} = 0 otherwise. Let

and ${{H}}_{0}\text{}=\text{}\mathrm{\text{span}}\text{}{\left\{{\varphi}_{i}\right\}}_{1\le i\le m}$ endowed with norm ${\Vert}\varphi {{\Vert}}^{2}=\sum _{\nu =0}^{m-1}{\left[({D}^{\nu}\varphi )(0)\right]}^{2}$, then ${D}^{(m)}({{H}}_{0})=0$, where *D*^{m} denotes the *m*th derivative. It can be shown that a reproducing kernel exists for ${{H}}_{0}$ and is defined as

therefore for any ${f}_{0}\in {H}$, we have ${f}_{0}(t)=\sum _{\nu =1}^{m}{d}_{\nu}{\varphi}_{\nu}(t).$

Let ${{B}}_{m}=\left\{f:f\in {{C}}^{m-1}\left[0,1\right],{f}^{\nu}(0)=0,\nu =0,1,\dots ,m-1\right\}$ and define

The space

with norm $\Vert f{\Vert}^{2}={\displaystyle {\int}_{0}^{1}({f}^{(m)}(}t){)}^{2}dt$ can be shown to be an RKHS with reproducing kernel

so that for any ${f}_{1}\in {{H}}_{1}$, we have ${f}_{1}(t)=\sum _{i=1}^{n}{c}_{i}{\xi}_{i},$ where ${\xi}_{i}(\xb7)={R}^{1}({s}_{i},\xb7).$ It follows from the properties of the RKHS, we can construct a direct sum space ${{G}}_{m}={{H}}_{0}+{{H}}_{1}$ since

The reproducing kernel for ${{G}}_{m}$ can be shown [41] to be

with norm

therefore for any $f\in {{G}}_{m}$, we have *f* = *f*_{0} + *f*_{1} with ${f}_{0}\in {{H}}_{0},{f}_{1}\in {{H}}_{1}$ and

Let $d={({d}_{1},\dots {d}_{m})}^{T}$ and $c={({c}_{1},\dots ,{c}_{n})}^{T}$ be coefficient vectors, $f={\left(f({t}_{1}),\mathrm{...},f({t}_{p})\right)}^{T}$ represent the discretization of *f* on discrete grids ${t}_{j},j=1,2,\cdots {n}_{t},T\in {\mathbb{R}}^{n\times m},\Sigma \in {\mathbb{R}}^{n\times n}$ with **T**_{jν} = ϕ_{ν}(*t*_{j}), **Σ**_{ji} = ξ_{i}(*t*_{j}), then * f* =

*+*

**Td****Σ**

*. Given the observation data*

**c**where * η* denotes additive Gaussian noise. To find

*from*

**f***, we need to estimate the coefficients*

**g***and*

**d***. Following Wahba [41], the coefficients*

**c***are estimated from the discrete measurements*

**d, c***by*

**g**where the term μ**c**^{T}**Σ*** c* is a non-smoothness penalty. Closed form solutions for

*and*

**c***can be obtained. The parameter*

**d***m*controls the total degree of the polynomials used to define

*. For example, when*

**T***m*= 3 the basis functions are given as

*ϕ*

_{1}(

*x*) = 1,

*ϕ*

_{2}(

*x*) =

*x*and ${\varphi}_{3}(x)={x}^{2}/2$.

#### 2.1.2. Image Smoothing Using Splines

Extending the one-dimensional spline model to two-dimensional data, we consider a similar discretization and set up a model similar to that defined in (11). Let $f={\left(f({t}_{1}),\dots ,f({t}_{{n}_{t}})\right)}^{T}$, *t*_{i} ∈ [0, 1] × [0, 1] be a discretization for an observed noisy image. From Wahba [41], an estimate of * f* can be obtained from the additive noise data model in (10) for images by minimizing the following

where *m* defines the order of the partial derivatives in ${{L}}_{2}([0,1]\times [0,1])$, and with the penalty function defined by

The null space of *J*_{m} is the $M=\left(\begin{array}{c}2+m-1\\ 2\end{array}\right)$-dimensional space spanned by polynomials in two variables with total degree at most *m* − 1. In this paper, and we choose *m* = 3 so that *M* = 6 and the null space of *J*_{m} is spanned by monomials ϕ_{1}, ϕ_{2}, …, ϕ_{6} given by

Duchon [44] proved that a unique minimizer *f*_{μ} exists for (12) with representation

where *E*_{m} is defined as

and ${\theta}_{m,d}=\frac{{(-1)}^{d/2+m+1}}{{2}^{2m-1}{\pi}^{d/2}(m-1)!(m-d/2)!}$.

Based on the work of Duchon [44] and Meinguet [45] we can rewrite (12) to find minimizers * c* and

*by*

**d**where $T\in {\mathbb{R}}^{{n}_{t}\times M}$, with **T**_{i, ν} = ϕ_{ν}(*t*_{i}) and $K\in {\mathbb{R}}^{{n}_{t}\times {n}_{t}}$ with **K**_{i, j} = *E*_{m}(*t*_{i}, *t*_{j}). *E*_{m}(*s, t*) is the two-dimensional equivalent of ξ_{i}(*t*) in the one-dimensional case. After obtaining the coefficients, we compute * f* using the relationship

*=*

**f***+*

**Td***.*

**Kc**### 2.2. Approximated Heaviside Function

The one-dimensional Heaviside step function is defined as

Due to the singularity at *x* = 0, we approximate ϕ by,

We refer to this as the approximated Heaviside function (AHF), an approximation to ϕ(*x*) as ε → 0. The variable ε controls the smoothness of the approximation [40]. A two-dimensional variation of the AHF is given by

with variables θ and *c* determining the rotations and locations of the edges as shown in Figure 1, we show the 2D and 3D surface view of the AHF at two different pairs of θ and *c*. Kainen *et al*. [46] proved that a function $f\in {{L}}_{2}(\left[0,1\right]\times \left[0,1\right])$ can be approximately represented by the weighted linear combination of approximated Heaviside functions:

**Figure 1**. Illustrating the surfaces corresponding to the approximated Heaviside functions for varying pairs of θ and *c*.

We define $\text{\Psi}\in {\mathbb{R}}^{{n}_{t}\times {n}_{\theta}}$ with ${\Psi}_{i,j}=\psi \left(\left(\begin{array}{c}\mathrm{cos}{\theta}_{j}\\ \mathrm{sin}{\theta}_{j}\end{array}\right)\xb7\left(\begin{array}{c}{x}_{i}\\ {y}_{i}\end{array}\right)+{c}_{j}\right),\text{\hspace{0.17em}\hspace{0.17em}}{t}_{i}=({x}_{i},{y}_{i}),\text{and\hspace{0.17em}}{n}_{\theta}$ is the number of rotations considered. Suppose * f, g* are the vectorized high and low resolution discretizations of

*f*, respectively. The vectorized high resolution image $f\in {\mathbb{R}}^{{n}_{1}{n}_{2}}$ can be obtained by ${f}_{i}=f({t}_{i}^{h})$ with ${t}_{i}^{h}=({x}_{i}^{h},{y}_{i}^{h})$, ${x}_{i}^{h}\in \{0,\frac{1}{{n}_{1}-1},\frac{2}{{n}_{2}-1},\dots ,1\}$, ${y}_{i}^{h}\in \{0,\frac{1}{{n}_{2}-1},\frac{2}{{n}_{2}-1},\dots ,1\}$ on a finer grid. Similarly, vectorized the low-resolution discretization $g\in {\mathbb{R}}^{{m}_{1}{m}_{2}}$ can be obtained by ${g}_{i}=f({t}_{i}^{\ell})$ with ${t}_{i}^{\ell}=({x}_{i}^{\ell},{y}_{i}^{\ell})$, ${x}_{i}^{\ell}\in \{0,\frac{1}{{m}_{1}-1},\frac{2}{{m}_{1}-1},\dots ,1\}$, and ${y}_{i}^{\ell}\in \{0,\frac{1}{{m}_{2}-1},\frac{2}{{m}_{2}-1},\dots ,1\}$ on a coarse grid. Now, assuming that the underlying analog image intensity function is approximated by the sum of a RKHS function and the variations of AHF with ${T}^{\ell}\in {\mathbb{R}}^{{m}_{1}{m}_{2}\times M}$, ${K}^{\ell}\in {\mathbb{R}}^{{m}_{1}{m}_{2}\times {m}_{1}{m}_{2}}$, and ${\text{\Psi}}^{\ell}\in {\mathbb{R}}^{{m}_{1}{m}_{2}\times {n}_{\theta}}$, we have

Thus, given the low resolution input * g*, the coefficient vectors $c\in {\mathbb{R}}^{{m}_{1}{m}_{2}}$,

*∈ℝ*

**d**^{M}and $e\in {\mathbb{R}}^{{n}_{\theta}}$ for the residual, smooth and edge components are obtained by solving the following minimization problem:

where * B* is the identity matrix in the absence of blur,

*is the downsampling operator, and the superscripts*

**S***h*and ℓ refers to a fine and course scale matrices. The smooth components of the image are modeled by the RKHS approach, while AHF caters for the edges. Since the dictionary

**Ψ**is pretty exhaustive, i.e., accounting for multiple edge orientations, it is reasonable to use the ||·||

_{1}to enforces the sparsity of as all the orientations may not be present for any given image. Once the coefficients have been obtained, we have

where ${T}^{h}\in {\mathbb{R}}^{{n}_{1}{n}_{2}\times M}$, ${K}^{\ell}\in {\mathbb{R}}^{{n}_{1}{n}_{2}\times {m}_{1}{m}_{2}}$, and ${\text{\Psi}}^{\ell}\in {\mathbb{R}}^{{n}_{1}{n}_{2}\times {n}_{\theta}}$. Next, we discuss our proposed model and examine some numerical experiments in the chapters that follow.

## 3. Proposed Models

In this section, we present our proposed models for SISR and pansharpening based on the functions defined in section 2. We propose a 3D patch-based approach that infers the HR patches from LR patches that has been grouped into classes based on their structural similarity. As a result, we impose similarity constraints within the classes so that the coefficients for neighboring patches in a group do not differ substantially. In addition, we use sparse constrained optimization techniques that simplify the minimization of the resulting energy functional by solving a series of subproblems, each with a closed form solution. Two algorithms will be discussed. The first is a 3D SR model for true color images and MS images with at most four bands. The second algorithm is an extension of our SISR model to pansharpening for MS images with at least four spectral bands. This is a two-step hybrid approach that incorporates CS and MRA methods into the first algorithm to obtain an enhanced resolution result.

### 3.1. Single Image Super-Resolution

Given a LR image $Y\in {\mathbb{R}}^{{m}_{1}\times {m}_{2}\times \lambda}$ we estimate the target HR image $\hat{X}\in {\mathbb{R}}^{{n}_{1}\times {n}_{2}\times \lambda}$ with *n*_{i} = ω*m*_{i}, where ω is a scale factor and λ is the number of bands as follows. We begin by decomposing the LR image into a set of overlapping patches ${{P}}_{Y}={\left\{{Y}_{{p}_{i}}\right\}}_{i=1}^{{N}_{p}}$, ${Y}_{{p}_{i}}\in {\mathbb{R}}^{{m}_{p}\times {m}_{p}\times \lambda}$. The size of the square patches, and the overlap between adjacent patches depend on the dimensions of the input image. We consider the use of overlapping patches to improve local consistency of the estimated image across the regions of overlap. The estimated image $\hat{X}$ will be represented by the same number of patches, *N*_{p}. Next, for a suitably chosen *k*, we use the *k-means* algorithm with careful seeding [47] to cluster the patches into *k* classes ${{P}}_{Y}^{1},\dots {{P}}_{Y}^{k}$, as shown in Figure 2. In this paper, the value of *k* is determined empirically to obtain the best result. Clustering is done so that each high resolution patch generated preserves some relationships with its neighboring patches.

Following the clustering, we consider each set of patches ${{P}}_{Y}^{i}={\{{Y}_{{p}_{j}}\}}_{j\in {I}_{i}}$, for some index set *I*_{i} with 1 ≤ *i* ≤ *k*. Let ${Y}_{(\lambda ,{I}_{i})}\in {\mathbb{R}}^{{m}_{p}{m}_{p}\times \lambda {N}_{{I}_{i}}}$ and ${\hat{X}}_{(\lambda ,{I}_{i})}\in {\mathbb{R}}^{{n}_{p}{n}_{p}\times \lambda {N}_{{I}_{i}}}$ represent the images corresponding to the set of LR and HR patches ${{P}}_{Y}^{i}$ and ${{P}}_{\hat{X}}^{i}$ defined by

where *n*_{p} = ω*m*_{p} for some scale factor ω ∈ ℕ, **y**_{i,pj} and ${\hat{x}}_{i,{p}_{j}}$ represents the vectorization of the *i*-th band of the *j*-th patch, with 1 ≤ *i* ≤ λ and *j* ∈ *I*_{i}. Here, *N*_{Ii} represents the total number of patches in the set ${{P}}_{Y}^{i}$. Using the RKHS and AHF functions as in Deng et al. [40] we can find an embedding for the high-dimensional data that is also structure preserving. We further assume that the HR and LR image patches have similar local geometry and are related by the equations

where $D\text{}=\text{}[{d}_{1},\dots {d}_{\lambda {N}_{{I}_{i}}}]\in {\mathbb{R}}^{M\times \lambda {N}_{{I}_{i}}},C=[{c}_{1},\dots {c}_{\lambda {N}_{{I}_{i}}}]\in {\mathbb{R}}^{{m}_{1}{m}_{2}\times \lambda {N}_{{I}_{i}}}$ and $E=[{e}_{1},\dots {e}_{\lambda {N}_{{I}_{i}}}]\in {\mathbb{R}}^{{n}_{\theta}\times \lambda {N}_{{I}_{i}}}$ are coefficient matrices to be determined. The matrices * T, K* and

**Ψ**represent the evaluations of the smooth, residual and edge components of the image intensity function on discrete grids. The superscript ℓ and

*h*denote the coarse and fine grids and correspond to lower and higher resolution, respectively. We obtain the fine and coarse scale matrices

*and*

**T***following the discretization models outlined in section 2.1 Similarly, we follow the discretization in section 2.2 to generate*

**K****Ψ**

^{ℓ}and

**Ψ**

^{h}. The coefficient matrices are obtained by solving the following minimization problem for each patch indexed by

*i*:

with tr(·) as the trace and adaptive weights

where * D* = [

**d**_{1}, …

**d**_{λNIi}] and μ

_{1}, μ

_{2}, μ

_{3}≥ 0. Note that $\text{exp}\left(-\frac{{\Vert}{d}_{j}-{d}_{k}{{\Vert}}_{2}^{2}}{\sigma}\right){\Vert}{d}_{j}-{d}_{k}{{\Vert}}_{2}^{2}$ is a unimodal function of ${\Vert}{d}_{j}-{d}_{k}{{\Vert}}_{2}^{2}$ that is maximized at 1 and minimized at zero and ∞. To simplify the computations, the adaptive weights are computed using the coefficients from a previous iteration. The first term of (26) is the data fidelity. There are three regularization terms in our proposed model. We assume that coefficients

**d**_{i}are likely to vary smoothly within the same class of images. When

**d**_{i}and

**d**_{j}fall into the same class,

*w*

_{ij}tends to be larger, forcing the next iterate of

**d**_{i}and

**d**_{j}to be close as well. The second regularization term is a structure guided regularity following from (16) to guarantee that the coefficients of the residual components do not grow too large. Finally, we impose sparsity constraints using the ||·||

_{1, 1}norm [48] defined by

since we assume that edges are sparse in the image patches. The resulting minimization can be solved using *splitting* techniques to obtain closed form solutions for the coefficients. We solve (26) iteratively using the alternating direction method of multipliers (ADMM) algorithm [49, 50].

Due to the non-differentiability of the ||·||_{1, 1} norm, we introduce a new variable * U* and solve the equality-constrained optimization problem

By introducing the augmented Lagrangian and completing the squares, we change the constrained optimization problem (29) into the following non-constrained optimization where * V* is Lagrangian multiplier,

with step length γ ≥ 0. ADMM reduces (30) to solving the following separable subproblems:

• * D* subproblem:

Given that * D* = [

**d**_{1}, …

**d**_{λNIi}],

*= [*

**C**

**c**_{1}, …

**c**_{λNIi}],

*= [*

**E**

**e**_{1}, …

**e**_{λNIi}] and we can solve for each column of

*separately*

**D**where *y* is the corresponding column in **Y**_{(λ,Ii)} matching the column index of **d**_{j}. This gives a solution

where ${T}^{{\ell}_{T}}$ represents the transpose of **T**^{ℓ}.

• * C* subproblem:

From which we get

• * E* subproblem:

Similarly, this gives the solution

• * U* subproblem:

A solution to this problem arising from the group sparsity constraint can be obtained by applying ℓ_{1} *shrinkage* to the rows of * U*, i.e.,

where

for any * x* ∈ ℝ

^{n}and γ > 0.

The update for * V* is obtained by

The estimated set of patches ${{P}}_{\hat{X}}$ can be generated using (25) once the coefficients have been obtained. After running through the procedure for all the *k* sets, we consolidate the patches to form the enhanced resolution result $\hat{X}$. The proposed algorithm for SISR is summarized in Algorithm 1. Step 3 contains most of the key components of the algorithm so we provide more details. An outer layer loop is used to pick up residuals and put them back into the super resolution algorithm to further enhance the results. τ is the total number of outer layer iterations, * B* and

*are the blurring and downsampling matrices, respectively. In the absence of blur,*

**S***becomes the identity matrix. Starting from one cluster of low resolution image patches, Step 3(b) solves the minimization problem (30) to obtain the coefficients*

**B***, from which one is able to assemble a higher resolution image as shown in Step 3(c). From this higher resolution image, we can create a lower resolution image by applying*

**C, D, E***operator on it. Step 3(d) calculates the error between the actual low resolution and the simulated one. If the higher resolution image is close to the ground truth, we expect the error to be small. Otherwise, we feeding this difference into the minimization problem (30) can further recover more details. These details are added to the previously recovered higher resolution image to achieve a better image [40].*

**S****B**### 3.2. Pansharpening

Given a LR MS image * Y* and a high spatial resolution PAN image (

*) matching the same scene, our goal is to estimate a HR MS image $\hat{X}$ with the spatial dimensions of the PAN. We propose a two-stage approach to solve the pansharpening problem. To begin with, we obtain a preliminary estimate for the target image $\stackrel{~}{Y}$ using CS or MRA, two fast and simple pansharpening techniques. $\stackrel{~}{Y}$ has the same size as $\hat{X}$ but the result $\stackrel{~}{Y}$ obtained using these methods suffer from spectral and spatial distortion drawbacks. The second step in our proposed two-stage approach provides a way to reduce the drawbacks. Next, we feed Algorithm 1 with input $\stackrel{~}{Y}$ and proceed with the resulting optimization model to generate $\hat{X}$. Considering the image $\stackrel{~}{Y}$, the modified optimization model becomes*

**P**where ${\stackrel{~}{Y}}_{(\lambda ,{I}_{i})}=[{\stackrel{~}{y}}_{1,{p}_{1}},\dots ,{\stackrel{~}{y}}_{\lambda ,{p}_{1}},\dots ,{\stackrel{~}{y}}_{1,{p}_{{N}_{{I}_{i}}}},\dots ,{\stackrel{~}{y}}_{\lambda ,{p}_{{N}_{{I}_{i}}}}]\in {\mathbb{R}}^{{n}_{p}{n}_{p}\times \lambda {N}_{{I}_{i}}}$ for the set of patches ${{P}}_{\stackrel{~}{Y}}^{i}$, 1 ≤ *i* ≤ *k*. K-means clustering is used to decompose the number of overlapping patches into *k* classes, so that pixels corresponding to similar kinds of objects in the image will be collectively estimated in each step. The optimal value of *k* for a given image was determined using the silhouette method [51] which measures how similar a point is to its own cluster compared to other clusters. The adaptive weights are defined the same way as in (27). Applying ADMM to (31) we can solve the separable subproblems that ensue for the coefficients * D, C* and

*as in the single image model. We summarize the procedure in Algorithm 2.*

**E**## 4. Experimental Results

### 4.1. Simulation Protocol

To begin with the simulation procedure, we test the performance of the proposed algorithm by generating the reduced resolution image from a given image for some chosen downscaling factors. We then compute the enhanced resolution result using the proposed algorithm and compare its performance with other methods. The experiments are implemented using MATLAB(R2018a).

For the SISR experiments we compare our three-dimensional single image super-resolution algorithm to bicubic interpolation and the RKHS algorithm by Deng et al. [40]. The authors in Deng et al. [40] implement their SR algorithm following an independent band procedure. For color images (RGB), they transform the image to the “YCbCr” color space and apply model in Deng et al. [40] to the luma component (Y) only since humans are more sensitive to luminance changes. The “Cb” and “Cr” components are the blue-difference and the red-difference chroma components, respectively. This way, the time complexity of the algorithm is reduced, as compared to an optimization based methods using all the channels at the same time. To obtain the desired result, the chroma components (Cb, Cr) are upscaled by bicubic interpolation. The components are then transformed back to the RGB color space for further analysis. In our approach, we apply Algorithm 1 to the three-dimensional color images directly, without undertaking any transformations.

We undertake the pansharpening experiments using semi-synthetic test data. This may be attributed to the fact that existing sensors cannot readily provide images at the increased spatial resolution that is desired. A generally acceptable procedure attributed to Ranchin and Wald [23] for operating at reduced resolution is as follows (Illustration in Figure 3):

• The enhanced result should be as identical as possible to the original multispectral image once it has been degraded.

• The enhanced result should be identical to the corresponding high resolution image that a capable sensor would observe at the increased resolution.

• Each band of the enhance result should be as identical as possible to the corresponding bands of the image that would have been observed by a sensor with high resolution.

We apply Algorithm 2 to the observed multispectral data and compare it with existing state-of-the-art approaches.

### 4.2. Parameter Selection

Parameter selection is essential to obtain good results. However, to show the stability of our proposed method, we fix the parameters for all the experiments. Fine tuning of the parameters might lead to slightly better results. Parameter choices for the optimization of our proposed algorithms are outlined as follows:

• Scale factor (ω): In the experiments shown above, we set ω = 2 for the SISR problem, and set ω = 4 for the pansharpening problem. Quantitative measures obtained suggest that for ω > 4 both problems yield unsatisfactory results as the spatial and spectral measures are less competitive. However, the choice of ω depends on the dimensions of the observed data. For large ω, we can select larger patch sizes to improve upon the target.

• Patch size (*m*_{p}): Patch sizes can vary according to the size of the given image. We use a default patch size, *m*_{p} × *m*_{p} × λ in our experiments. Reasonable quantitative measures were obtained for 6 ≤ *m*_{p} ≤ 8 for the two problems considered. Outside the given bounds, the algorithms either take time to converge or yield unsatisfactory results.

• Overlap size: The limits of the overlapping region range from 2 to 4 are limited by the patch size. We set the overlap size to 2.

• Number of classes (*k*): In our experiments, the value of *k* was decided based on the solution that gave the best result from the silhouette method [51], i.e., by combing through predetermined values. Future work will involve investigating whether it is possible to automate this step.

• Construction of approximation matrices * T, K*, and

**Ψ**: We generate the matrices based on the procedure outlined in section 3, with

*M*= 6, ε = 1

*e*

^{−3}, and set 20 evenly distributed angles on [0, 2π] in computing the fine and coarse scale matrices for

*and*

**T**,**K****Ψ**, respectively.

• Regularization penalties μ_{1}, μ_{2}, μ_{3} and γ: Without knowledge of optimal values for the penalty, we sweep through a range of values to determine the optimal result. For the experiments considered, μ_{1}, μ_{2}, μ_{3}, γ ∈ {10^{−8}, ⋯ , 10^{2}} and choose the one that leads to the best results evaluated by quantitative and qualitative measures. The step length γ for updating of * V* influences the convergence of the algorithm greatly.

• Preprocessing for Algorithm 2: Our proposed pansharpening method provides a flexible way to improve the results for both CS and MRA algorithms. In the experiments shown above, we preprocess the images using the GS algorithm. Future work will detail which preprocessing step works best for multispectral images.

### 4.3. Quantitative Criteria

To measure the similarity of the enhanced resolution image ($\hat{X}$) to the reference image (* X*), we compute the following image quality metrics:

• Correlation coefficient (CC): The correlation coefficient measures the similarity of spectral features between the reference and pansharpened image. It is defined as

where **X**_{i} and ${\hat{X}}_{i}$ are the matrices of the *i*-th band of * X* and $\hat{X}$, with μ

_{Xi}and ${\mu}_{{\hat{X}}_{i}}$ as their respective mean gray values. The standard deviations for ${\hat{X}}_{i}$ and

**X**_{i}are ${\sigma}_{{\hat{X}}_{i}}$ and σ

_{Xi}, respectively.

• Root mean square error (RMSE):

where

• Peak signal-to-noise ratio (PSNR): It is an expression of the ratio between the maximum possible value of the signal and the distorting noise that affects the quality of the representation.

• Structural Similarity (SSIM): The SSIM index is computed by examining various windows of the reference and target image. It is a full reference metric designed to improve upon the traditional methods such as PNSR and MSE.

• Spectral angle mapper (SAM): This measure denotes the absolute value of the angle between the reference and estimated spectral vectors.

Let ${\hat{X}}_{\left\{i\right\}}={({\widehat{x}}_{i,1},\dots ,{\widehat{x}}_{i,\lambda})}^{T},1\le i\le {n}_{1}{n}_{2}$ be a vector collecting the intensity at all spectral bands at pixel *i*,

• Erreur relative globale adimensionnelle de synthése (ERGAS): The relative dimensionless global error in the synthesis reflects the overall quality of the pansharpened image.

• *Q* index: It defines a universal measure that combines the loss of correlation, luminance, and contrast distortion of an image and is defined by

where ${\sigma}_{({X}_{i},{\hat{X}}_{i})}$ is the covariance between **X**_{i} and ${\hat{X}}_{i}$.

### 4.4. Results

In this section, we compare the visual results and quantitative measures of the proposed approaches with some existing super-resolution algorithms. The results of Algorithm 1 will be compared to bicubic interpolation and the RKHS model by Deng et al. [40]. Results of Algorithm 2 are compared with CS methods such as PCA, IHS, Brovey, GS and Indusion, as well as MRA methods such as HPF, SFIM, ATWT, AWLP, GLP techniques, and the Deng et al. pansharpening model [43]. We begin with numerical tests for true color images and consider patches of size 8 × 8 × λ. The size of the overlap across the vertical and horizontal dimensions is 2.

In Figure 4, we show one example of decomposing an original image into three components. One can see that the edge information is separated successfully from the smooth component and the residual. In Figure 5, we compare 3D super resolution result of the proposed model with bicubic interpolation and RKHS [40] on a baboon image. We can see that the proposed method leads to a color image with highest visual quality. The bicubic interpolation result tends to be blurry. RKHS [40] result is better but not as good as the proposed. In Table 1, we list some quantitative metric comparison for more testing images. It is observed that the proposed 3D single image super resolution method (detailed in Algorithm 1) is consistently the best among the three approaches.

**Figure 4**. Decomposition of components for the grayscale test image. From top to bottom and left to right: the original, **T**^{h}**D**, **K**^{h}* C*,

**Ψ**

^{h}

*.*

**E****Figure 5**. Results for *baboon* using a scale factor ω = 2. From top to bottom and left to right: ground truth, bicubic interpolation, RKHS [40] and the proposed.

For pansharpening, we compare with 10 other methods in the literature using five different quantitative metrics (see Table 2). It is observed that the proposed pansharpening method (detailed in Algorithm 2) is mostly the best and occasionally the second among all. The performance is consistently superior.

Note that we did not conduct comparison with deep neural network based methods. The proposed approach is based on single image modeling while the deep neural network approach relies on a lot of external data. We don't think it is a fair comparison.

## 5. Conclusions

In this paper, we have proposed a technique for single image SR by modeling the image as a linear combination of regular functions in tandem with sparse coding. We show that the proposed scheme is an improvement upon similar existing approaches as it outperforms these algorithms. Besides this advantage, the proposed methods can also benefit image decomposition.

In future work, we will apply the model to multispectral and hyperspectral image SR where sparse coding can be useful due to the redundant nature of the input.

## Data Availability Statement

Some of the datasets for this study can be found in http://openremotesensing.net.

## Author Contributions

WG contributes on providing the main ideas including the mathematical models and algorithms proposed in the paper. RL was a former Ph.D. student of WG. Under WG's guidance, RL did computer programming to implement the ideas and conducted numerical experiments to test the effective of the proposed method compared to the state of the art. XZ and CG are domain scientists who help us evaluate the numerical results and provide comments.

## Funding

WG was supported by National Science Foundation (DMS-1521582).

## Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

## References

1. Farsiu S, Robinson D, Elad M, Milanfar P. Advances and challenges in super-resolution. *Int J Imaging Syst Technol*. (2004) **14**:47–57. doi: 10.1002/ima.20007

2. Yang C-Y, Ma C, Yang M-H. Single-image super-resolution: a benchmark. In: *European Conference on Computer Vision*. Cham: Springer (2014). p. 372–86. doi: 10.1007/978-3-319-10593-2_25

3. Park SC, Park MK, Kang MG. Super-resolution image reconstruction: a technical overview. *IEEE Signal Process Mag*. (2003) **20**:21–36. doi: 10.1109/MSP.2003.1203207

4. Baker S, Kanade T. Limits on super-resolution and how to break them. *IEEE Trans Pattern Anal Mach Intell*. (2002) **24**:1167–83. doi: 10.1109/TPAMI.2002.1033210

5. Lin Z, Shum H-Y. Fundamental limits of reconstruction-based superresolution algorithms under local translation. *IEEE Trans Pattern Anal Mach Intell*. (2004) **26**:83–97. doi: 10.1109/TPAMI.2004.1261081

6. Thévenaz P, Blu T, Unser M. Interpolation revisited [medical images application]. *IEEE Trans Pattern Anal Mach Intell*. (2000) **19**:739–758. doi: 10.1109/42.875199

7. Keys R. Cubic convolution interpolation for digital image processing. *IEEE Trans Acoust Speech Signal Process*. (1981) **29**:1153–60. doi: 10.1109/TASSP.1981.1163711

8. Glasner D, Bagon S, Irani M. Super-resolution from a single image. In: *IEEE 12th International Conference on Computer Vision, 2009*. IEEE (2009). p. 349–56. doi: 10.1109/ICCV.2009.5459271

9. Freedman G, Fattal R. Image and video upscaling from local selfexamples. *ACM Trans Graph*. (2011) **30**:12. doi: 10.1145/2010324.1964943

10. Gao X, Zhang K, Tao D, Li X. Joint learning for single-image super-resolution via a coupled constraint. *IEEE Trans Image Process*. (2012) **21**:469–80. doi: 10.1109/TIP.2011.2161482

11. Wang Z, Yang Y, Wang Z, Chang S, Yang J, Huang TS. Learning super-resolution jointly from external and internal examples. *IEEE Trans Image Process*. (2015) **24**:4359–71. doi: 10.1109/TIP.2015.2462113

12. Yang J, Wright J, Huang TS, Ma Y. Image super-resolution via sparse representation. *IEEE Trans Image Process*. (2010) **19**:2861–73. doi: 10.1109/TIP.2010.2050625

13. Bioucas-Dias JM, Figueiredo MA. A new twist: two-step iterative shrinkage/thresholding algorithms for image restoration. *IEEE Trans Image Process*. (2007) **16**:2992–3004. doi: 10.1109/TIP.2007.909319

14. Rudin LI, Osher S. Total variation based image restoration with free local constraints. In: *IEEE International Conference on Image Processing, 1994*. ICIP-94. IEEE (1994). p. 31–5.

15. Getreuer P. Contour stencils: Total variation along curves for adaptive image interpolation. *SIAM J Imaging Sci*. (2011) **4**:954–79. doi: 10.1137/100802785

16. Kim KI, Kwon Y. Single-image super-resolution using sparse regression and natural image prior. *IEEE Trans Pattern Anal Mach Intell*. (2010) **32**:1127–33. doi: 10.1109/TPAMI.2010.25

17. Sun J, Xu Z, Shum H-Y. Image super-resolution using gradient profile prior. In: *IEEE Conference on Computer Vision and Pattern Recognition, 2008*. CVPR. IEEE (2008). p. 1–8.

18. Fattal R. Image upsampling via imposed edge statistics. *ACM Trans Graph*. (2007) **26**:95. doi: 10.1145/1276377.1276496

19. Lium B, Son S, Kim H, Nah S, Mu K. Enhanced deep residual networks for single image super-resolution. In: *The IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops*. (2017).

20. Yang W, Zhang X, Tian Y, Wang W, Xue J, Liao Q. Deep learning for single image super-resolution: a brief review. *IEEE Trans Multimedia*. (2019) **21**:3106–21. doi: 10.1109/TMM.2019.2919431

21. Wang Y, Wang L, Wang H, Li P. End-to-end image super-resolution via deep and shallow convolutional networks. *IEEE Access*. (2019) **7**:31959–70. doi: 10.1109/ACCESS.2019.2903582

22. Li Z, Yang J, Liu Z, Yang X, Jeon G, Wu W. Feedback network for image super-resolution. In: *IEEE Conference on Computer Vision and Pattern Recognition CVPR*. Long Beach, CA: IEEE (2019). doi: 10.1109/CVPR.2019.00399

23. Ranchin T, Wald L. Fusion of high spatial and spectral resolution images: the ARSIS concept and its implementation. *Photogram Eng Remote Sens*. (2000) **66**:49–61.

24. Loncan L, de Almeida LB, Bioucas-Dias JM, Briottet X, Chanussot J, Dobigeon N, et al. Hyperspectral pansharpening: a review. *IEEE Geosci Remote Sens Mag*. (2015) **3**:27–46. doi: 10.1109/MGRS.2015.2440094

25. Vivone G, Alparone L, Chanussot J, Dalla Mura M, Garzelli A, Licciardi GA, et al. A critical comparison among pansharpening algorithms. *IEEE Trans Geosci Remote Sens*. (2015) **53**:2565–86. doi: 10.1109/TGRS.2014.2361734

26. Shettigara V. A generalized component substitution technique for spatial enhancement of multispectral images using a higher resolution data set. *Photogramm Eng Remote Sens*. (1992) **58**:561–7.

27. Shah VP, Younan NH, King RL. An efficient pan-sharpening method via a combined adaptive PCA approach and contourlets. *IEEE Trans Geosci Remote Sens*. (2008) **46**:1323–35. doi: 10.1109/TGRS.2008.916211

28. Tu T-M, Su S-C, Shyu H-C, Huang PS. A new look at IHS-like image fusion methods. *Inform Fusion*. (2001) **2**:177–86. doi: 10.1016/S1566-2535(01)00036-7

29. Carper JW, Lillesand TM, Kiefer RW. The use of intensityhue- saturation transformations for merging spot panchromatic and multispectral image data. *Photogramm Eng Remote Sens*. (1990) **56**:459–67.

30. Chavez P, Sides SC, Anderson JA. Comparison of three different methods to merge multiresolution and multispectral data: landsat TM and SPOT panchromatic. *Photogramm Eng Remote Sens*. (1991) **57**:295–303.

31. Gillespie AR, Kahle AB, Walker RE. Color enhancement of highly correlated images. I. Decorrelation and HSI contrast stretches. *Remote Sens Environ*. (1986) **20**:209–35. doi: 10.1016/0034-4257(86)90044-1

32. Laben CA, Brower BV. *Process for Enhancing the Spatial Resolution of Multispectral Imagery Using Pan-Sharpening*. US Patent 6011875 (2000).

33. Liu J Smoothing filter-based intensity modulation: a spectral preserve image fusion technique for improving spatial details. *Int J Remote Sens*. (2000) **21**:3461–72. doi: 10.1080/014311600750037499

34. Aiazzi B, Alparone L, Baronti S, Garzelli A, Selva M. An MTF-based spectral distortion minimizing model for pan-sharpening of very high resolution multispectral images of urban areas. In: *2nd GRSS/ISPRS Joint Workshop on Remote Sensing and Data Fusion over Urban Areas*. IEEE (2003). p. 90–4.

35. Aiazzi B, Alparone L, Baronti S, Garzelli A, Selva M. MTF-tailored multiscale fusion of high-resolution MS and pan imagery. *Photogramm Eng Remote Sens*. (2006) **72**:591–6. doi: 10.14358/PERS.72.5.591

36. Vivone G, Restaino R, Dalla Mura M, Licciardi G, Chanussot J. Contrast and error-based fusion schemes for multispectral image pansharpening. *IEEE Geosci Remote Sens Lett*. (2014) **11**:930–4. doi: 10.1109/LGRS.2013.2281996

37. Otazu X, González-Audícana M, Fors O, Núñez J. Introduction of sensor spectral response into image fusion methods. Application to wavelet-based methods. *IEEE Trans Geosci Remote Sens*. (2005) **43**:2376–85. doi: 10.1109/TGRS.2005.856106

38. Hou L, Zhang X. Pansharpening image fusion using cross-channel correlation: a framelet-based approach. *J Math Imaging Vis*. (2016) **55**:36–49. doi: 10.1007/s10851-015-0612-x

39. He X, Condat L, Bioucas-Dias JM, Chanussot J, Xia J. A new pansharpening method based on spatial and spectral sparsity priors. *IEEE Trans Image Process*. (2014) **23**:4160–74. doi: 10.1109/TIP.2014.2333661

40. Deng L-J, Guo W, Huang T-Z. Single-image super-resolution via an iterative reproducing kernel hilbert space method. *IEEE Trans Circ Syst Video Technol*. (2016) **26**:2001–14. doi: 10.1109/TCSVT.2015.2475895

41. Wahba G, Spline Models for Observational Data. *Society for Industrial and Applied Mathematics*. Philadelphia, PA: Siam (1990). doi: 10.1137/1.9781611970128

42. Irani M, Peleg S. Super resolution from image sequences. In: *10th International Conference on Pattern Recognition*. IEEE (1990). p. 115–20.

43. Deng L-J, Vivone G, Guo W, Dalla Mura M, Chanussot J. A variational pansharpening approach based on reproducible kernel hilbert 12 space and heaviside function. *IEEE Trans Image Process*. (2018). doi: 10.1109/ICIP.2017.8296338

44. Duchon J. Splines minimizing rotation-invariant semi-norms in Sobolev spaces. In: *Constructive theory of Functions of Several Variables*. Berlin: Springer (1977). p. 85–100. doi: 10.1007/BFb0086566

45. Meinguet J. Multivariate interpolation at arbitrary points made simple. *Z Angew Math Phys*. (1979) **30**:292–304. doi: 10.1007/BF01601941

46. Kainen P, Kůrková V, Vogt A. Best approximation by linear combinations of characteristic functions of half-spaces. *J Approx Theory*. (2003) **122**:151–9. doi: 10.1016/S0021-9045(03)00072-8

47. Arthur D, Vassilvitskii S. k-means++: the advantages of careful seeding. In: *Proceedings of the Eighteenth Annual ACM-SIAM Symposium on Discrete Algorithms. Society for Industrial and Applied Mathematics*. (2007). p. 1027–35.

48. Kowalski M. Sparse regression using mixed norms. *Appl Comput Harm Anal*. (2009) **27**:303–24. doi: 10.1016/j.acha.2009.05.006

49. Boyd S, Parikh N, Chu E, Peleato B, Eckstein J. Distributed optimization and statistical learning via the alternating direction method of multipliers. *Found Trends Mach Learn*. (2011) **3**:1–122. doi: 10.1561/2200000016

50. Eckstein J, Bertsekas DP. On the Douglas-Rachford splitting method and the proximal point algorithm for maximal monotone operators. *Math Program*. (1992) 55–**3**:293–318. doi: 10.1007/BF01581204

Keywords: super-resolution, reproducible kernel Hilbert space (RKHS), heaviside, sparse representation, multispectral imaging

Citation: Lartey R, Guo W, Zhu X and Grohnfeldt C (2020) Analog Image Modeling for 3D Single Image Super Resolution and Pansharpening. *Front. Appl. Math. Stat.* 6:22. doi: 10.3389/fams.2020.00022

Received: 01 February 2020; Accepted: 11 May 2020;

Published: 12 June 2020.

Edited by:

Ke Shi, Old Dominion University, United StatesReviewed by:

Jian Lu, Shenzhen University, ChinaXin Guo, Hong Kong Polytechnic University, Hong Kong

Copyright © 2020 Lartey, Guo, Zhu and Grohnfeldt. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Weihong Guo, wxg49@case.edu