Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci., 23 November 2022
Sec. Visual Neuroscience
This article is part of the Research Topic Neuroscience-Inspired Visual Sensing and Understanding View all 7 articles

Dynamically attentive viewport sequence for no-reference quality assessment of omnidirectional images

  • 1School of Information Science and Engineering, Ningbo University, Ningbo, China
  • 2College of Science and Technology, Ningbo University, Ningbo, China

Omnidirectional images (ODIs) have drawn great attention in virtual reality (VR) due to the capability of providing an immersive experience to users. However, ODIs are usually subject to various quality degradations during different processing stages. Thus, the quality assessment of ODIs is of critical importance to the community of VR. The quality assessment of ODIs is quite different from that of traditional 2D images. Existing IQA methods focus on extracting features from spherical scenes while ignoring the characteristics of actual viewing behavior of humans in continuously browsing an ODI through HMD and failing to characterize the temporal dynamics of the browsing process in terms of the temporal order of viewports. In this article, we resort to the law of gravity to detect the dynamically attentive regions of humans when viewing ODIs. In this article, we propose a novel no-reference (NR) ODI quality evaluation method by making efforts on two aspects including the construction of Dynamically Attentive Viewport Sequence (DAVS) from ODIs and the extraction of Quality-Aware Features (QAFs) from DAVS. The construction of DAVS aims to build a sequence of viewports that are likely to be explored by viewers based on the prediction of visual scanpath when viewers are freely exploring the ODI within the exploration time via HMD. A DAVS that contains only global motion can then be obtained by sampling a series of viewports from the ODI along the predicted visual scanpath. The subsequent quality evaluation of ODIs is performed merely based on the DAVS. The extraction of QAFs aims to obtain effective feature representations that are highly discriminative in terms of perceived distortion and visual quality. Finally, we can adopt a regression model to map the extracted QAFs to a single predicted quality score. Experimental results on two datasets demonstrate that the proposed method is able to deliver state-of-the-art performance.

Introduction

The omnidirectional image (ODI), which records and delivers 360-degree surround information, plays an important role in virtual reality (VR) photography. The brand-new viewing experience enabled by ODIs is substantially different from traditional 2D plane images, as humans are allowed to freely change their viewport to explore the immersive virtual environments through head-mounted display (HMD) (Li et al., 2018; Zhang et al., 2018; Tran et al., 2019; Deng et al., 2021). Due to its capability of providing natural immersions of real-world scenarios, ODIs have attracted lots of attentions from both academia and industry. In the meanwhile, ODIs have been put into widespread use in many practical VR applications (Chen et al., 2019; ISO, 2019; Alain et al., 2020).

The visual quality of ODIs is a researching-worth topic as ODI content with poor visual quality may cause both physical and mental discomforts. Compared with the traditional 2D plane image, the visual quality issues of ODIs are much more acute and challenging. On the one hand, the currently mainstream methods for acquiring a typical ODI are to stitch multiple images which are captured by using a wide-angle camera array with partially overlapped field of view. These images from multiple cameras are stitched to produce an omnidirectional panorama in the spherical format. However, the stitching process will inevitably introduce stitching artifacts at the stitching boundaries, which are scarcely occurred in traditional 2D plane images (Ho and Budagavi, 2017). On the other hand, compared with traditional 2D plane images, the storage space requirement and the spatial resolution of ODIs are much higher, e.g., 4K, 8K, or higher. Therefore, ODIs are often heavily compressed to facilitate transmission and storage (Kim et al., 2020). As a result, ODIs with serious compression artifacts inevitably lead to even worse quality-of-experience. In addition, ODIs are usually with the spherical format to be displayed on HMD. Therefore, the human viewing behavior when freely exploring the ODIs with HMD is dramatically different with that of 2D plane images. This dramatically different viewing behavior will also affect the human quality perception of ODIs accordingly. On considering the above quality issues, it is necessary to develop effective ODI quality metrics by jointly considering the effects of different distortions as well as the human viewing behaviors in viewing ODIs with HMD.

Despite its high importance, the problem of ODI quality evaluation has been not well addressed so far. One of the most important challenges is that human viewing behavior in browsing ODIs through HMD is dramatically different from that in viewing 2D plane images by human eyes directly. Typically, when viewers are browsing a spherical ODI through HMD, they can obtain immersive and interactive viewing experiences by freely changing the viewpoint, and only the visual contents within the current viewport can be viewed at a certain time (Chen et al., 2020; Xu M. et al., 2020). Besides, viewers usually tend to focus on the regions near the equator during the exploration of ODIs with HMD. Therefore, the prediction of viewports plays a critical role in designing accurate ODI quality evaluation metrics as it can extract the most important visual contents from ODIs to facilitate the quality evaluation process. To account for this, some previous efforts have been made to extract viewports from ODIs for ODI quality evaluation. For example, the VGCN-based ODI quality evaluation method proposed in Xu J. et al. (2020) was dedicated to extracting viewports with higher probabilities of being explored by viewers according to the human visual sensitivity to structural information and adopted the graph convolution network (GCN) to predict ODI quality score by implicitly modeling the interactions among different viewports. In Sun et al. (2019), the authors proposed to project the equirectangular image into six equally sized viewport images and then generate a corresponding channel for further study. These previous works have achieved promising performance and demonstrated the validity and importance of viewport generation toward creating reliable ODI quality evaluation metrics. Despite the effectiveness, the viewport generation strategies of these methods have a common limitation, i.e., they both ignore the actual viewing behavior of humans in continuously browsing an ODI through HMD and fail to characterize the temporal dynamics of the browsing process in terms of the temporal order of viewports. Therefore, research efforts dedicated to more accurate viewport sequence generation by accounting for the temporal order of viewports will undoubtedly facilitate the quality evaluation of ODIs.

In this article, we propose a novel no-reference (NR) ODI quality evaluation method by making efforts on two aspects including the construction of Dynamically Attentive Viewport Sequence (DAVS) from ODIs and the extraction of Quality-Aware Features (QAFs) from DAVS. The proposed method is named Spatiotemporal Scene Statistics of Dynamically Attentive Viewport Sequence (S3DAVS) in short. The construction of DAVS aims to build a sequence of viewports that are likely to be explored by viewers based on the prediction of visual scanpath when viewers are freely exploring the ODI within the exploration time via HMD. A DAVS that contains only global motion can then be obtained by sampling a series of viewports from the ODI along the predicted visual scanpath. As a result, the obtained DAVS can be considered a human viewing behavior-characterized compact representation of the whole ODI. The subsequent quality evaluation of ODIs is performed merely based on the DAVS. The extraction of QAFs aims to obtain effective feature representations that are highly discriminative in terms of perceived distortion and visual quality. Finally, we can adopt a regression model to map the extracted QAFs to a single predicted quality score.

The contributions of this work are 2-fold. First, we make the first attempt to predict the visual scanpath based on which a DAVS is accordingly obtained as one kind of human viewing behavior-characterized compact representation of the whole ODI. Second, we model the spatiotemporal scene statistics by analyzing the 3D-MSCN and spatiotemporal Gabor response maps of the DAVS to serve as the QAFs for the quality evaluation of ODIs. Finally, we conduct extensive experiments on two benchmark databases to validate the effectiveness of our proposed S3DAVS method.

The rest of this article is arranged as follows. Section “Related works” provides a brief review of some representative ODI quality metrics. Section “Proposed spatiotemporal scene statistics of dynamically attentive viewport sequence approach” illustrates the details of our proposed ODI quality evaluation metric. Section “Experimental results” presents the experiments and performance comparisons. Section “Conclusion” concludes and discusses future works.

Related works

Compared with traditional 2D plane images, ODIs are always extremely high-resolution and a large amount of data which hereby directly increases the burden of transmission and storage. Therefore, the visual quality issues of ODIs are much more acute and challenging, calling for dedicated ODI quality evaluation metrics. During the past decades, image quality evaluation has been widely investigated and can be roughly categorized into three categories including full-reference (FR), reduced-reference (RR), and no-reference (NR) according to the participation amount of information from reference image (i.e., without distortion) (Mittal et al., 2013; Zhu et al., 2019; Liu et al., 2021). Compared with the FR/RR methods which require full/partial reference information, the NR methods do not require any information from the reference image, which can have much wider applications in practical systems. In this section, we will provide a brief review of some representative FR and NR ODI quality metrics in the literature.

FR-OIQA

There have been many FR image quality assessment (IQA) metrics designed for 2D plane images (Wang et al., 2003; Zhou et al., 2004; Sheikh et al., 2006; Zhang et al., 2011; Xue et al., 2014). However, the existing 2D FR-IQA metrics cannot be directly applied to evaluate ODIs because the geometric distortions induced by projection will be wrongly treated and evaluated. Therefore, the previous research efforts mostly concentrate on studying to improve existing 2D FR-IQA metrics by removing the influence of the projection-related geometric distortions. Yu et al. (2015) proposed a spherical PSNR (S-PSNR) metric to improve the performance of traditional PSNR by sampling pixel locations from a sphere uniformly and getting pixel contents from the reference image and the distorted image according to the relation of spherical coordinates to 2D coordinates, then calculated the error between the pixels on the reference image and the pixels on the distorted image. Zakharchenko et al. (2017) introduce a Craster’s parabolic projection (CPP-PSNR) method by projecting the reference image and distorted image into a shared CPP format domain before calculating the PSNR. However, the drawback of CPP-PSNR methods is that interpolation may introduce errors and decrease the accuracy. Sun et al. (2017) advocated calculating weighted-to-spherically uniform PSNR (WS-PSNR) on the 2D format of the omnidirectional image directly, which may decrease the negative effect of interpolation in CPP-PSNR. However, PSNR is not highly correlated with human perception (Horé and Ziou, 2010). Inspired by the good consistency of structure similarity (SSIM) metrics with subjective perception in 2D-IQA, Zhou et al. (2018) introduce a weighted-to-spherically uniform SSIM metric for FR-OIQA. Besides, some studies suggest figuring out more information from other dimensions to improve performance. For example, the phase consistency-guided method (Xia et al., 2021) proposes to utilize the abundant structure and texture features of high-order phase information from ODIs. Recently, with the development of deep learning, Kim et al. (2020) constructed a novel FR-OIQA model by learning the visual and positional features with the guidance of human perception through adversarial learning.

NR-OIQA

The limitation of FR-OIQA metrics is the dependence on reference images which may be unavailable in practical applications. Therefore, NR-IQA metrics that do not require any reference information are desired (Gu et al., 2015). In the literature, several NR-IQA metrics for 2D images have been proposed (Moorthy and Bovik, 2011; Min et al., 2018; Yan et al., 2019). However, as demonstrated in Xu J. et al. (2020), NR-OIQA involves new challenges, such as stitching artifacts, sphere representation, and wide field of view (FoV). Therefore, a reliable no-reference quality metric for ODIs is extremely needed. There have been some algorithms proposed for NR-IQA of ODIs. Zhou W. et al. (2022) captured the multi-frequency information by decomposing the projected ERP maps into multiple images on sub-bands. However, the conventional method of studying the overall ODI is not necessary, because the salient region in ODIs mainly locate on the equator (Sitzmann et al., 2018). Therefore, the thought of exploring the information from a viewport-based image is proposed and obtained broad acceptance. Sun et al. (2019) projected each ODI into six viewport images and then proposed a multi-channel CNN framework including six parallel ResNet34. However, the way to obtain viewport-based images is directed by the behavior of humans when they view omnidirectional images through HMD (Li et al., 2019; Xu et al., 2019; Jiang et al., 2021; Zou et al., 2021). Xu J. et al. (2020) noticed this problem and constructed a viewport-oriented graph convolutional network (VGCN) to address the perceptual quality assessment of ODIs.

Proposed spatiotemporal scene statistics of dynamically attentive viewport sequence approach

The framework of our proposed method is shown in Figure 1. Our proposed method involves three modules, including DAVS generator, QAF extractor, and image quality regressor. DAVS which is composed of a series of viewports sampled from the ODI along the scanpath can be considered a human viewing behavior-characterized compact representation of the whole ODI. Specifically, part (a) is an unprocessed ERP image from the OIQA database. The spherical format ODI converted from ERP format ODI is presented in part (b). The predicted results of scanpath are shown in part (c). We get the center location of each viewport on spherical format ODI from precomputed locations in the ERP format ODI provided in part (c) by coordinate mapping. Then, we offer the value of FOV to set the size of the viewport, and finally obtain the complete viewport content from part (b). All those viewport contents are shown in part (f). In the QAF extractor, the image sequence is generated by arranging the viewport images along the order in the scanpath. The QAF extraction is performed on the DAVS and involves obtaining effective feature representations that are highly discriminative in terms of visual quality. Finally, the extracted QAFs are used to predict the quality score via a learned regression model.

FIGURE 1
www.frontiersin.org

Figure 1. The overall framework of our proposed S3DAVS method. It contains DAVS generator, QAF extractor, and image quality regressor. The viewport sequence produced by the DAVS generator will be sent into the QAF extractor and then the image quality regressor will predict the final quality score based on the features coming from the QAF extractor. Source for the photos in this figure: open source OIQA Database.

Dynamically attentive viewport sequence construction

1) Scanpath prediction: Sitzmann et al. (2018) proposed that humans mainly pay attention to salient regions located on the equator in ODIs. Therefore, our original intention of building this omnidirectional image quality assessment algorithm is to simulate human behavior when they observe ODI through HMD (Zhu et al., 2020). Besides, different trajectories of visual scanpath will produce different contents of DAVS and accordingly affect the perceptual quality assessment. Based on these two considerations, the key to DAVS construction is accurately predicting the scanpath of viewers when they are exploring the virtual scene via HMD. The fixation is decided by visual attention shift, i.e., a phenomenon of the temporal dynamics of human visual attention (Yang et al., 2021). Current approaches in modeling scanpath prediction are based on the saliency map that illustrates the probability of each human visual attention (Ling et al., 2018). However, these approaches fail to characterize the temporal dynamics of visual attention. Recently, it has been revealed that the law of gravitation can well explain the mechanisms behind the dynamic process of attention shift in exploring a visual scene. Inspired by this, we intuitively propose a gravitational model-based visual scanpath prediction approach for ODIs. Given an ODI image as input, this model can yield a continuous function that describes the trajectory of the fixations over the exploration time as output.

We think the process that humans pay attention to the interesting area can be regarded as a field effect in HMD. Therefore, it is necessary to define the virtual mass corresponding with ODI content and the virtual field in ODIs for building a relationship between ODI content with fixation. Here, we employ the model proposed by Zanca and Gori (2017) and Zanca et al. (2020) for the utilization of gravitation theory. The ODI content noticed by humans should contain abundant details and motion information. Details reflect masses proportional to the magnitude of the gradient, while motions illustrate proportional to the optical flow which mainly occurs when humans shift to the next fixation. If we set virtual mass μ that obeys a certain distribution as the assembling of all the fixation during the exploration, and think it degenerates to a single distributional mass concentrated in x, then the virtual mass can be expressed as μ(y,t) = δ(yx), y = a(t),where a(t) represents the focus of attention at time t. The μ(y,t) consists of two elements: gradient of brightness μ1 and optical flow μ2. The field we set makes effects on the virtual mass we have designed, and then the Green function (G) will construct the correlation between the field and its corresponding mass. Therefore, our scanpath can be associated with attention a(t) by the potential,

G ( a - x ) = - 1 2 π log ( || a - x || ) (1)

Then the gravitational field e at time t can be written as:

e ( a - x ) = 1 2 π a - x || a - x || 2 (2)

where Equation (2) illustrates that the strength of field is inversely proportional to the distance between the a(t) and x. Then the overall field can be demonstrated by,

E ( a ( t ) ) = - 1 2 π R d x a ( t ) - x || a ( t ) - x || 2 μ ( x , t ) (3)
E ( a ( t ) ) = - ( e * μ ) ( a ( t ) ) (4)

where Equation (3) can be rewritten as Equation (4) using convolution operation. It depicts the links between the virtual mass in the field and the attention of viewers. Subscript R represents the overall field. However, the method first explores the focus of attention in the most attentive regions which may hinder the method from searching unexplored locations and finishing a complete exploration of the scene. Therefore, the shift of attention is needed to be triggered by an extra setting. Zanca proposed to introduce the inhibitory function (I(t)) which will return value 0 for the pixels that have not been explored and value 1 for the pixels that have already been explored. Then all the focus of attention a(t)can finally be represented as follows:

I . ( t ) = β ( g ( x - a ( t ) ) - I ( t ) ) (5)
a . ( t ) = z ( t ) (6)
z . ( t ) = - λ z ( t ) - ( e * μ ) ( t , a ( t ) ) (7)

Here, g(u)=e-u22δ2 and 0 < β < 1. The dumping term a.(t) prevents strong oscillations and makes the overall dynamics closer to human scanpath. Figure 2 shows the scanpath prediction of four images in the OIQA database. We could obtain specific coordinates of these viewport centers to be utilized in the following part.

FIGURE 2
www.frontiersin.org

Figure 2. Examples of scanpath prediction results. The red squares in each image denote fixation locations based on which the viewport images are sampled. Source for the photos in this figure: open source OIQA Database.

2) Scanpath-based Viewport Sampling: Based on the predicted scanpath data, we can obtain a series of viewport images from the ODI by sampling along the trajectory of the scanpath. The precomputed locations denoted by {s1,s2,…,st} are human fixations that have been listed along the timeline, but they only represent the center of attentive regions on the ERP ODIs. Each st includes st,x and st,y, where, (st,x, st,y) correspond with width and height values on the ERP image. To obtain the realistic viewport content from the spherical scene, we first map those locations onto a unit sphere and then define those coordinates as the predicted scanpath which is written as {p1,p2,…, pt}, where the subscript of standpt records the overall exploration time and keep the consistency between locations from the ERP image and locations from the spherical image. Specifically, given the current attentive location in the ERP image is st, we calculate its spherical location pt and consider it to be the actual fixation. Then, based on the theory about near peripheral vision (Besharse and Bok, 2011), we first set the field of view (FoV) to [-π/6, π/6], after which the scale of viewport content can be determined (Sui et al., 2022). Finally, bicubic interpolation will be used to sample the viewport image and construct the sequence according to the order of the time. Figure 3 shows how viewers perceive spherical content with HMD at time t. In the spherical coordinate, the behavior of viewer position mainly decided by rotation matrix (Rai et al., 2017) and is mainly calculated by the Cartesian theory for the location transformation. Figure 4 shows partial viewport content from the DAVS. (A) is an original distorted ERP image in the database and (B) displays the examples of different viewports in DAVS. Those overlapping viewport contents illustrate a dynamically attentive process when humans view the ODI from one attentive point to the next and finally composite into a sequence that curve the dynamically spatiotemporal information which will facilitate us to extract spatiotemporal features from the DAVS. Comparing the content in these viewports, some of these viewport contents are rich in luminance information, while others are not. This phenomenon corresponds to the theory of not depending on the saliency map but predicting the scanpath with the law of gravity, so the scanpath is more accurate than the common method.

FIGURE 3
www.frontiersin.org

Figure 3. The illustration of the spherical scene. The viewer with HMD stands at the center of the sphere. The location of attention can be described with both longitude and latitude.

FIGURE 4
www.frontiersin.org

Figure 4. Visualization of viewports in DAVS obtained by the scanpath-based viewport sampling. (A) ODI with the ERP format and (B) examples of different viewports in DAVS. Source for the photos in this figure: open source OIQA Database.

Quality-aware feature extraction and quality prediction

As stated, the built DAVS can be considered as a human viewing behavior-characterized compact representation of the whole ODI. The remaining issue is how to perform the quality evaluation of an ODI based on the DAVS. As known, the most important step of quality evaluation is to extract effective QAFs that are highly descriptive of distortion level (i.e., visual quality level). Since the DAVS is a sequence composed of a series of viewports sampled from the ODI along the scanpath, we can naturally treat the DAVS as a pseudo video with only global camera motion and each viewport in DAVS just corresponds to a specific frame in the video. Obviously, just like a natural video, the DAVS also contains information regarding the spatial variations in pixels along with the dynamic motion in successive frames (Mittal et al., 2016). Thus, the QAF extraction from DAVS should well account for the characteristics in the spatiotemporal domain. In this work, we propose to model the spatiotemporal scene statistics of the DAVS as the QAFs of the corresponding ODI. The obtained spatiotemporal scene statistics are used as the QAFs based on which quality prediction is performed.

1) Spatiotemporal MSCN Sequence: It has been widely demonstrated that the local mean subtracted contrast normalized (MSCN) coefficients of a pristine natural image can be well modeled by a Gaussian distribution (Mittal et al., 2012). However, when a natural image suffers distortions, the distribution of its MSCN coefficients will deviate from the original distribution (Mittal et al., 2012). These distributions can be modeled either by a generalized Gaussian distribution (GGD) or asymmetric GGD (AGGD). The parameters of GGD and AGGD can be used as the QAFs of natural images and have achieved great success in designing NR-IQA metrics during the past decade. The main reason for working on MSCN images is due to the decorrelation of local pixel dependency. Similarly, the DAVS which records the dynamic visual contents seen by the viewers along the scanpath has a high correlation among neighboring pixels in both spatial and temporal domains. In order to decorrelate such local dependency in the DAVS, we propose to use the spatiotemporal MSCN. Specifically, let us denote the DAVS as V(x,y,t), and the spatiotemporal MSCN coefficients of V(x,y,t) is calculated as follows:

V ^ ( x , y , t ) = V ( x , y , t ) - μ ( x , y , t ) σ ( x , y , t ) + c (8)

where x and y are the spatial indices and t is the temporal index. A small constant c is imposed on the dominator to avoid instability when σ(x,y,t) approaches to zero and we empirically set c = 1 in this work. μ(x,y,t) and σ(x,y,t) are the mean and standard deviation, respectively, which are defined as follows:

μ ( x , y , t ) = j = - J J k = - K K l = - L L ω j , k , l V ( x + j , y + k , t + l ) (9)
σ ( x , y , t ) = j = - J J k = - K K l = - L L ω j , k , l [ V ( x + j , y + k , t + l ) - μ ( x , y , t ) ] 2 (10)

where ω is a symmetric normalized 3D Gaussian weighting function with zero mean and standard deviation of 1.166. According to Dendi and Channappayya (2020), we set J = K = L = 2.

2) Spatiotemporal Gabor Filter Response Sequence: It has been hypothesized that the HVS employs spatiotemporal bandpass filters to analyze and process dynamic visual signals (Simoncelli and Olshausen, 2001). Our proposed approach is motivated by this hypothesis that spatiotemporal Gabor filters are a good approximation of the bandpass behavior of the HVS (Petkov and Subramanian, 2007; Tu et al., 2020; Gotz-Hahn et al., 2021). The spatiotemporal Gabor filters combine information over space and time, which advocates a suitable model for feature analysis of the DAVS. Mathematically, the spatiotemporal Gabor filter is defined by the product of three factors including a Gaussian envelope function that limits the spatial extent, a cosine wave moving with phase speed v in the θ direction, and a Gaussian function that determines the decay along time:

G v , θ , ϕ ( x , y , t ) = γ 2 π σ 2 exp ( - ( ( x ¯ + v c t ) 2 + γ 2 y ¯ 2 ) 2 σ 2 ) (11)
cos ( 2 π λ ( x ¯ + vt ) + φ ) 1 2 π τ exp ( - ( t - μ t ) 2 2 τ 2 )

where x¯=xcos(θ)+ysin(θ),y¯=-xsin(θ)+ycos(θ). γ is the rate that specifies the ellipticity of the Gaussian envelope in the spatial domain and is set to γ = 0.5 for matching to the elongated receptive field along the y¯ axis. σ describes the standard deviation of Gaussian and determines the size of the receptive field. v is the phase speed of the cosine factor, which determines the speed of motion. In addition, the speed which the center of the spatial Gaussian moves along the x¯ axis is specified by the parameter vc. In our implementation, we simply set vc = v. λ is the wavelength of the cosine wave and it is obtained through the relation λ=21+v2. θ ∈ [0,2π] determines the motion direction and the spatial orientation of the filter. The phase offset φ ∈ [−π,π] determines the symmetry in the spatial domain. The Gaussian distribution with mean μt = 1.75 and standard deviation τ = 2.75 is used to model the decay in intensities along time.

With the above-defined spatiotemporal Gabor filter, the bandpass response of a specific spatiotemporal MSCN sequence can be obtained by convolving it with a bank of spatiotemporal Gabor filters:

R v , θ , ϕ ( x , y , t ) = V ^ ( x , y , t ) * G v , θ , φ ( x , y , t ) (12)

where * denotes the convolution operation. In this work, we generate the bandpass spatiotemporal Gabor filter banks by varying the values of v, θ, and φ. Specifically, we select three different speeds {v = 0,1,2}, four different orientations {θ = 0,π/3,2π/3,π}, and two different phase offsets, i.e., φ = 0 for the symmetry case and φ = π/2 for the anti-symmetry case. As a result, the total number of spatiotemporal Gabor filters is 24. It means that we can obtain 24 bandpass spatiotemporal Gabor filter response sequences for an input DAVS.

3) AGGD Modeling and Parameter Estimation: In the legacy NR-IQA works, it has been widely demonstrated that the distribution of the coefficients in a 2D MSCN map can be well modeled by GGD or AGGD. Inspired by this, we in this work also employ the AGGD to model the distributions of the coefficients in the spatiotemporal MSCN and bandpass spatiotemporal Gabor filter response sequences, respectively. The AGGD is a flexible model that can effectively characterize a large variety of unimodal data with only three parameters. Mathematically, the AGGD model is described as follows:

f ( x ; γ , β l , β r ) = { γ ( β l + β r ) Γ ( 1 γ ) exp ( - ( - x β l ) γ ) ; x 0 γ ( β l + β r ) Γ ( 1 γ ) exp ( - ( - x β r ) γ ) ; x > 0 (13)

where γ,βlr are three parameters controlling the shape of the distribution, and Γ(⋅) is defined as follows:

Γ ( a ) = 0 t a - 1 e - t dt ; a > 0 (14)

We adopt the moment estimation method suggested in Lasmar et al. (2009) to estimate the three parameters γ,βlr. Besides, we also compute another parameter η = γ/(βl + βr). Finally, we use four parameters [γ,βlr,η] to represent an AGGD model.

4) Final Feature Representation: By applying the AGGD to model the distributions of the coefficients in the spatiotemporal MSCN and bandpass spatiotemporal Gabor filter response sequences, we can obtain a hybrid parameter set to serve as the QAF representation of the DAVS:

f ^ D A V S = [ f ^ V ^ , f ^ R v , θ , φ ] (15)

where f^V^ is a 12-dimensional feature vector containing the AGGD parameters obtained by applying the AGGD model on the spatiotemporal MSCN sequence at three coarse-to-fine scales (the coarser scale is first processed by a low-pass filter, followed by a down-sampling operation with a factor of 2 and f^Rv,θ,φ is a 288-dimensional feature vector also containing the AGGD parameters at three coarse-to-fine scales and they are obtained by applying the AGGD model on the bandpass spatiotemporal Gabor filter response sequences with all v, θ, and φ.

5) Quality Score Regression: After obtaining the overall feature representation of the DAVS (also the ODI), the remaining issue is how to predict the quality score of an input ODI based on the extracted feature representations. Given the subjective quality is provided in the form of scaler, the quality prediction is a typical regression problem from the perspective of machine learning. Therefore, we learn a regression model via support vector regression (SVR) for mapping the 300-dimensional QAF vector into a single quality score with the usage of a radial basis function kernel. Once the SVR model is built by training, it can be used for the quality prediction of an input ODI in the test stage with its corresponding 300-dimensional QAF vector as input.

Experimental results

Experimental protocols

We utilize two publicly available omnidirectional image quality databases in the experiments. They are the OIQA database (Duan et al., 2018) and CVIQD database (Sun et al., 2018). The OIQA database contains 16 image scenes with four distortion types, including JPEG compression (JPEG), JPEG2000 compression (JP2K), Gaussian blur (BLUR), and Gaussian white noise (WN). Further, each distortion type involves five distortion levels. So, the OIQA database includes 16 pristine ODIs and 320 distorted ODIs in total. Subjective rating scores in the form of mean opinion score (MOS) are given in the range of Xue et al. (2014) and Tran et al. (2019), where a higher score means better visual quality. The CVIQD database consists of 528 ODIs including 16 pristine images and 512 compressed ODIs. Three popular coding techniques, i.e., JPEG, H.264/AVC, and H.265/HEVC, are applied to simulate the compression artifacts. Subjective rating scores in the form of mean opinion score (MOS) are given in the range of [0, 100], where a higher score means a better visual quality.

Our proposed method was run on a computer with a 3.60 GHz Intel Core i7 processor, 64 GB main memory, and Nvidia GeForce GTX 3090 graphics. We utilize three criteria to validate the performance of IQA models, including Pearson’s linear correlation coefficient (PLCC), Spearman’s rank-order correlation coefficient (SRCC), and root mean squared error (RMSE). For PLCC and SRCC, the higher the better, while for RMSE, the tendency is the opposite. For a perfect match between prediction scores and ground-truth subjective scores, we should have PLCC = SRCC = 1 and RMSE = 0. Before computing PLCC and RMSE, according to the recommendation of the video quality expert group (VQE) (Video Quality Experts Group, 2003), we apply a standard five-parameter logistic function to adjust the predicted scores to minimize the non-linearity of the subjective rating scores:

s = α 1 ( 1 2 - 1 1 + exp ( α 2 ( p - α 3 ) ) ) + α 4 p + α 5 (16)

where p represents the predicted score and s denotes the mapped score. α1 to α5 are parameters for fitting this function.

Since our proposed method requires the training process to build the quality prediction model, it is necessary to describe the details regarding the training process. Specifically, we equally divide the whole dataset into five non-overlapping subsets with each subset containing 20% samples. Then, we apply the fivefold validation strategy to test the model performance. Specifically, the whole dataset is trained and tested five times, with four subsets as the training data and the remaining one subset as the testing data. After training and testing the whole dataset five times, we can get the predicted scores of all samples in the dataset. Finally, the PLCC, SRCC, and RMSE values are calculated between all the predicted scores and ground-truth subjective scores provided in each dataset.

Single-dataset performance comparison

We compare the performance of our proposed S3DAVS model with eight representative NR-IQA models, including BRISQUE (Mittal et al., 2012), DIIVINE (Moorthy and Bovik, 2011), SSEQ (Liu et al., 2014), OG-IQA (Liu et al., 2016), NRSL (Li et al., 2016), SSP-BOIQA (Zheng et al., 2020), MC360IQA (Sun et al., 2019), and Zhou Y. et al. (2022) work. The former five models are traditional NR-IQA models that are developed for 2D natural images, while the latter three models are specifically designed for ODIs. In these three blind omnidirectional image quality assessments (BOIQA), MC360IQA (Sun et al., 2019) and Zhou Y. et al. (2022) work employ deep learning methods, while SSP-BOIQA (Zheng et al., 2020) focuses on segmenting ODI into three regions and extracting features with weights. Based on those training-based models, we re-trained the quality prediction models on each dataset with the same 5-fold validation strategy. Specifically, before testing the data, each dataset should generate three sub-datasets with three different scales respectively, and the reduction factor is 1/2. We fuse the features from sub-datasets into final feature vectors. Then, in the test session, a 5-fold cross-validation method is used to verify the performance of the model. The performance results in terms of PLCC, SRCC, and RMSE on the OIQA and CVIQD databases are shown in Tables 1, 2, respectively. We highlight the best performance model in each column. From these two tables, we have several observations. First, for both the OIQA and CVIQD databases, our proposed S3DAVS model, although inferior to some compared methods on some individual distortion types, presents the highest PLCC, SRCC, and RMSE values when considering the overall database. It means that our method is a good candidate for OIQA when the distortion type is unknown. Second, since the OIQA database contains more distorted types than the CVIQD database, the overall results on OIQA are commonly lower than those on the CVIQD database. According to the above experimental results, we could get the conclusion that the S3DAVS model owns the best performance among those models. Moreover, we can observe that the results of SSP-BOIQA (Zheng et al., 2020) which is designed for ODIs specifically are even worse than traditional 2D models. This is mainly due to the fact that SSP-BOIQA only focused on transforming the ERP image into bipolar and equatorial regions, ignoring the extraction of features from specific contents that are highly consistent with human perception.

TABLE 1
www.frontiersin.org

Table 1. Performance comparison of the OIQA database.

TABLE 2
www.frontiersin.org

Table 2. Performance comparison on CVIQD database.

Statistical significance test

Besides the comparisons in terms of PLCC, SRCC, and RMSE, we employ that the t-test to demonstrate the superiority of our proposed S3DAVS model over other compared methods is significant. T-test could be divided into three types: one-sample t-test, independent samples t-test, and paired t-test, and is used to judge whether the performance difference between two models is significant or not. In this section, we first get 100 results of indicators and select the PLCC as the sample data, and then use the independent samples t-test method to calculate the final significance criteria. The formula can be written as:

t = m A - m B S A 2 n A + S B 2 n B (17)

For each model, m represents the means of samples, S2 represents the variance of samples, and n is the number of samples. t illustrates the significant difference between the two compared models. In Figure 5, the blue block indicates that the model at a row is worse than model in the column and is labeled as “-1.” The green block indicates that there are no obvious differences between the models in row and column and is labeled as “0.” The orange block means the model at row is worse than that in column, and is labeled as “1.” According to the data distribution of each row in the statistical significance figure, the higher the number of “1,” the better the algorithm. It can be observed that our proposed model is significantly better than all the others.

FIGURE 5
www.frontiersin.org

Figure 5. Statistical significance comparison by the t-test between our proposed method and other methods. (A,B) Depict the results on the OIQA and CVIQD databases, respectively.

Ablation study

The proposed method mainly focuses on extracting statistical features in the spatiotemporal domain and analyzing the viewport content in a way that is consistent with human viewing behavior. According to the aforementioned feature extractor, we know that our method mainly depends on the spatiotemporal MSCN (ST-MSCN) and the spatiotemporal Gabor filter (ST-Gabor) to obtain feature vectors. To figure out the performance of each module, we test the performance by considering each single feature set, i.e., only using the feature set from the ST-MSCN or the ST-Gabor filter. The results are shown in Tables 3, 4, respectively. By comparing the results in these two tables, we can easily find that the features from the ST-Gabor play a dominant role, while the spatiotemporal MSCN features play an auxiliary role in our S3DAVS. Overall, the performance will be further improved by jointly considering the spatiotemporal statistical features of the ST-Gabor and the ST-MSCN.

TABLE 3
www.frontiersin.org

Table 3. Performance comparison of the OIQA database with each single feature set.

TABLE 4
www.frontiersin.org

Table 4. Performance comparison on the CVIQD database with each single feature set.

Cross-dataset performance comparison

To test the generalization ability of an image quality metric, cross-database validation is necessary. Given two databases, we use one database for training and another one for testing. Because the OIQA database contains more distorted types, such as Gaussian noise and Gaussian blur, whereas the CVIQD database only contains compression distortion, we only tested JPEG and JP2K in the OIQA database, and the results are shown in Table 5. As shown in the table, the results on the CVIQD database are generally higher than the results on the OIQA database. The reason is that the model trained on the OIQA database learns more feature types and could discriminate the distortion types in the CVIQD database more accurately. Overall, in comparison to other existing methods, our proposed model has a better generalization capability.

TABLE 5
www.frontiersin.org

Table 5. Performance results of cross-database validation.

Conclusion

This article has presented a novel no-reference (NR) ODI quality evaluation method based on the construction of Dynamically Attentive Viewport Sequence (DAVS) from ODIs and the extraction of Quality-Aware Features (QAFs) from DAVS. The construction of DAVS aims to build a sequence of viewports that are likely to be explored by viewers based on the prediction of visual scanpath when viewers are freely exploring the ODI within the exploration time via HMD. A DAVS that contains only global motion can then be obtained by sampling a series of viewports from the ODI along the predicted visual scanpath. The subsequent quality evaluation of ODIs is performed merely based on the DAVS. The extraction of QAFs aims to obtain effective feature representations that are highly discriminative in terms of perceived distortion and visual quality. Finally, a regression model is built to map the extracted QAFs to a single predicted quality score. Experimental results on two datasets demonstrate that the proposed method is able to deliver state-of-the-art performance. However, its shortcomings are also obvious. Although this method captures effective information to analyze the feature of ODI and predict the quality, it is time-consuming and computationally complex, and it needs to be improved. Taking the OIQA database as an example, there are 320 ODIs in this database, and the smallest resolution is 11,332 × 5,666. The test result shows it takes about 1.983 min for an image to complete all the steps. Besides, the scanpath can only present the movements of human eyes, but the movements of the head are not clear. We could explore these directions in our future work.

Data availability statement

The original contributions presented in this study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

YW and HL contributed to conception and design of the study and wrote the first draft of the manuscript. YW wrote the software codes of the algorithm and conducted experiments. HL and QJ performed the statistical analysis. All authors contributed to manuscript revision, read, and approved the submitted version.

Funding

This work was supported by the Natural Science Foundation of China under Grant No. 62271277 and the Zhejiang Natural Science Foundation of China under Grant No. LR22F020002.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Alain, M., Zerman, E., and Ozcinar, C. (2020). “Immersive imaging technologies: From capture to display,” in Proceedings of the 28th ACM International Conference on Multimedia (MM’20), (New York, NY: Association for Computing Machinery), 4787–4788. doi: 10.1145/3394171.3418550

CrossRef Full Text | Google Scholar

Besharse, J., and Bok, D. (2011). The Retina and its Disorders. Cambridge: Academic Press.

Google Scholar

Chen, M., Jin, Y., Goodall, T., Yu, X., and Bovik, A. C. (2019). Study of 3D virtual reality picture quality. IEEE J. Sel. Top. Signal Process. 14, 89–102. doi: 10.1109/JSTSP.2019.2956408

CrossRef Full Text | Google Scholar

Chen, Z., Xu, J., Lin, C., and Zhou, W. (2020). Stereoscopic omnidirectional image quality assessment based on predictive coding theory. IEEE J. Sel. Top. Signal Process. 14, 103–117. doi: 10.1109/JSTSP.2020.2968182

CrossRef Full Text | Google Scholar

Dendi, S., and Channappayya, S. S. (2020). No-reference video quality assessment using natural spatiotemporal scene statistics. IEEE Trans. Image Process. 29, 5612–5624. doi: 10.1109/TIP.2020.2984879

PubMed Abstract | CrossRef Full Text | Google Scholar

Deng, X., Wang, H., Xu, M., Guo, Y., Song, Y., and Yang, L. (2021). “LAU-Net: Latitude adaptive upscaling network for omnidirectional image super-resolution,” in 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), (New York, NY: IEEE), 9185–9194. doi: 10.1109/CVPR46437.2021.00907

CrossRef Full Text | Google Scholar

Duan, H., Zhai, G., Min, X., Zhu, Y., and Yang, X. (2018). “Perceptual quality assessment of omnidirectional images,” in 2018 IEEE International Symposium on Circuits and Systems (ISCAS), (New York, NY: IEEE), 1–5. doi: 10.1109/ISCAS.2018.8351786

CrossRef Full Text | Google Scholar

Gotz-Hahn, F., Hosu, V., Lin, H., and Saupe, D. (2021). KonVid-150k: A dataset for no-reference video quality assessment of videos in-the-wild. IEEE Access 9, 72139–72160. doi: 10.1109/ACCESS.2021.3077642

CrossRef Full Text | Google Scholar

Gu, K., Zhai, G., Yang, X., and Zhang, W. (2015). Using free energy principle for blind image quality assessment. IEEE Trans. Multimed. 17, 50–63. doi: 10.1109/TMM.2014.2373812

CrossRef Full Text | Google Scholar

Ho, T., and Budagavi, M. (2017). “Dual-fisheye lens stitching for 360-degree imaging,” in 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), (New York, NY: IEEE), 2172–2176. doi: 10.1109/ICASSP.2017.7952541

CrossRef Full Text | Google Scholar

Horé, A., and Ziou, D. (2010). “Image quality metrics: PSNR vs. SSIM,” in 2010 20th International Conference on Pattern Recognition, (New York, NY: IEEE), 2366–2369. doi: 10.1109/ICPR.2010.579

CrossRef Full Text | Google Scholar

ISO (2019). Information Technology Coded Representation of Immersive Media—Part 2: Omnidirectional Media Format. Geneva: ISO.

Google Scholar

Jiang, H., Jiang, G., Yu, M., Zhang, Y., Yang, Y., Peng, Z., et al. (2021). Cubemap-based perception-driven blind quality assessment for 360-degree images. IEEE Trans. Image Process. 30, 2364–2377. doi: 10.1109/TIP.2021.3052073

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim, H. G., Lim, H. T., and Yong, M. R. (2020). Deep virtual reality image quality assessment with human perception guider for omnidirectional image. IEEE Trans. Circuits Syst. Video Technol. 30, 917–928. doi: 10.1109/TCSVT.2019.2898732

CrossRef Full Text | Google Scholar

Lasmar, N. E., Stitou, Y., and Berthoumieu, Y. (2009). “Multiscale skewed heavy tailed model for texture analysis,” in 2009 16th IEEE International Conference on Image Processing (ICIP), (New York, NY: IEEE), 2281–2284. doi: 10.1109/ICIP.2009.5414404

CrossRef Full Text | Google Scholar

Li, C., Xu, M., Du, X., and Wang, Z. (2018). “Bridge the gap between VQA and human behavior on omnidirectional video,” in ACM Multimedia Conference Proceedings-American Computer Association, (New York, NY: Association for Computing Machinery), 932–940. doi: 10.1145/3240508.3240581

CrossRef Full Text | Google Scholar

Li, C., Xu, M., Jiang, L., Zhang, S., and Tao, X. (2019). Viewport proposal CNN for 360° video quality assessment. 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New York: IEEE 10169–10178

Google Scholar

Li, Q., Lin, W., Xu, J., and Fang, Y. (2016). Blind image quality assessment using statistical structural and luminance features. IEEE Trans. Multimed. 18, 2457–2469. doi: 10.1109/TMM.2016.2601028

CrossRef Full Text | Google Scholar

Ling, J., Zhang, K., Zhang, Y., Yang, D., and Chen, Z. (2018). A saliency prediction model on 360-degree images using color dictionary based sparse representation. Signal Process. Image Commun. 69, 60–68. doi: 10.1016/j.image.2018.03.007

CrossRef Full Text | Google Scholar

Liu, L., Liu, B., Huang, H., and Bovik, A. C. (2014). No-reference image quality assessment based on spatial and spectral entropies. Signal Process. Image Commun. 29, 856–863. doi: 10.1016/j.image.2014.06.006

CrossRef Full Text | Google Scholar

Liu, L., Yi, H., Zhao, Q., Hua, H., and Bovik, A. C. (2016). Blind image quality assessment by relative gradient statistics and adaboosting neural network. Signal Process. Image Commun. 40, 1–15. doi: 10.1016/j.image.2015.10.005

CrossRef Full Text | Google Scholar

Liu, Y., Yu, H., Huang, B., Yue, G., and Song, B. (2021). Blind omnidirectional image quality assessment based on structure and natural features. IEEE Trans. Instrument. Meas. 70, 1–11. doi: 10.1109/TIM.2021.3102691

CrossRef Full Text | Google Scholar

Min, X., Zhai, G., Gu, K., Liu, Y., and Yang, X. (2018). Blind image quality estimation via distortion aggravation. IEEE Trans. Broadcast. 64, 508–517. doi: 10.1109/TBC.2018.2816783

CrossRef Full Text | Google Scholar

Mittal, A., Moorthy, A. K., and Bovik, A. C. (2012). No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 21, 4695–4708. doi: 10.1109/TIP.2012.2214050

PubMed Abstract | CrossRef Full Text | Google Scholar

Mittal, A., Saad, M. A., and Bovik, A. C. (2016). A completely blind video integrity oracle. IEEE Trans. Image Process. 25, 289–300. doi: 10.1109/TIP.2015.2502725

PubMed Abstract | CrossRef Full Text | Google Scholar

Mittal, A., Soundararajan, R., and Bovik, A. C. (2013). Making a “completely blind” image quality analyzer. IEEE Signal Process. Lett. 20, 209–212. doi: 10.1109/LSP.2012.2227726

CrossRef Full Text | Google Scholar

Moorthy, K., and Bovik, A. C. (2011). Blind image quality assessment: From natural scene statistics to perceptual quality. IEEE Trans. Image Process. 20, 3350–3364. doi: 10.1109/TIP.2011.2147325

PubMed Abstract | CrossRef Full Text | Google Scholar

Petkov, N., and Subramanian, E. (2007). Motion detection, noise reduction, texture suppression, and contour enhancement by spatiotemporal Gabor filters with surround inhibition. Biol. Cybern. 97, 423–439. doi: 10.1007/s00422-007-0182-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Rai, Y., Callet, P. L., and Guillotel, P. (2017). “Which saliency weighting for omnidirectional image quality assessment,” in 2017 Ninth International Conference on Quality of Multimedia Experience (QoMEX), (Erfurt: IEEE), 1–6. doi: 10.1109/QoMEX.2017.7965659

CrossRef Full Text | Google Scholar

Sheikh, H. R., Sabir, M. F., and Bovik, A. C. (2006). A statistical evaluation of recent full reference image quality assessment algorithms. IEEE Trans. Image Process. 15, 3440–3451. doi: 10.1109/TIP.2006.881959

PubMed Abstract | CrossRef Full Text | Google Scholar

Simoncelli, E. P., and Olshausen, B. A. (2001). Natural image statistics and neural representation. Annu. Rev. Neurosci. 24, 1193–1216. doi: 10.1146/annurev.neuro.24.1.1193

PubMed Abstract | CrossRef Full Text | Google Scholar

Sitzmann, V., Serrano, A., Pavel, A., Agrawala, M., and Wetzstein, G. (2018). Saliency in VR: how do people explore virtual environments. IEEE Trans. Vis. Comput. Graph. 24, 1633–1642. doi: 10.1109/TVCG.2018.2793599

PubMed Abstract | CrossRef Full Text | Google Scholar

Sui, X., Ma, K., Yao, Y., and Fang, Y. (2022). Perceptual quality assessment of omnidirectional images as moving camera videos. IEEE Trans. Vis. Comput. Graph. 28, 3022–3034. doi: 10.1109/TVCG.2021.3050888

PubMed Abstract | CrossRef Full Text | Google Scholar

Sun, W., Gu, K., Ma, S., Zhu, W., Liu, N., and Zhai, G. (2018). “A large-scale compressed 360-degree spherical image database: From subjective quality evaluation to objective model comparison,” in 2018 IEEE 20th International Workshop on Multimedia Signal Processing (MMSP), (New York, NY: IEEE), 1–6. doi: 10.1109/MMSP.2018.8547102

CrossRef Full Text | Google Scholar

Sun, W., Min, X., Zhai, G., Gu, K., and Ma, S. (2019). MC360IQA: a multi-channel cnn for blind 360-degree image quality assessment. IEEE J. Sel. Top. Signal Process. 14, 64–77. doi: 10.1109/JSTSP.2019.2955024

CrossRef Full Text | Google Scholar

Sun, Y., Lu, A., and Yu, L. (2017). Weighted-to-spherically-uniform quality evaluation for omnidirectional video. IEEE Signal Process. Lett. 24, 1408–1412. doi: 10.1109/LSP.2017.2720693

CrossRef Full Text | Google Scholar

Tran, H. T. T., Nguyen, D. V., Ngoc, N. P., Hoang, T. H., Huong, T. T., and Thang, T. C. (2019). Impacts of retina-related zones on quality perception of omnidirectional image. IEEE Access 7, 166997–167009. doi: 10.1109/ACCESS.2019.2953983

CrossRef Full Text | Google Scholar

Tu, Z., Chen, C. J., Chen, L. H., Birkbeck, N., Adsumilli, B., and Bovik, A. C. (2020). “A comparative evaluation of temporal pooling methods for blind video quality assessment,” in 2020 IEEE International Conference on Image Processing (ICIP), (New York, NY: IEEE), 141–145. doi: 10.1109/ICIP40778.2020.9191169

CrossRef Full Text | Google Scholar

Video Quality Experts Group (2003). Final Report from the Video Quality Experts Group on the Validation of Objective Models of Video Quality Assessment, Phase II. San Jose, CA: Video Quality Experts Group.

Google Scholar

Wang, Z., Simoncelli, E. P., and Bovik, A. C. (2003). Multiscale structural similarity for image quality assessment. The Thrity-Seventh Asilomar Conference on Signals, Systems Computers, 2, 1398–1402. doi: 10.1109/ACSSC.2003.1292216

CrossRef Full Text | Google Scholar

Xia, Y., Wang, Y., and Wang, C. (2021). Phase consistency guided full-reference panoramic image quality assessment algorithm. J. Image Graph. 26, 1625–1636. doi: 10.11834/JIG.200546

CrossRef Full Text | Google Scholar

Xu, J., Luo, Z., Zhou, W., Zhang, W., and Chen, Z. (2019). “Quality assessment of stereoscopic 360-degree images from multi-viewports,” in 2019 Picture Coding Symposium (PCS), (New York, NY: IEEE), 1–5. doi: 10.1109/PCS48520.2019.8954555

CrossRef Full Text | Google Scholar

Xu, J., Zhou, W., and Chen, Z. (2020). Blind omnidirectional image quality assessment with viewport oriented graph convolutional networks. IEEE Trans. Circuits Syst. Video Technol. 31, 1724–1737. doi: 10.1109/TCSVT.2020.3015186

CrossRef Full Text | Google Scholar

Xu, M., Li, C., Zhang, S., and Callet, P. L. (2020). State-of-the-art in 360° video/image processing: perception, assessment and compression. IEEE J. Sel. Top. Signal Process. 14, 5–26. doi: 10.1109/JSTSP.2020.2966864

CrossRef Full Text | Google Scholar

Xue, W., Zhang, L., Mou, X., and Bovik, A. C. (2014). Gradient magnitude similarity deviation: a highly efficient perceptual image quality index. IEEE Trans. Image Process. 23, 668–695. doi: 10.1109/TIP.2013.2293423

PubMed Abstract | CrossRef Full Text | Google Scholar

Yan, J., Li, J., and Fu, X. (2019). No-reference quality assessment of contrast-distorted images using contrast enhancement. arXiv [Preprint]. doi: 10.48550/arXiv.1904.08879

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, L., Xu, M., Xin, D., and Feng, B. (2021). “Spatial attention-based non-reference perceptual quality prediction network for omnidirectional images,” in 2021 IEEE International Conference on Multimedia and Expo (ICME), (Shenzhen: IEEE), 1–6. doi: 10.1109/ICME51207.2021.9428390

CrossRef Full Text | Google Scholar

Yu, M., Lakshman, H., and Girod, B. (2015). “A Framework to Evaluate Omnidirectional Video Coding Schemes,” in 2015 IEEE International Symposium on Mixed and Augmented Reality, (New York, NY: IEEE), 31–36. doi: 10.1109/ISMAR.2015.12

CrossRef Full Text | Google Scholar

Zakharchenko, V., Choi, K. P., Alshina, E., and Park, J. H. (2017). “Omnidirectional Video Quality Metrics and Evaluation Process,” in 2017 Data Compression Conference (DCC), (New York, NY: IEEE), 472–472. doi: 10.1109/DCC.2017.90

CrossRef Full Text | Google Scholar

Zanca, D., Gori, M. (2017). “Variational laws of visual attention for dynamic scenes,” in Advances in Neural Information Processing Systems 30. (eds), I. Guyon and U. V. Luxburg and S. Bengio and H. Wallach and R. Fergus and S. Vishwanathan and R. Garnett, (New York, NY: Acm Digital Library), 3826–3835. doi: 10.5555/3294996.3295139

CrossRef Full Text | Google Scholar

Zanca, D., Melacci, S., and Gori, M. (2020). Gravitational laws of focus of attention. IEEE Trans. Pattern Anal. Mach. Intel. 42, 2983–2995. doi: 10.1109/TPAMI.2019.2920636

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, L., Zhang, L., Mou, X., and Zhang, D. (2011). FSIM: a feature similarity index for image quality assessment. IEEE Trans. Image Process. 20, 2378–2386. doi: 10.1109/TIP.2011.2109730

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, Y., Wang, Y., Liu, F., Liu, Z., Li, Y., Yang, D., et al. (2018). Subjective panoramic video quality assessment database for coding applications. IEEE Trans. Broadcast. 64, 461–473. doi: 10.1109/TBC.2018.2811627

CrossRef Full Text | Google Scholar

Zheng, X., Jiang, G., Yu, M., and Jiang, H. (2020). Segmented spherical projection-based blind omnidirectional image quality assessment. IEEE Access 8, 31647–31659. doi: 10.1109/ACCESS.2020.2972158

CrossRef Full Text | Google Scholar

Zhou, W., Bovik, A. C., Sheikh, H. R., and Simoncelli, E. P. (2004). Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13, 600–612. doi: 10.1109/TIP.2003.819861

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhou, W., Xu, J., Jiang, Q., and Chen, Z. (2022). No-reference quality assessment for 360-degree images by analysis of multifrequency information and local-global naturalness. IEEE Trans. Circuits Syst. Video Technol. 32, 1778–1791. doi: 10.1109/TCSVT.2021.3081182

CrossRef Full Text | Google Scholar

Zhou, Y., Sun, Y., Li, L., Gu, K., and Fang, Y. (2022). Omnidirectional image quality assessment by distortion discrimination assisted multi-stream network. IEEE Trans. Circuits Syst. Video Technol. 32, 1767–1777.

Google Scholar

Zhou, Y., Yu, M., Ma, H., Shao, H., and Jiang, G. (2018). “Weighted-to-Spherically-Uniform SSIM Objective Quality Evaluation for Panoramic Video,” in 2018 14th IEEE International Conference on Signal Processing (ICSP), (New York, NY: IEEE), 54–57. doi: 10.1109/ICSP.2018.8652269

CrossRef Full Text | Google Scholar

Zhu, W., Zhai, G., Min, X., Hu, M., Liu, J., Guo, G., et al. (2019). Multi-channel decomposition in tandem with free-energy principle for reduced-reference image quality assessment. IEEE Trans. Multimedia. 21, 2334–2346. doi: 10.1109/TMM.2019.2902484

CrossRef Full Text | Google Scholar

Zhu, Y., Zhai, G., Min, X., and Zhou, J. (2020). The prediction of saliency map for head and eye movements in 360 degree images. IEEE Trans Multimedia. 22, 2331–2344. doi: 10.1109/TMM.2019.2957986

CrossRef Full Text | Google Scholar

Zou, W., Zhang, W., and Yang, F. (2021). Modeling the perceptual quality for viewport-adaptive omnidirectional video streaming considering dynamic quality boundary artifact. IEEE Trans. Circuits Syst. Video Technol. 31, 4241–4254. doi: 10.1109/TCSVT.2021.3050157

CrossRef Full Text | Google Scholar

Keywords: omnidirectional images, image quality assessment, no-reference, spatiotemporal scene statistics, virtual reality

Citation: Wang Y, Li H and Jiang Q (2022) Dynamically attentive viewport sequence for no-reference quality assessment of omnidirectional images. Front. Neurosci. 16:1022041. doi: 10.3389/fnins.2022.1022041

Received: 18 August 2022; Accepted: 13 October 2022;
Published: 23 November 2022.

Edited by:

Qingbo Wu, University of Electronic Science and Technology of China, China

Reviewed by:

Leida Li, Xidian University, China
Guanghui Yue, Shenzhen University, China

Copyright © 2022 Wang, Li and Jiang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Hong Li, ky_lihong@nbu.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.