Next Article in Journal
A Wearable-Sensor System with AI Technology for Real-Time Biomechanical Feedback Training in Hammer Throw
Next Article in Special Issue
Optical Panel Inspection Using Explicit Band Gaussian Filtering Methods in Discrete Cosine Domain
Previous Article in Journal
Current Technologies for Detection of COVID-19: Biosensors, Artificial Intelligence and Internet of Medical Things (IoMT): Review
Previous Article in Special Issue
Inline Quality Monitoring of Reverse Extruded Aluminum Parts with Cathodic Dip-Paint Coating (KTL)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Conv-Former: A Novel Network Combining Convolution and Self-Attention for Image Quality Assessment

1
Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China
2
College of Materials Science and Opto-Electronic Technology, University of Chinese Academy of Sciences, Beijing 100049, China
3
Department of Electrical and Optical Engineering, Space Engineering University, Beijing 101416, China
*
Author to whom correspondence should be addressed.
Submission received: 7 November 2022 / Revised: 21 December 2022 / Accepted: 26 December 2022 / Published: 30 December 2022

Abstract

:
To address the challenge of no-reference image quality assessment (NR-IQA) for authentically and synthetically distorted images, we propose a novel network called the Combining Convolution and Self-Attention for Image Quality Assessment network (Conv-Former). Our model uses a multi-stage transformer architecture similar to that of ResNet-50 to represent appropriate perceptual mechanisms in image quality assessment (IQA) to build an accurate IQA model. We employ adaptive learnable position embedding to handle images with arbitrary resolution. We propose a new transformer block (TB) by taking advantage of transformers to capture long-range dependencies, and of local information perception (LIP) to model local features for enhanced representation learning. The module increases the model’s understanding of the image content. Dual path pooling (DPP) is used to keep more contextual image quality information in feature downsampling. Experimental results verify that Conv-Former not only outperforms the state-of-the-art methods on authentic image databases, but also achieves competing performances on synthetic image databases which demonstrate the strong fitting performance and generalization capability of our proposed model.

1. Introduction

Image quality assessment (IQA) is a branch of computer vision research that aims to give computers the same ability to judge image quality as humans. IQA is a crucial research area since it can be applied to image restoration and used to distinguish between distinct visual perception experiences [1]. These days, digital images captured by cameras on cell phones, professional imaging equipment, and remote sensing satellites are widely used [2]. Digital images introduce noise signals in the acquisition process, and they can lose perceptual information during compression, storage, and transmission [3]. As a result, IQA is becoming increasingly important and has numerous applications, such as designing image restoration models to remove blur, noise, clouds, and other artifacts from images using IQA algorithms. It can also assist image acquisition equipment in evaluating and debugging the product’s imaging parameters and determining whether the imaging system is degraded. Although subjective human assessment can truly reflect human visual perception, and the assessment results are direct, accurate, and reliable, which is the ultimate basis for judging image quality, the implementation process is time-consuming and costly. It is also easily influenced by personal subjective emotions and preferences, and it has many limitations in meeting practical application needs [4]. To promote the use of IQA in real-world engineering, accurate and effective objective IQA algorithms must be developed.
The objective IQA method is an evaluation method that builds a mathematical model based on the visual system of the human eye and scores the image to be measured. This method is low-cost, has the advantages of batch processing and reproducible results, and can be more easily applied to a variety of scenarios. Traditionally, objective IQA algorithms are classified into three categories based on whether a reference image is required or not: full reference IQA (FR-IQA) [5], reduced-reference IQA (RR-IQA) [6], and no reference IQA (NR-IQA) [7]. The full-reference image quality evaluation method uses the image under test to compare with the original reference image in an adequately defined image space; the reduced-reference image quality evaluation does not use the original image directly but uses some information of the original image such as structural features and distortion types to evaluate the image quality, and the no-reference image quality evaluation does not use any original reference image information in the evaluation process. The main goal of all three image quality evaluation methods is to predict a quality score relevant to human visual perception. NR-IQA has attracted a large number of researchers’ attention in recent years because no reference information is available or may not even exist in many realistic situations. NR-IQA has been a challenging problem that has not been well addressed by many methods.
Early models for NR-IQA were based on manual feature extraction [8,9,10], which relied heavily on our summarized knowledge of the probabilistic architecture of the visual world, the mechanisms of image degradation, and the composition of the human visual system (HVS). In recent years with the development of convolutional neural networks (CNN), more and more vision tasks have benefited from this [11,12,13]. Current deep learning-based image quality assessment methods have achieved remarkable success in extracting visual features using CNN [4,14], which have the following advantages over manual feature extraction:
(1)
By carefully designing a deep neural network with the problem to be solved and the input data, it is possible to automatically learn the relationships implicit within the data from the training dataset without the need for tedious manual feature extraction;
(2)
Deep neural network models can contain thousands of parameters; thus, deep features can have better differentiation and representation capabilities. Compared with manually extracted features, it has more prominent advantages in extracting multi-level features and contextual information of images;
(3)
Deep learning can change the model architecture by simply adjusting the parameters, which enables the network to automatically model itself according to the specific characteristics of the task, with good generalization and efficiency.
In recent years, Self-Attention based transformer architecture has achieved great success in the field of natural language processing (NLP) by establishing a long-range interaction in a scalable manner [15,16,17,18] and continues to make constant and ground-breaking progress in building on this foundation. Researchers have applied the transformer directly to computer vision by slicing images into patches, such as ViT [19] and DETR [20]. The transformer has shown great promise in the field of computer vision, outperforming CNN in various mainstream tasks such as image classification [11] and target detection [12], and is likely to replace CNN as the new backbone in the future, mainly due to the transformer’s ability to capture long-range pixel interactions and aggregate global information from the entire input sequence.
Transformer mainly uses Multi-Headed Self-Attention (MHSA) to model long-range interactions, as shown in Figure 1a below is the schematic architecture of MHSA. The input image X i m a g e H × W × C is cropped into N = H W / P 2 image blocks of size P × P × C , and each image block is expanded into a one-dimensional vector to finally obtain X N × p 2 c , where H, and W represent the length and width of the image, respectively, N represents the number of image blocks, i.e., the number of tokens, and d = P 2 C represents the feature dimension of each token. As in Equation (1) X is linearly transformed, and together with the class token a new vector is formed as the input to the transformer encoder.
X i n p u t = [ X c l a s s ; X 1 W ; X 2 W ; X N W ] + E p o s
where X c l a s s is the class token added to implement the classification task, W is the matrix that implements the linear mapping, and E p o s is the position embedding.
Implementing Self-Attention requires defining three key elements. The query Q = X W Q , the key K = X W K , and the value V = X W V . Where W Q , W K , W V d × d is the weight matrix that implements the linear mapping. The output Z can be formulated as:
Z = softmax ( Q K T / d ) V = Attention ( Q , K , V )
where d represents approximate normalization and the matrix product of Q K T calculates the similarity between each pair of tokens to achieve Self-Attention. As shown in Figure 1b, the input X is i-quantized, i.e., X = [ x 1 , x 2 , , x i ] , and x i N × d i is entered separately into the Self-Attention module to obtain the multi-headed attention values by the connected single attention values Z = [ z 1 , z 2 , , z i ] . The formulation can be further rewritten as:
Z = [ Attention ( x i w i Q , x i w i K , x i w i V ) ] 1 : i W o
where Attention(⋅) is the standard self-attentive function based on QKV, [ ] 1 : i indicating that the results obtained are concatenated.
Although the transformer has made breakthroughs in computer vision tasks, there is still a significant performance and computational cost gap between the simple multilayer transformer encoder architecture and the previous convolutional neural network, making the transformer unavailable on limited hardware resources. Figure 2 shows an overview of the ResNet50 [21], the vision transformer [19], and the multi-stage transformer [22]. The main idea of ResNet50 is to divide the feature map extraction into different stages, so that, given an input image, we can generate feature maps at different scales, which has proved to be very useful in many intensive prediction tasks [21,23]. The image is split directly into non-overlapping patches in the vision transformer and then fed into the transformer encoder after linear projection. Inspired by the two architectures mentioned above, The researchers designed a novel transformer network architecture to generate hierarchical feature representations like ResNet50, which usually apply a pooling layer before each stage to reduce the size of intermediate features by 2x downsampling, and stack several encoder blocks in each stage. Compared to traditional vision transformer architecture, this multi-stage transformer network can significantly reduce the number of parameters, allowing us to train and extract multi-scale feature representations that have been shown to be beneficial for many vision tasks. We can obtain four hierarchical feature maps at different resolutions, similar to a typical convolutional neural network. Our proposed Conv-Former network also exploits the advantages of this hierarchical architecture and improves upon it.
In this paper, we aim to design an IQA model (Conv-Former) using the long-range interaction capabilities of a transformer and the local feature extraction capabilities of CNN. The network can give predictions that are more consistent with human visual system perception. Therefore, we introduce a multi-stage transformer network architecture to our IQA model. Using a local information perception module and a transformer to capture local information-aware features and global semantic features in an image, the network is able to collect fine-grained detail and global information using both local and global features. The dual-path pooling allows the multi-stage transformer to capture as much contextual information as possible, and the Conv-Former has experimentally proven to be highly capable of local feature perception and image content understanding, both of which are quite important in IQA tasks. The main contributions of this paper are as follows:
(1)
We designed an end-to-end neural network model called Conv-Former for no-reference image quality assessment. The overall architecture uses a multi-stage architecture similar to that of ResNet-50 to obtain multi-scale features, which can significantly reduce the number of parameters compared to the traditional transformer architecture. At the same time, multi-scale features are more conducive to the extraction of image quality features. This architecture enables the generation of appropriate perceptual mechanisms in image quality assessment to build an accurate IQA model;
(2)
In this work, we introduce an effective hybrid architecture for image quality assessment networks that utilize local information from CNNs and global semantic information captured by the transformer to further improve the accuracy of IQA, implemented by replacing the linear layer that generates the qkv matrix with a local information-aware module that is able to further obtain local information in image quality, acquire fine-grained features and obtain detailed and overall information representation in the image. Network analysis experiments also demonstrate that the network outperforms other models for understanding the content of the input images. This enables the neural network to focus better on the subtle differences between images and thus obtain a more accurate image quality score. In order to reduce the image quality information loss in the feature downsampling process under multi-stage architecture, we designed the dual path pooling module to keep more contextual information;
(3)
The position embedding of traditional transformer networks cannot adapt to the input of different resolution images and the use of local information perception modules. Therefore, this paper proposes an adaptive 2D position embedding module, which solves the problem that traditional CNN networks cannot input images with different resolutions, and at the same time, the 2D position embedding is more in line with the characteristics of images. It can effectively represent the position information between tokens;
(4)
We experimented with Conv-Former on two different authentic image quality assessment datasets, LIVE Challenge (LIVEC) and KonIQ-10k, as well as on synthetic datasets LIVE, TID2013, and CSIQ, and compared the performance of the algorithm on different distortion types. The extensive experimental results show that Conv-Former has competitive results, which demonstrate the strong fitting performance and generalization capability of our proposed model. As shown in Figure 3, we can find that the results of Conv-Former are more in line with the Mean Opinion Score (MOS).

2. Related Work

2.1. Attention Mechanism in CNN

The attention mechanism is an important feature of the human visual system, which means that only a portion of all visible information is noticed by humans. The attention mechanism is introduced in CNNs by simulating the human visual perception process, which can ignore the interference of irrelevant information and thus improve the generalization performance of the network. This has enabled CNNs to make breakthroughs in areas such as object detection, image generation, and target tracking.
The attention mechanism was originally used to encode long input sentences as part of the encoder-decoder framework in recurrent neural networks (RNN) and has since been widely used in RNN [24]. Attention is widely used to enhance the representation of features. For example, Hu et al. [25] propose that SENet uses channel attention to explicitly model the interdependencies between feature maps and adaptively acquire the importance of each feature map by learning and then updating the original data based on this importance. In this way, SENet increases the importance of features that are more useful for the task and decreases the importance of useless features to achieve better results. By embedding this module into other networks, the computational resources of the neural network can be more rationally allocated with a small increase in the cost of the number of parameters, resulting in a significant improvement in network performance. Wang et al. [26] proposed efficient channel attention by improving the SENet, which is a local cross-channel interaction strategy without dimensionality reduction and an adaptive selection of the one-dimensional convolutional kernel size to obtain more accurate attention information by aggregating cross-channel information through a one-dimensional convolutional layer. Convolutional Block Attention Module (CBAM) [27] is constructed by combining the spatial attention module (SAM) and the channel. The CBAM is built by combining the spatial attention module (SAM) and the channel attention module (CAM), aggregating attention information from both spatial and channel aspects respectively, and fusing the information to a certain extent to obtain more comprehensive and reliable attention information and provide more appropriate guidance on the allocation of computational resources. Based on CBAM, fu et al. [28] proposed DA-Net, which also integrates channel attention and spatial attention. Unlike CBAM, where the acquisition of attention information in both directions is parallel, DA-Net captures global feature dependencies in the spatial and channel dimensions, using a spatial attention module to learn spatial interdependencies of features and a channel attention module to model channel interdependencies.

2.2. No-Reference Image Quality Assessment

No-reference image quality assessment means that no reference image is required, and the quality is assessed only based on the distorted image’s characteristics, which is also more in line with practical needs, as reference images are difficult to obtain or do not exist in practical applications. NR-IQA methods can be divided into two categories, distortion type specific IQA methods [29,30,31] and generic IQA methods [8,9,10]. Distortion type specific IQA methods are designed based on specific distortion types, such as noise, JPEG compression artifacts, blurred artifacts, and other distortion types, and these methods design specific feature extraction methods by looking at the histogram of the pixel distribution of the image after distortion, after which the quality prediction score of the image is obtained. However, this method is limited in that it can only detect quality losses caused by specific distortions and is therefore not widely used. The generic NR-IQA method is more effective because the image distortion type is usually unknown in advance.
As shown in Figure 4, most of the traditional NR-IQA methods are based on natural scene statistics (NSS) methods, which first manually extract features from distorted images and then use probabilistic or regression models for quality prediction of distorted images. In early NSS-based NR-IQA methods, features are extracted from transform domains such as wavelet or cosine transform domains.
Moorth et al. [32,33] proposed a class of image quality evaluation methods divided into two stages, first identifying distortion types and then performing the distortion-specific quality assessment. For example, Blind Image Quality Index (BIQI) [32], an image authenticity and integrity image quality evaluation algorithm based on distortion type identification (DIIVINE) [33], both of which are based on training a support vector machine (SVM) to obtain a classifier for the image distortion type, then extracting the image features and relying on a support vector regressor to regress the quality prediction scores for each DIIVINE improves the process of extracting image features based on BIQI, uses NSS to estimate the coefficient distribution of wavelets, and extracts global features to determine the image quality score. Saad et al. proposed a blind image integrity labeling algorithm based on discrete cosine transform statistics (BLIINDS) [34], which extracts features from the DCT domain and then uses a multivariate Gaussian model to obtain the quality scores of distorted images. Saad et al. later proposed an optimized BLIINDS (BLIINDS-II) [35] by extracting more complex DCT features, using a generalized Gaussian mixture model to fit different multiscale discrete cosine transform coefficient distributions as frequency domain statistics and a Bayesian which wasmore time-consuming because of the transformation of the image domain. To avoid the transformation of the domain, methods based on spatial domain features have emerged. Mittal et al. [8] proposed a spatial domain non-reference image quality assessment algorithm (BRISQUE), which uses the local mean and variance of the image to calculate the local normalized brightness of the image, and then uses a generalized Gaussian model to model the local normalized brightness distribution as its spatial domain natural statistical features to obtain the prediction score. Ye et al. [36] proposed a codebook-based manual feature extraction NR-IQA algorithm, which uses a K-mean clustering method to learn codebooks directly from training image blocks, then codebooks are used to encode on test images to obtain features of the images, and finally, SVR is used to predict the quality scores of distorted images. Zhang et al. [37] used these features to extract salient regions of semantic objects for quality estimation. Xu et al. [9] improved the feature set and predicted the quality score by merging the higher-order statistical information of the images. However, these manual feature extraction methods require specialized design and are very time-consuming. In addition, scene statistical features characterize image quality from a global perspective and thus cannot measure local distortions common in real distorted images.
Inspired by the breakthroughs in deep learning for other vision tasks [11,12,13,21], researchers have proposed several learning-based methods for image quality assessment that extract quality-related image features and automatically learn correction parameters through deep learning. Thus, better results than traditional manual feature extraction are obtained.

3. Proposed Method

In this section, we describe in detail the architecture of our proposed Conv-Former model and outline the specific roles each block plays. First, we describe the overview of the Conv-Former block. On this basis, the key modules of the algorithm are described, such as the local information-awareness module and the adaptive position embedding.

3.1. Overview

An overview of the model is depicted in Figure 5. The processing of the input image can be divided into three stages and let the output features of each stage be F 1 , F 2 and F 3 respectively. Each feature consists of spatial tokens [ X s p a t i a l 1 , , X s p a t i a l 3 ] and classification token [ X c l a s s 1 , , X c l a s s 3 ] and can be expressed by Equation (4). In this paper, the feature channel numbers D1, D2, and D3 are taken to be 192, 384, and 512 respectively.
F 1 = [ X c l a s s 1 ; X s p a t i a l 1 ] , X c l a s s 1 1 × D 1 , X s p a t i a l 1 F 2 = [ X c l a s s 2 ; X s p a t i a l 2 ] , X c l a s s 2 1 × D 2 , X s p a t i a l 2 F 3 = [ X c l a s s 3 ; X s p a t i a l 3 ] , X c l a s s 3 1 × D 3 , X s p a t i a l 3
First, we reshape the input image X H × W × C to the feature map F 1 H p × W p × D 1 by convolutional feature extraction, where (H, W) is the resolution of the original image, C is the number of channels, p is the resolution decay after the convolution operation, D 1 is the dimension of the feature map F 1 , and N = H W / P 2 is the resulting number of tokens, which serves as the effective input sequence length for the Transformer. The Adaptive Position Embedding described in Section 3.2 is added to the patch embeddings to retain positional information. As shown in Equation (6), The process from feature F 1 to F 2 is similar to the process from F 2 to F 3 , with three layers of transformer modules and a layer of Dual path pooling(DDP) modules for 2x down adoption in between, reducing the resolution of the features while increasing the number of feature channels, and after obtaining feature F 3 , the classification token containing the image quality information is fed alone into the MLP to obtain the final image quality assessment score.
X F 1 = [ X c l a s s 1 ; C o n v ( X ) + E p o s ] , E p o s F 1 H × F 1 W , X c l a s s 1 1 × D 1 F 1 F 2 = D P P ( T B ( T B ( T B ( F 1 ) ) ) ) F 2 F 3 = D P P ( T B ( T B ( T B ( F 2 ) ) ) ) F 3 S c o r e = M L P ( X c l a s s 3 ) , X c l a s s 3 1 × D 3

3.2. Adaptive 2D Position Embedding

In CNN-based image quality evaluation models, the input images need to be resized or cropped to a fixed shape for batch training. However, this pre-processing changes the aspect ratio and composition of the image, which affects the image quality. However, by processing the position encoding part of the transformer-based network, it is possible to input images of any resolution into the network. No pre-processing of the input image is required, in line with the human visual system.
Position encoding is an integral part of the transformer architecture, through which the position encoding can be determined to explicitly model the position of the token and improve the representational power of the model. Its effectiveness has been well demonstrated in the field of natural language processing [15,16,18]. Since images can be considered as two-dimensional sequences, there is a need to extend the one-dimensional position encoding to two-dimensional position encoding, regardless of the input image size. The method mentioned in this section can effectively provide the position information required for object localization.
The specific implementation steps are as follows. Suppose the size of the input feature map is h × w × c , then we define a learnable parameter matrix L s × s , where the size of the matrix s is a hyperparameter, set to 10 in this paper, and we obtain the position code A h × w by adaptively deflating the learnable matrix, the size of the position code is consistent with the feature map, as shown in Figure 6, let ( h i , w j ) be a point on the position code matrix A, then the corresponding position code at that point can be determined by the following equation.
A ( h i , w j ) = L ( R o u n d ( s h × h i ) , R o u n d ( s w × w i ) )
where R o u n d ( ) stands for rounding the floating point number inside the brackets and A (   ,   ) , L (   ,   ) , represent the coded values at the corresponding positions respectively.

3.3. Transformer Block

Local features can be captured in CNN by convolutional operations, and although global features can be captured by continuously deepening the neural network, the global features suffer a significant loss in the process. With the advent of transformer, the long-range dependencies of token are captured by Self-Attention and multi-layer perceptron (MLP) architecture, but such architectures ignore local detail features. As shown in Figure 7, in order to combine the advantages of local features and global representations and thus improve the performance of the transformer network, we designed a novel local information perception (LIP) module to generate QKV that improves the discriminability between background and foreground. As shown in Equation (7), Let the input tensor X h × w × d be projected into the query vector Q L × d , key vector K L × d , and value tensor V L × d , where d is the dimension size of each token and L = h × w + 1 is the number of tokens.
Q = c o n c a t e [ f l a t t e n ( L I P Q ( X ) ) , T c l s ] K = c o n c a t e [ f l a t t e n ( L I P K ( X ) ) , T c l s ] V = c o n c a t e [ f l a t t e n ( L I P V ( X ) ) , T c l s ]
where T c l s stands for classification token.
L I P ( X ) = C o n v ( C o n v ( x ) + D W C o n v ( x ) )

3.4. Dual Path Pooling

For the feature downsampling at the end of each stage, we designed a dual-path pooling (DPP) layer, as shown in Figure 5. It consists of two branches: one is a 3 × 3 depthwise convolution with a step size of two; the other is a pooling layer and a 1 × 1 convolution. It is possible to achieve twice as much downsampling. During feature downsampling, the features on both paths are fused together by channel stacking to retain more contextual information. Experimental results show that DPP performs better than a direct maximum pooling layer. In equation terms, this can be described as follows.
I = c o n c a t ( D W c o n v ( x , 3 × 3 ) + c o n v ( m a x p o o l i n g ( x ) , 1 × 1 ) )

4. Experiments

4.1. Datasets

In this work, five widely used datasets in the field of image quality assessment were used, which can be split into authentic datasets and synthetic datasets based on the method of obtaining distorted images. The synthetic datasets include LIVE [38], TID2013 [39], and CSIQ [40]. The authentic distortion image dataset includes the LIVE Challenge (LIVEC) [41] and KonIQ-10k [42] datasets. A detailed description of them is given in Table 1.
The University of Texas at Austin’s Image and Video Engineering Laboratory established the LIVE image quality assessment dataset [38] in 2006. It consists of 779 distorted images developed from 29 source images using a total of five different forms of distortion (JP2K compression, JPEG compression, additive white Gaussian noise, Gaussian blur, and Simulated fast-fading Rayleigh channel). The scores are expressed by the Differential Mean Opinion Score (DMOS), the difference between the human eye’s evaluation score of the reference image and the distorted image, with lower values indicating higher image visual quality.
The TID2013 dataset is an extension of the TID2008 dataset [43] and contains 3000 distorted images based on 25 reference images with 24 different distortion types and five distortion levels. Image distortion categories include Additive Gaussian noise, Impulse noise, Chromatic aberrations, and so on. The Mean Opinion Score (MOS) values [0, 9] are employed. The higher the value, the greater the visual quality. Because the TID2013 dataset contains more types of distortion, it places more demands on the algorithm, and many traditional methods cannot be used effectively. In Figure 8, we compare the attention maps of the three different network architectures.
The Computational Perception and Image Quality Lab at Oklahoma State University created the CSIQ dataset [40], which contains 30 raw images and 866 images distorted by JPEG compression, JP2K compression, Gaussian blur, Gaussian white noise, Gaussian pink noise, or contrast variation, with five or four levels of each distortion type. The photos are 512 × 512 in size. The DMOS values acquired are in the [0, 1] range, with lower values suggesting greater visual quality. We show a selection of images from the dataset in Figure 9.
LIVE Challenge [41] contains 1162 images taken in a variety of natural environments, with complex losses due to the level of photography and imaging equipment used to capture them, typically a combination of overexposure or underexposure, blur, grain, or compression, with MOS ranging from [0, 100], the higher the value the better. We show a selection of images from the dataset in Figure 10.
The KonIQ-10k dataset consists of 10,073 images selected from the large public multimedia database YFCC100m [44]. The sampled images cover as wide and uniform a quality distribution as possible in terms of brightness, colour, contrast and sharpness, and the types of distortion present in these images include noise, JPEG artifacts, blending, lens motion blur, over-sharpening, and so on. The researchers conducted a large-scale crowdsourcing experiment based on the collected dataset, receiving 1.2 million assessments from 1467 observers utilizing statistical approaches such as taking the mean and deleting extreme scores to determine the final MOS values. The photos were 1024 × 768 in size. MOS values were in the [0, 5] range, with higher values indicating less distortion.

4.2. Evaluation Metrics

In order to quantitatively compare the performance of IQA algorithms, researchers often use the following three evaluation criteria.
(1)
Spearman rank-order correlation coefficient (SROCC), SRCC is used to measure the monotonicity of IQA algorithm predictions and is calculated as follows.
SRCC = 1 6 i d i 2 I ( I 2 1 )
where d i denotes the difference between the subjective quality score ranking of the i-th image and the objective quality score ranking, and I denotes the number of images in the test set.
(2)
The Pearson linear correlation coefficient (PLCC), PLCC is used to assess the accuracy and degree of linear correlation of IQA model predictions.
PLCC = i ( q i q m ) ( q ^ i q ^ m ) i ( q i q m ) 2 i ( q ^ i q ^ m ) 2
where q i and q ^ i denote the MOS value and algorithm prediction score of the i-th image, respectively, and q m q ^ m denote the mean MOS value and the mean algorithm prediction score of the test image samples, respectively.
(3)
The root mean square error (RMSE), RMSE is used to assess the consistency of the IQA model’s predictions. It is used to measure the absolute error between the algorithm’s predicted score and the subjective evaluation score and is calculated as follows.
RMSE = 1 n i = 1 N ( q i q ^ i ) 2

4.3. Implementation Details

In the experiments, for each dataset, 80% of the images were randomly selected for training and 20% for testing. The training is conducted using a SGD optimizer with a batch size of eight. We trained our models with an initial learning rate of 0.001, with a warm up cosine learning rate decay scheduler. We adopted MSE loss for training:
L r e g = m o d e l ( I D i s t ) s 2
where m o d e l ( I D i s t ) denotes the output of the proposed Conv-Former, s denotes the ground-truth normalized MOS or DMOS value. We implemented our proposed model Conv-Former in Pytorch version 1.12.0 and python version 3.9, which was trained using a single NVIDIA GeForce RTX 3090 GPU. The CUDA and CuDNN versions are 11.6 and 8.4.0 respectively.

4.4. Comparing with the State-of-The Art (SOTA)

We assessed the performance of our model with PLCC and SRCC. PLCC assesses the linear correlation between ground truth and the predicted quality scores, whereas SRCC describes the level of monotonic correlation.
We evaluated the effectiveness of Conv-Former on five benchmark datasets. For all of our tests, we followed the above experimental setup. It can be shown in Table 2 that Conv-Former outperforms or is competitive with 14 NR-IQA methods: BRISQUE, NIQE, DIIVINE, HOSA, WaDIQaM, BIECON, SFA, PQR, DBCNN, SHN, RankIQA, ResNet-ft, TRIQ and MUSIQ. We found that our method achieves the best PLCC/SRCC results in comparison to other works. Especially on the moe complex dataset TID2013, our proposed model achieved a solid improvement over previous work. Even though Conv-Former achieves 0.965 on PLCC and 0.964 on SRCC, which means the metric is consistent with the human perspective. The effective feature fusion by CNN and ViT and the proposed multi-scale prediction module make our method substantially superior to other transformer-based image quality assessment network; examples include TRIQ and MUSIQ. Although the model achieved better results on more complex datasets, there was no major improvement for datasets where most algorithms performed well, such as the LIVE dataset. In order to visualise the advantages of Conv-Former, we present the data in Table 2 as a histogram in Figure 11.
In order to evaluate the effectiveness of Conv-Former on different types of distorted images, we also compare the PLCC/SRCC performances on five kinds (JP2K compression, JPEG compression, additive white Gaussian noise, Gaussian blur, Simulated fast fading Rayleigh channel) of distorted types in the LIVE dataset and six kinds (JPEG compression, JP2K compression, Gaussian blur, Gaussian white noise, Gaussian pink noise, contrast variation) of distorted types in the CSIQ dataset. Table 3 and Table 4 show the results of different IQA methods on different types of distorted images. In the tables, we find our method achieves the best or most competitive performances on all different types of distorted images than other IQA methods.
To further investigate the effectiveness of the proposed Conv-Former, we demonstrate the scatter plots between the MOS and the prediction scores and analyze the correlation. Figure 12 and Figure 13 show the scatter plots of different IQA methods on the CSIQ and LIVEC datasets, respectively. The red points denote the testing instances. The blue line is the ideal linear relationship between MOS and the prediction score. The green line is the curve fitted using the test instance. All values are normalized in the range of −5 to 5 for a better view. In the Figure 12 and Figure 13, we can find that the results of Conv-Former are more in line with the MOS. Compared with WaDIQaM and BIECON, our Conv-Former has fewer outliers and performs more consistently with the MOS.
To evaluate the generalization of our proposed Conv-Former, we conduct the cross dataset evaluation on LIVE, CSIQ, TID2013, and LIVEC. We train the model on one train dataset separately, then test it on the full set of the other three benchmark datasets. As shown in Table 5, Conv-Former achieves good generalization ability.

4.5. Ablation Studies

In this section, to evaluate the efficiency of our proposed components, we analyze the effectiveness of the proposed network by conducting ablation studies. With different configuration and implementation strategies, we evaluate the effect of each of the four major components: Multi-Scale transformer architecture (MS), Adaptive Position Embedding (APE), Dual Path Pooling (DPP), and local information perception (LIP) module. We conduct ablation experiments on the LIVEC, KonIQ, CSIQ, and TID2013 databases. The results are shown in Table 6.
We first examine the effectiveness of our proposed Adaptive Position Embedding module. From the results, we can see a significant improvement in SRCC and PLCC with the use of positional coding. In KonIQ, the SRCC and PLCC increased by 2.8% and 2.2%, respectively. As a result, this demonstrates the critical importance of Position Embedding for our proposed Conv-Former. All subsequent ablation experiments were carried out with the Position Embedding. Table 6 shows that the best performance can be achieved when all three components are available. The table shows that the lack of any of the following three components in our Conv-Former will negatively impact the objective performance metrics.

4.6. Analysis and Discussion

To further validate the proposed method’s effectiveness and analyze the internal mechanism of the difference in the performance of the neural networks, in Figure 8, we compare the attention maps of the three different network architectures.
The proposed Conv-Former tends to focus on the regions that significantly impact the quality assessment scores, as seen from the attention maps so that the obtained image quality assessment scores can be remarkably consistent with the subjective assessment results of the human eye. For example, people tend to assess the quality of an image based on the target when it is in a pure background, and some images are blurred to highlight the target, so people tend to focus on the clear parts to assess the quality of the image. This aligns with our intuition, resulting in superior results compared to other methods. Compared to the traditional approach using CNN, the approach using the transformer tends to activate more significant regions rather than local areas, implying enhanced long-range feature dependency. Since the local information-aware module provides detailed local features, the Conv-former can retain important detailed local features that are often corrupted by the vision transformer. In addition, the target attention areas can be more complete in a significant area context, meaning that Conv-former learns feature representations with higher discriminative power.
On the other hand, humans perceive different image qualities in different ways when the image’s content is different. Our proposed Conv-Former can identify the image content well in the process of image quality evaluation and can try to understand the image before predicting it, which is more in line with the laws of human perception of the objective world.

5. Conclusions

In this paper, we propose a novel network, combining convolution and Self-Attention for an image quality assessment network (Conv-Former) for the no-reference image quality assessment task. Our model can obtain global features through transformer and local information perception (LIP). We evaluate the effectiveness of Conv-Former on five benchmark datasets, we found that our method achieves the best PLCC/SRCC results compared to other works. Especially on the more complex dataset TID2013, our proposed model achieved a solid improvement over previous work. Even though Conv-Former achieves 0.965 on PLCC and 0.964 on SRCC, which means the metric is consistent with the human perspective. Experimental results showed that our proposed approach outperforms the state-of-the-art (SOTA) methods on IQA databases, which has strong generalization ability and provides prospects for the broader application of IQA tasks. Future work will focus on developing more generic IQA models in which a single model can be adapted to diverse image content and imaging devices. Our proposed model should have a sufficient dataset so that the trained model can have a stronger generalization ability and obtain excellent results on images where the model has not been trained. The size of the model is also a key factor in its ability to be deployed in practical applications, and we will continue to optimize the model in the future so that it has fewer parameters and runs faster.

Author Contributions

All authors were involved in the formulation of the problem and the design of the methodology; L.H. designed the experiment and wrote the manuscript; H.L. (Hailong Liu) and G.B. analysed the accuracy of the experimental data; Y.Z. (Yuchen Zhao), Z.Y., Y.F. and H.L. (Hengyi Lv) reviewed and guided the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China (62005269).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Acknowledgments

The authors thank the editors and reviewers for their hard work and valuable advice.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Le, Q.-T.; Ladret, P.; Nguyen, H.-T.; Caplier, A. Computational Analysis of Correlations between Image Aesthetic and Image Naturalness in the Relation with Image Quality. J. Imaging 2022, 8, 166. [Google Scholar] [CrossRef] [PubMed]
  2. Talebi, H.; Milanfar, P. NIMA: Neural Image Assessment. IEEE Trans. Image Process. 2018, 27, 3998–4011. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Han, L.T.; Zhao, Y.C.; Lv, H.Y.; Zhang, Y.S.; Liu, H.L.; Bi, G.L. Remote Sensing Image Denoising Based on Deep and Shallow Feature Fusion and Attention Mechanism. Remote Sens. 2022, 14, 23. [Google Scholar] [CrossRef]
  4. Zhang, W.X.; Ma, K.D.; Yan, J.; Deng, D.X.; Wang, Z. Blind Image Quality Assessment Using a Deep Bilinear Convolutional Neural Network. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 36–47. [Google Scholar] [CrossRef] [Green Version]
  5. Eskicioglu, A.M.; Fisher, P.S. Image quality measures and their performance. IEEE Trans. Commun. 1995, 43, 2959–2965. [Google Scholar] [CrossRef] [Green Version]
  6. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  7. Hui, Q.; Sheng, Y.X.; Yang, L.K.; Li, Q.M.; Chai, L. Reduced-Reference Image Quality Assessment for Single-Image Super-Resolution Based on Wavelet Domain. In Proceedings of the 31st Chinese Control and Decision Conference (CCDC), Nanchang, China, 3–5 June 2019; pp. 2067–2071. [Google Scholar]
  8. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-Reference Image Quality Assessment in the Spatial Domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef]
  9. Xu, J.T.; Ye, P.; Li, Q.H.; Du, H.Q.; Liu, Y.; Doermann, D. Blind Image Quality Assessment Based on High Order Statistics Aggregation. IEEE Trans. Image Process. 2016, 25, 4444–4457. [Google Scholar] [CrossRef]
  10. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a “Completely Blind” Image Quality Analyzer. IEEE Signal Process. Lett. 2013, 20, 209–212. [Google Scholar] [CrossRef]
  11. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  12. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 1–9. [Google Scholar] [CrossRef]
  13. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 779–788. [Google Scholar]
  14. Kim, J.; Lee, S. Fully Deep Blind Image Quality Predictor. IEEE J. Sel. Top. Signal Process. 2017, 11, 206–220. [Google Scholar] [CrossRef]
  15. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  16. Lan, Z.; Chen, M.; Goodman, S.; Gimpel, K.; Sharma, P.; Soricut, R. Albert: A lite bert for self-supervised learning of language representations. arXiv 2019, arXiv:1909.11942. [Google Scholar]
  17. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. 2020, 21, 1–67. [Google Scholar]
  18. Yang, Z.; Dai, Z.; Yang, Y.; Carbonell, J.; Salakhutdinov, R.R.; Le, Q.V. Xlnet: Generalized autoregressive pretraining for language understanding. Adv. Neural Inf. Process. Syst. 2019, 32, 1–11. [Google Scholar]
  19. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  20. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  21. He, K.M.; Zhang, X.Y.; Ren, S.Q.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  22. Guo, J.; Han, K.; Wu, H.; Tang, Y.; Chen, X.; Wang, Y.; Xu, C. Cmt: Convolutional neural networks meet vision transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 12175–12185. [Google Scholar]
  23. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  24. Graves, A.; Mohamed, A.-r.; Hinton, G. Speech recognition with deep recurrent neural networks. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 6645–6649. [Google Scholar]
  25. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  26. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  27. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  28. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3146–3154. [Google Scholar]
  29. Wang, S.; Ma, K.; Yeganeh, H.; Wang, Z.; Lin, W. A patch-structure representation method for quality assessment of contrast changed images. IEEE Signal Process. Lett. 2015, 22, 2387–2390. [Google Scholar] [CrossRef]
  30. Liu, Y.; Li, X. No-reference quality assessment for contrast-distorted images. IEEE Access 2020, 8, 84105–84115. [Google Scholar] [CrossRef]
  31. Gu, K.; Lin, W.; Zhai, G.; Yang, X.; Zhang, W.; Chen, C.W. No-reference quality metric of contrast-distorted images based on information maximization. IEEE Trans. Cybern. 2016, 47, 4559–4565. [Google Scholar] [CrossRef]
  32. Moorthy, A.K.; Bovik, A.C. A Two-Step Framework for Constructing Blind Image Quality Indices. IEEE Signal Process. Lett. 2010, 17, 513–516. [Google Scholar] [CrossRef]
  33. Moorthy, A.K.; Bovik, A.C. Blind Image Quality Assessment: From Natural Scene Statistics to Perceptual Quality. IEEE Trans. Image Process. 2011, 20, 3350–3364. [Google Scholar] [CrossRef] [PubMed]
  34. Saad, M.A.; Bovik, A.C.; Charrier, C. A DCT Statistics-Based Blind Image Quality Index. IEEE Signal Process. Lett. 2010, 17, 583–586. [Google Scholar] [CrossRef] [Green Version]
  35. Saad, M.A.; Bovik, A.C.; Charrier, C. Blind Image Quality Assessment: A Natural Scene Statistics Approach in the DCT Domain. IEEE Trans. Image Process. 2012, 21, 3339–3352. [Google Scholar] [CrossRef] [PubMed]
  36. Ye, P.; Kumar, J.; Kang, L.; Doermann, D. Unsupervised Feature Learning Framework for No-reference Image Quality Assessment. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012; pp. 1098–1105. [Google Scholar]
  37. Zhang, P.; Zhou, W.G.; Wu, L.; Li, H.Q. SOM: Semantic Obviousness Metric for Image Quality Assessment. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 2394–2402. [Google Scholar]
  38. Sheikh, H.R.; Sabir, M.F.; Bovik, A.C. A statistical evaluation of recent full reference image quality assessment algorithms. IEEE Trans. Image Process. 2006, 15, 3440–3451. [Google Scholar] [CrossRef]
  39. Ponomarenko, N.; Jin, L.; Ieremeiev, O.; Lukin, V.; Egiazarian, K.; Astola, J.; Vozel, B.; Chehdi, K.; Carli, M.; Battisti, F.; et al. Image database TID2013: Peculiarities, results and perspectives. Signal Process.-Image Commun. 2015, 30, 57–77. [Google Scholar] [CrossRef]
  40. Larson, E.C.; Chandler, D.M. Most apparent distortion: Full-reference image quality assessment and the role of strategy. J. Electron. Imaging 2010, 19, 21. [Google Scholar]
  41. Ghadiyaram, D.; Bovik, A.C. Massive online crowdsourced study of subjective and objective picture quality. IEEE Trans. Image Process. 2015, 25, 372–387. [Google Scholar] [CrossRef] [Green Version]
  42. Hosu, V.; Lin, H.; Sziranyi, T.; Saupe, D. KonIQ-10k: An ecologically valid database for deep learning of blind image quality assessment. IEEE Trans. Image Process. 2020, 29, 4041–4056. [Google Scholar] [CrossRef] [Green Version]
  43. Ponomarenko, N.; Lukin, V.; Zelensky, A.; Egiazarian, K.; Carli, M.; Battisti, F. TID2008—A database for evaluation of full-reference visual quality assessment metrics. Adv. Mod. Radioelectron. 2009, 10, 30–45. [Google Scholar]
  44. Thomee, B.; Shamma, D.A.; Friedland, G.; Elizalde, B.; Ni, K.; Poland, D.; Borth, D.; Li, L.-J. YFCC100M: The new data in multimedia research. Commun. ACM 2016, 59, 64–73. [Google Scholar] [CrossRef]
Figure 1. Three key architectures in transformers are depicted schematically. (a) Self-Attention; (b) Multi-head Self-Attention; (c) Transformer encoder.
Figure 1. Three key architectures in transformers are depicted schematically. (a) Self-Attention; (b) Multi-head Self-Attention; (c) Transformer encoder.
Sensors 23 00427 g001
Figure 2. Comparison of three different neural network architectures (a) ResNet-50 [21] (b) Vision transformer [19] (c) Multi-stage transformer [22].
Figure 2. Comparison of three different neural network architectures (a) ResNet-50 [21] (b) Vision transformer [19] (c) Multi-stage transformer [22].
Sensors 23 00427 g002
Figure 3. Representative traditional IQA methods and deep learning-based methods are compared with our proposed method on dataset TID2013. Higher correlation means better performance of the IQA method. (a) BRISQUE (b) NIQE (c) DBCNN (d) ours.
Figure 3. Representative traditional IQA methods and deep learning-based methods are compared with our proposed method on dataset TID2013. Higher correlation means better performance of the IQA method. (a) BRISQUE (b) NIQE (c) DBCNN (d) ours.
Sensors 23 00427 g003
Figure 4. The general process of traditional no-reference image quality assessment methods.
Figure 4. The general process of traditional no-reference image quality assessment methods.
Sensors 23 00427 g004
Figure 5. The overview of the network architecture of the proposed Conv-Former.
Figure 5. The overview of the network architecture of the proposed Conv-Former.
Sensors 23 00427 g005
Figure 6. Schematic drawing of the implementation of adaptive 2D position embedding.
Figure 6. Schematic drawing of the implementation of adaptive 2D position embedding.
Sensors 23 00427 g006
Figure 7. The architecture of the proposed Transformer Block.
Figure 7. The architecture of the proposed Transformer Block.
Sensors 23 00427 g007
Figure 8. Distorted images in the synthetic dataset CSIQ and the corresponding reference images, corresponding to the distortion types (a) White Noise (b) Gaussian Blur (c) Contrast stretching (d) Pink noise (e) JPEG Compression (f) JPEG2000 Compression.
Figure 8. Distorted images in the synthetic dataset CSIQ and the corresponding reference images, corresponding to the distortion types (a) White Noise (b) Gaussian Blur (c) Contrast stretching (d) Pink noise (e) JPEG Compression (f) JPEG2000 Compression.
Sensors 23 00427 g008
Figure 9. (af) are the distorted images from the authentic dataset LIVE Challenge.
Figure 9. (af) are the distorted images from the authentic dataset LIVE Challenge.
Sensors 23 00427 g009
Figure 10. Histogram representation of the data in Table 2.
Figure 10. Histogram representation of the data in Table 2.
Sensors 23 00427 g010
Figure 11. Scatter plots of ground-truth mean opinion scores (MOS) against predicted scores of (af) different IQA algorithms on CSIQ datasets. The red points denote the testing instances. The blue line is the ideal linear relationship between MOS and the prediction score. The green line is the curve fitted using the test instance. Higher correlation means better performance of the IQA method. (a) BRISQUE; (b) NIQE; (c) DIIVINE; (d) HOSA; (e) CORNIA; (f) BIECON; (g) SFA; (h) WaDIQaM; (i) ours.
Figure 11. Scatter plots of ground-truth mean opinion scores (MOS) against predicted scores of (af) different IQA algorithms on CSIQ datasets. The red points denote the testing instances. The blue line is the ideal linear relationship between MOS and the prediction score. The green line is the curve fitted using the test instance. Higher correlation means better performance of the IQA method. (a) BRISQUE; (b) NIQE; (c) DIIVINE; (d) HOSA; (e) CORNIA; (f) BIECON; (g) SFA; (h) WaDIQaM; (i) ours.
Sensors 23 00427 g011
Figure 12. Scatter plots of ground-truth mean opinion scores (MOS) against predicted scores of (af) different algorithms on LIVEC datasets. The red points denote the testing instances. The blue line is the ideal linear relationship between MOS and the prediction score. The green line is the curve fitted using the test instance. Higher correlation means better performance of the IQA method. (a) BRISQUE; (b) NIQE; (c) DIIVINE; (d) HOSA; (e) CORNIA; (f) BIECON; (g) SFA; (h) WaDIQaM; (i) ours.
Figure 12. Scatter plots of ground-truth mean opinion scores (MOS) against predicted scores of (af) different algorithms on LIVEC datasets. The red points denote the testing instances. The blue line is the ideal linear relationship between MOS and the prediction score. The green line is the curve fitted using the test instance. Higher correlation means better performance of the IQA method. (a) BRISQUE; (b) NIQE; (c) DIIVINE; (d) HOSA; (e) CORNIA; (f) BIECON; (g) SFA; (h) WaDIQaM; (i) ours.
Sensors 23 00427 g012aSensors 23 00427 g012b
Figure 13. Comparison of attention maps for three different architectures of neural networks. All attentions are normalized in the range 0 to 1. The red areas mean the higher attention values, and the blue areas mean the lower attention values. The visual sensitive areas of our approach are more distinguishable.
Figure 13. Comparison of attention maps for three different architectures of neural networks. All attentions are normalized in the range 0 to 1. The red areas mean the higher attention values, and the blue areas mean the lower attention values. The visual sensitive areas of our approach are more distinguishable.
Sensors 23 00427 g013aSensors 23 00427 g013b
Table 1. IQA datasets for performance evaluation and model training.
Table 1. IQA datasets for performance evaluation and model training.
DatabaseReference
Image
Distorted
Image
Distorted TypeScore TypeScore Range
LIVE297795DMOS *[0, 100]
TID201325300024MOS[0, 9]
CSIQ308666DMOS[0, 1]
LIVE Challenge-1162-MOS[0, 100]
KonIQ-10k-10,073-MOS[0, 5]
* DMOS indicates the difference in mean opinion scores (MOS) between the test image and its reference image.
Table 2. Performance comparison of the IQA methods on five standard IQA databases, i.e., LIVEC, KonIQ, TID2013, LIVE, CSIQ. The top-performing method is highlighted in boldface.
Table 2. Performance comparison of the IQA methods on five standard IQA databases, i.e., LIVEC, KonIQ, TID2013, LIVE, CSIQ. The top-performing method is highlighted in boldface.
MethodsLIVECKonIQTID2013LIVECSIQ
PLCCSRCCPLCCSRCCPLCCSRCCPLCCSRCCPLCCSRCC
BRISQUE0.6290.6080.6810.6650.6940.6040.9350.9390.8290.746
NIQE0.4150.3730.4890.5160.4260.3170.9080.9080.7250.627
DIIVINE0.5950.5760.4790.4340.6720.5830.8810.8790.8360.784
HOSA0.6780.6400.6940.6710.8150.7350.9470.9460.8230.741
WaDIQaM0.6800.6710.8050.7970.7870.7610.9630.9540.9730.955
BIECON0.6130.5950.6510.6180.7620.7170.9620.9610.8230.815
SFA0.8330.8120.8720.8560.8730.8610.8950.8830.8180.796
PQR0.8820.8570.8840.8800.7980.7390.9710.9650.9010.873
DBCNN0.8690.8510.8840.875--0.9710.9680.9590.946
SHN0.8820.8590.9170.906--0.9660.9620.9420.923
RankIQA- *---0.8100.7800.9820.981--
ResNet-ft0.8490.819--0.7560.7120.9540.9500.9050.876
TRIQ--0.9220.910------
MUSIQ--0.9280.916------
Ours0.8990.8680.9380.9240.9650.9640.9790.9710.9780.964
* indicates that no testing was undertaken on this dataset in the original.
Table 3. Average PLCC and SRCC results of individual distortion types on the LIVE databases. The top-performing method is highlighted in boldface.
Table 3. Average PLCC and SRCC results of individual distortion types on the LIVE databases. The top-performing method is highlighted in boldface.
MethodsJPEGJPEG200WNGBFF
PLCCBRISQUE0.9710.9400.9890.9650.894
HOSA0.9670.9490.9830.9670.967
CORNIA0.9620.9440.9740.9610.943
DBCNN0.9860.9670.9880.9560.961
Ours0.9870.9770.9860.9740.970
SRCCBRISQUE0.9650.9290.9820.9660.828
HOSA0.9540.9350.9750.9540.954
CORNIA0.9470.9240.9580.9510.921
DBCNN0.9720.9550.9800.9350.930
Ours0.9740.9600.9710.9650.965
Table 4. Average PLCC and SRCC results of individual distortion types on the CSIQ databases. The top performing method is highlighted in bold face.
Table 4. Average PLCC and SRCC results of individual distortion types on the CSIQ databases. The top performing method is highlighted in bold face.
MethodsJPEGJPEG200WNGBPNCC
PLCCBRISQUE0.8280.8870.7420.8910.4960.835
HOSA0.7590.8990.6560.9120.6010.744
CORNIA0.5630.8830.6870.9040.6320.543
DBCNN0.9820.9710.9560.9690.9500.895
MEON0.9790.9250.9580.946--
Ours0.9660.9870.9510.9760.9820.947
SRCCBRISQUE0.8060.8400.7230.8200.3780.804
HOSA0.7330.8180.6040.8410.5000.716
CORNIA0.5130.8310.6640.8360.4930.462
DBCNN0.9400.9530.9480.9470.9400.870
MEON0.9480.8980.9510.918--
Ours0.9690.9800.9750.9450.9650.924
Table 5. SRCC evaluations on cross database tests. The top-performing method is highlighted in boldface.
Table 5. SRCC evaluations on cross database tests. The top-performing method is highlighted in boldface.
TrainTestMethods
BRISQUEM3FRIQUEECORNIAHOSADB-CNNOurs
LIVECSIQ0.5620.6210.7220.6490.5940.7580.762
TID20130.3580.3440.4610.3600.3610.5240.563
LIVEC0.3370.2260.4110.4430.4630.5670.572
CSIQLIVE0.8470.7970.8790.8530.7730.8770.864
TID20130.4540.3280.4630.3120.3290.5400.572
LIVEC0.1310.1830.2640.3930.2910.4520.463
TID2013LIVE0.7900.8730.7550.8460.8460.8910.894
CSIQ0.5900.6050.6350.6720.6120.8070.853
LIVEC0.2540.1120.1810.2930.3190.4570.524
LIVECLIVE0.2380.0590.6440.5880.5370.7460.752
CSIQ0.2410.1090.5920.4460.3360.6970.711
TID20130.2800.0580.4240.4030.3990.4240.417
Table 6. PLCC and SRCC results of ablation experiments on the LIVEC, KonIQ, CSIQ, and TID2013 database.
Table 6. PLCC and SRCC results of ablation experiments on the LIVEC, KonIQ, CSIQ, and TID2013 database.
MethodsLIVECKonIQCSIQTID2013
APEMSLIPDPPPLCCSRCCPLCCSRCCPLCCSRCCPLCCSRCC
×0.8710.8560.9220.9130.9720.9530.9580.945
×0.8930.8660.9250.9160.9730.9600.9590.950
×0.8810.8600.9280.9210.9700.9610.9550.948
×0.8900.8620.9330.9230.9750.9630.9630.960
××0.8820.8580.9300.9210.9690.9560.9610.950
××0.8940.8590.9240.9140.9720.9610.9600.952
××0.8780.8540.9290.9180.9650.9530.9490.938
0.8990.8680.9380.9240.9780.9640.9650.964
“✓” means that Conv-Former contains this component, “×” means that Conv-Former does not contain this component.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, L.; Lv, H.; Zhao, Y.; Liu, H.; Bi, G.; Yin, Z.; Fang, Y. Conv-Former: A Novel Network Combining Convolution and Self-Attention for Image Quality Assessment. Sensors 2023, 23, 427. https://0-doi-org.brum.beds.ac.uk/10.3390/s23010427

AMA Style

Han L, Lv H, Zhao Y, Liu H, Bi G, Yin Z, Fang Y. Conv-Former: A Novel Network Combining Convolution and Self-Attention for Image Quality Assessment. Sensors. 2023; 23(1):427. https://0-doi-org.brum.beds.ac.uk/10.3390/s23010427

Chicago/Turabian Style

Han, Lintao, Hengyi Lv, Yuchen Zhao, Hailong Liu, Guoling Bi, Zhiyong Yin, and Yuqiang Fang. 2023. "Conv-Former: A Novel Network Combining Convolution and Self-Attention for Image Quality Assessment" Sensors 23, no. 1: 427. https://0-doi-org.brum.beds.ac.uk/10.3390/s23010427

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop