Next Article in Journal
Optical Convolutional Neural Networks: Methodology and Advances (Invited)
Previous Article in Journal
Modeling Human Encounter Situation Awareness Results Using Support Vector Machine Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adaptive Feature Fusion and Kernel-Based Regression Modeling to Improve Blind Image Quality Assessment

Electronics and Telecommunications Research Institute (ETRI), Gwangju 61012, Republic of Korea
Appl. Sci. 2023, 13(13), 7522; https://doi.org/10.3390/app13137522
Submission received: 23 May 2023 / Revised: 2 June 2023 / Accepted: 15 June 2023 / Published: 26 June 2023

Abstract

:
In the fields of image processing and computer vision, evaluating blind image quality (BIQA) is still a difficult task. In this paper, a unique BIQA framework is presented that integrates feature extraction, feature selection, and regression using a support vector machine (SVM). Various image characteristics are included in the framework, such as wavelet transform, prewitt and gaussian, log and gaussian, and prewitt, sobel, and gaussian. An SVM regression model is trained using these features to predict the quality ratings of photographs. The proposed model uses the Information Gain attribute approach for feature selection to improve the performance of the regression model and decrease the size of the feature space. Three commonly used benchmark datasets, TID2013, CSIQ, and LIVE, are utilized to assess the performance of the proposed methodology. The study examines how various feature types and feature selection strategies affect the functionality of the framework through thorough experiments. The experimental findings demonstrate that our suggested framework reaches the highest levels of accuracy and robustness. This suggests that it has a lot of potential to improve the accuracy and dependability of BIQA approaches. Additionally, its use is broadened to include image transmission, compression, and restoration. Overall, the results demonstrate our framework’s promise and ability to advance studies into image quality assessment.

1. Introduction

Images have become a vital channel for conveying information due to the rapid volant of digital technology. This led to an immersive user experience and the exponential stacking of digital images in large data volumes. A decline in the quality of digital images takes place during the cyclic period of acquiring the images, performing image compression, processing, and transmitting the images to the data storage point. This inevitable fact consequently governs the immersive user experience, leading to abysmal image quality, thus making a space for an image quality assessment IQA approach to be filled in [1]. On the other side of the spectrum, being able to robustly assess the image quality deeply affects various computer vision applications, which broadly include bio-medical imaging, self-driving vehicles, and object detection, to name a few [2,3].
Among the two quality assessment approaches, subjective and objective IQA, the latter is widely utilized, even though subjective IQA is a definitive and nearly error-free approach. The induction of human evaluation in the latter approach makes it a time extensive and tedious task, thus giving a selective priority to objective IQA, which automatically deduces the image quality by deploying computational methods capable of speculating the image quality in a manner that is harmonious with the image quality scores of human subjects [4]. Objective IQA is sub-classed into full reference IQA, reduce-reference IQA, and no-reference IQA, widely labeled as Blind IQA. The availability of distortion-free versions of images, termed as pristine images, classifies the three types. The reference or pristine image is unavailable during the process of BIQA, making it a noteworthy assessment approach among the three objective IQA methods [4,5].
Moreover, BIQA is further classified into distortion-specific and general-purpose image quality assessment approaches. The latter approach is designed keeping in mind the fact that multiple distortions can occur in a digital image, making it superior to the distortion-specific approach. As the general-purpose approach is not limited to a specific type of image distortion, its ability to assess image quality highly relies on obtaining the features that capture the essential information of multiple image distortions [6]. Conventional BIQA methods utilizing hand-crafted features are based on platforms such as the human visual system HVS and natural scene statistics NSS. However, due to the higher extent of similarity among BIQA and human perception, compiling such hand-crafted features thath can effectively capture and represent the image quality degradation levels becomes a tedious task [7].
Improper lighting, constraints put on imaging lenses, and limitations of sensors can induce distortions that are more complex in nature, and such cases in digital images usually occur in reality-based scenarios. The traditional quality assessment methods usually diverged towards artificially distorted images with distortions such as JPEG, gaussian noise, and blue, to name a few [8]. However, the bottleneck in these datasets is that they contain a limited quantity of images, which are less diverse in terms of distortions. Furthermore, as multiple distortions may exist in an image in reality-based cases, such artificially distorted images do not emulate the complex nature of distortions that can occur in real world images [9].
Traditional BIQA models extract low-level features from distorted images and predict the image quality by applying machine learning regression models [10,11]. This way of applying NSS to represent image quality does not capture the local distortions in an image. A recent utilization of deep learning methods in IQA research prompted the use of convolutional neural networks CNN to extract image features in order to highlight degradation regions in an image. One major issue arises that deep learning based IQA methods are usually trained on smaller and artificially induced image datasets [12,13]. Moreover, CNNs are also deployed to extract features of distorted images to replace the conventional approach of hand-crafted features for IQA research [1]. CNN-based IQA methods should be capable of capturing two quantities, such as distortion and image visual content. This mainly arises from the fact that most CNN-based IQA methods are oriented to image-visual content, which emulates the process of image recognition [14]. The designed IQA method should be capable of handling generalization as various distortions can coexist in an image from a set of vast image distortions. Along with it, deep IQA methods required a large quantity of images to train a neural network, usually comprising more than one million images [15].
In recent times, deep neural networks have shown great improvement in different research fields such as bioinformatics [16,17,18], agriculture [19,20], and medical imaging [21,22]. The deep IQA methods utilize convolutional operations by deploying deep neural networks to extract features from distorted regions of the image [1]. These IQA methods cannot yield accurate predictions on reality-based or authentic datasets, as such methods cannot learn the patterns and feature representations that occur in these distorted images. The distortions as induced in artificial datasets range from Gaussian blur, Gaussian noise, JPEG compression, etc. [15]. While distortions in the authentic datasets are generally caused by naturally occurring conditions. This can be evidently observed by examining the gradient maps of both types of datasets. The high-frequency components evident in the gradient map are object-oriented in the case of a naturally distorted image, while such components are distributed across the whole image for artificially distorted images [23].
A no-reference IQA approach that can combine both types of distorted images, is an effective solution because it is more versatile in nature. Multiple such IQA approaches countenance this issue, but still, these methods have various issues, i.e., in order to predict the image quality, high-level semantic features are utilized. Furthermore, such approaches try to combine the effects of both types of distorted images by using a bilinear pooling operation, which is inadequate for capturing the relationship between these two types of distorted images [24].

2. Datasets

The TID2013 dataset comprises 25 reference photos and 3000 distorted images created with 24 various forms of distortion, such as JPEG compression, Gaussian blur, and white noise. The collection also contains subjective quality rankings for each warped image, which were gathered from human observers during a subjective experiment. The subjective quality scores range from 0 to 9, with 0 being the poorest and 9 representing the best.
The CSIQ dataset contains 30 reference photos and 866 distorted images created with six distinct forms of distortion, including JPEG2000 compression, JPEG compression, and Gaussian blur. The subjective quality scores for each warped image were also collected from human observers in a subjective experiment, and they range from 1 to 5, with 1 being the poorest quality and 5 being the greatest quality.
The LIVE dataset is more current, with 29 reference photos and 779 distorted images made using 5 different forms of distortion, including JPEG compression, Gaussian blur, and white noise. The subjective quality scores for each deformed image were determined by human observers in a subjective experiment, and they range from 0 to 100, with 0 being the lowest quality and 100 representing the greatest quality.
While these datasets have been frequently used to test the effectiveness of blind image quality evaluation algorithms in the literature, they do have certain drawbacks. For example, the distortion types contained in these datasets may not completely reflect the spectrum of real-world image distortions. Nonetheless, these datasets serve as a valuable standard for comparing the performance of various approaches in the field of blind image quality evaluation. Figure 1 represents the sample images from the datasets.

3. Proposed Methodology

Figure 2 represents the overall framework of the Blind Image Quality Assessment system used in this research paper. The framework consists of three main stages: Feature Extraction, Feature Selection, and Quality Prediction.
In the Feature Extraction stage, the reference and distorted images are preprocessed using several well-known image processing techniques to extract relevant features. The extracted features include Haar wavelet coefficients, Prewitt and Sobel edge detection, and Log filters.
In the Feature Selection stage, the extracted features are passed through the WEKA machine learning tool to select the most relevant features using the Information Gain attribute algorithm aiming to reduce the in dimensionality of the feature map and improve the efficiency of the proposed model. The selected characteristics are supplied into the Support Vector Machine (SVM) Regression model in the Quality Prediction step to forecast the quality of the distorted images.

3.1. Feature Extraction

Feature extraction is an important step in Blind Image Quality Assessment because it allows vital details to be extracted from deformed images. Numerous well-known algorithms for image processing were utilized in this study to extract characteristics from source and damaged images. The images were processed using the following filters:

3.1.1. Wavelet Transform

The Wavelet Transform is a common image processing approach for extracting features. It divides the image into frequency sub-bands that capture varying levels of detail and texture. The Haar wavelet transform was utilized in this study to extract characteristics from reference and distorted images. This is how the Haar wavelet transform is defined,
W l , m = 1 2 l q = 0 2 l 1 h q · x p q where l , p = 0 , 1 , 2 ,
where W l , p denotes the wavelet coefficient at scale l and position m, h q denotes the Haar wavelet filter, and x p denotes the input image pixel value.

3.1.2. Prewitt and Gaussian

The Prewitt filter is an edge recognition filter that improves the image’s edges. Prewitt filter was used for the pristine and damaged images in this study, followed by a Gaussian filter to minimize noise. The Prewitt filter can be defined by following equation,
G p = 1 0 1 1 0 1 1 0 1 I and G q = 1 1 1 0 0 0 1 1 1 I
where G p and G q are the Prewitt filters for identifying vertical and horizontal edges, respectively, and I is the input image. The Gaussian filter is defined as follows,
G ( p , q ) = 1 2 π σ 2 exp p 2 + q 2 2 σ 2
where G ( p , q ) denotes the Gaussian filter at location ( p , q ) , and σ denotes the standard deviation of the filter.

3.1.3. Log and Gaussian

The log filter is used to identify edges and minimize image noise. In this study, the reference and distorted images were subjected to a Log filter, which was then followed by a Gaussian filter to minimize noise. The Log filter may be defined in the following way,
L ( p , q ) = 1 π σ 4 1 p 2 + q 2 2 σ 2 exp p 2 + q 2 2 σ 2
where L ( p , q ) denotes the Log filter at location ( p , q ) , and σ denotes the standard deviation of the filter.

3.1.4. Prewitt, Sobel, and Gaussian

Another edge detection filter that improves image edges is the Sobel filter. Prewitt and Sobel filters were applied to the reference and distorted images in this study, followed by a Gaussian filter to remove noise.

3.2. Feature Selection

The suggested Blind Image Quality Assessment system relies on the choice of features to minimize the complexity of the feature vector and increase the machine’s performance. In this study, the WEKA machine learning tool is utilized to pick features.
The Information Gain attribute feature selector ranks attributes based on their capacity to convey details regarding the overall quality of the distorted images. By employing a feature as a split property, the evaluator calculates the reduction in entropy of the quality scores. The Information Gain score IG ( F , Q ) is derived using the following equation for feature set F and quality scores Q:
IG ( F , Q ) = H ( Q ) f F | { q Q : q f = f } | | Q | H ( { q Q : q f = f } )
In this equation, H ( Q ) denotes the entropy of the quality scores, f denotes an attribute, and q f denotes the quality rating of the image determined by that selected feature. The total number of images in Q for which the feature f is the split attribute is given by | q Q : q f = f | . The entropy of the quality ratings for images when the feature f is the divided attribute is H ( q Q : q f = f ) .
The Information Gain metric represents the quantity of information obtained regarding image quality by incorporating a feature into the model. The greater the Information Gain score, the more essential the attribute is for quality forecasting, thus increasing its position in the process of choosing the feature.

4. Quality Prediction Stage

In the final assessment stage, the selected characteristics are used to construct a regression model that is capable of predicting the score for the quality of the distorted images. Due to its capacity to handle feature vectors with high dimensionality and nonlinear correlations between features and quality ratings, Support Vector Machine (SVM) regression is employed as the prediction model in this research work. We extract numerous image characteristics from the proposed system, such as wavelet coefficients, edge detection findings, and filter responses. These characteristics contribute to a multidimensional feature space. The capacity of SVM to handle such complicated and high-dimensional data makes it ideal for our needs. The strength of SVM is its capacity to deal with non-linear connections between characteristics and target variables. The connection between image attributes and quality scores in image quality evaluation can be non-linear and complicated. The kernel-based trick of SVM allows it to implicitly translate characteristics to a higher-dimensional space, allowing for the successful modeling of non-linear connections. Moreover, SVM offers high generalization characteristics and is less susceptible to overfitting, both of which are important when constructing a robust and trustworthy BIQA model.
The SVM regression model establishes the relationship between the set of features F and the quality ratings Q. A set of training data D of N labeled image data is used to train the function of regression f ( p ) which is defined as,
f ( p ) = l = 1 M α l y l k ( p , p l ) + w
In this formula, p is an image sample expressed as a feature vector, and α l and y l are the initial samples’ Lagrange multipliers and labels for classes, respectively. The kernel function k ( p , p l ) shifts the feature vector p to a higher-dimensional feature space, allowing for nonlinear decision boundaries. The radial basis function (RBF) kernel function is utilized in this research paper,
k ( p , p l ) = exp ( γ | | p p l | | 2 )
where γ denotes the kernel parameter.
The below-mentioned optimization equation is used to learn the regression coefficients α l and bias term w:
min α , w 1 2 l = 1 M j = 1 M α l α j y l y j k ( p l , p j ) l = 1 M α i s . t . 0 α l C l = 1 M α l y l = 0
C is a hyperparameter in this optimization problem that determines the tradeoff between maximizing the margin and reducing the training error. The optimization issue is tackled iteratively using Sequential Minimal Optimization (SMO). Once trained on the training set, the SVM regression model can be used to predict the quality scores of fresh, unseen image samples using their associated feature vectors and the learnt regression function f ( p ) .

5. Evaluation Metrics

Several measures are used to compare projected quality scores with ground-truth quality scores to evaluate the performance of BIQA models. These measures are used to assess the relationship between projected and actual quality ratings. To evaluate the proposed framework, we have used four of the most commonly used evaluation metrics, which are discussed below.

5.1. Spearman Rank-Order Correlation Coefficient (SROCC)

The monotonic connection between two ranking variables is measured by SROCC. It is used in BIQA to assess the relationship between expected and actual quality ratings. The equation of SROCC is defined as,
ρ = 1 6 l = 1 m ( r l q l ) 2 m ( m 2 1 ) ,
where m is the total number of images in the dataset, r l is the projected quality rating rank of the lth image, and q l is the ground truth rating rank of the lth image. SROCC has a value between 1 and 1, with 1 indicating an ideal positive association, 0 indicating no correlation, and 1 indicating a perfect negative association.

5.1.1. Linear Correlation Coefficient (LCC)

The linear connection between two variables is measured by LCC. It is used in BIQA to assess the relationship between expected and actual quality ratings. The following is a definition of LCC,
L C C = l = 1 m ( p i p ¯ ) ( q l q ¯ ) l = 1 m ( p l p ¯ ) 2 l = 1 m ( q l q ¯ ) 2 ,
where p l and q l are the forecasted and true quality ratings of the lth image, and p ¯ and q ¯ are the average of the predicted and ground truth quality scores. LCC has a value between 1 and 1, with 1 indicating perfect positive correlation, 0 indicating no correlation, and 1 indicating a perfect negative correlation.

5.1.2. Kendall Rank-Order Correlation Coefficient (KROCC)

KROCC assesses the ordinal relationship between two variables. It is used in BIQA to assess the relationship between expected and actual quality ratings. The equation of KROCC is defined as,
τ = ( number of concordant pairs ) ( number of discordant pairs ) m 2 ,
where m is the total number of images within the database, and a pair of images ( p l , q l ) is considered to be concordant if their rankings in their predicted and actual scores are the same, and otherwise discordant. KROCC has a value between 1 and 1, with 1 indicating perfect positive correlation, 0 indicating no correlation, and 1 indicating a perfect negative correlation.

5.1.3. Root Mean Squared Error (RMSE)

The RMSE gauges the difference between the expected and actual quality ratings. The RMSE is defined as,
R M S E = 1 M l = 1 M ( q p r e d , l q t r u e , l ) 2
M is the total number of test images, q p r e d , l is the projected quality rating for image l, and q t r u e , l is the actual quality rating for image l. The smaller the RMSE, the greater the model’s performance.

6. Results and Discussion

In this section, comparative analysis is carried out between different features and between existing and proposed methods.

6.1. Performance Comparison of Different Features

The performance of the proposed Blind Image Quality Assessment (BIQA) methodology on three frequently used benchmark datasets is shown in Table 1. The table displays the results produced by employing various image attributes in the suggested technique. The proposed model employed four types of image features in this study: Haar wavelet coefficients, Prewitt and Sobel edge detection, Log filters, and Prewitt, Sobel, and Gaussian filter combinations.
Table 1 also includes the results of using the Information Gain feature selection algorithm with the WEKA tool on the Prewitt and Gaussian (WEKA PG) and Prewitt, Sobel, and Gaussian (WEKA PSG) features, which achieved a performance improvement over the non-selected features.

6.2. Feature Analysis

Figure 3 depicts the Prewitt and Gaussian (PG) filter feature vectors for four distinct forms of distortion from CSIQ datasets: JPEG compression, white noise, Gaussian blur, and JPEG 2000 compression. Each subfigure depicts the PG feature vector distribution for a distinct type of distortion.
As seen in the image, each form of distortion has a distinct feature vector pattern. In the first subfigure, for example, the feature vector of JPEG compression shows a different pattern when compared with the other forms of distortion. This shows that the suggested technique is capable of collecting and learning the unique characteristics of each form of distortion.
Furthermore, the results show that by extracting and selecting relevant features with the Information Gain Score algorithm and then training the SVM regression model, the proposed framework can accurately predict the quality of distorted images. On widely used benchmark datasets such as TID2013, CSIQ, and LIVE, approach outperforms state-of-the-art methods in terms of accuracy and robustness.

6.3. Comparison with Exisiting Techniques

Table 2 compares the performance of proposed Blind Image Quality Assessment (BIQA) model to that of various leading models, including FRIQUEE, NFERM, BRSIQUE, BLINDS-II, CORNIA, and DIVINE, in terms of LCC, SROCC, KROCC, and RMSE on all three benchmark datasets (CSIQ, LIVE, and TID2013). On all three datasets, proposed model outperformed all other models in terms of all evaluation metrics. On the TID2013 dataset, for example, the proposed model achieved an LCC of 0.9453, an SROCC of 0.9289, a KROCC of 0.7804, and an RMSE of 0.406, all of which are significantly better than the best results reported by the other models. Similarly, using the LIVE and CSIQ datasets, the suggested model consistently outperformed the other models in terms of all assessment measures.
Several factors contribute to the superior performance of the proposed BIQA model. For starters, combining multiple image features such as wavelet transform, Prewitt and Sobel edge detection, and Log filters, as well as selecting them using the Information Gain attribute evaluator, allows for the capture of a wide range of relevant image attributes that are important for quality assessment. This enables the algorithm to successfully discriminate between different forms of image distortion and reliably estimate their quality.
Second, using SVM regression as the underlying machine learning algorithm has several advantages, including the ability to handle high-dimensional feature spaces, non-linear relationships between features and quality scores, and the ability to generalize well to new, previously unseen data.
Furthermore, extensive experimental evaluation and analysis of various types of features and feature selection techniques aid in the identification of the most effective combination of features for BIQA. Overall, the proposed BIQA model provides a promising method for reliably and robustly assessing the quality of distorted images, with potential applications in image reduction, transmission, and restoration.
It is vital to highlight that the use of machine learning techniques, such as the SVM Regression model, provides computational efficiency by definition. ML algorithms are renowned for their capacity to effectively analyze data and generate predictions based on previously learnt patterns. Furthermore, the feature selection phase in the proposed system reduces the complexity of the feature space, enhancing computing efficiency even further.
Certain restrictions and possible downsides must be considered in suggested approach. To begin, the framework’s performance is strongly dependent on the feature extraction techniques used, such as wavelet transformation and edge detection. While these strategies have yielded promising results, their efficacy may vary based on the precise properties of the photos under consideration. Furthermore, the framework’s generalizability to various image types should be examined. Although the system has been evaluated on benchmark datasets, its performance may vary when applied to other image categories or domains.

7. Conclusions

This study presented a system for Blind Image Quality Assessment (BIQA) that integrates feature extraction, feature selection, and support vector machine (SVM) regression. The incorporation of four separate types of image features, as well as the usage of the Information Gain attribute approach for feature selection, contribute to the regression model’s increased performance. Using well-known benchmark datasets such as TID2013, CSIQ, and LIVE, this shows that the proposed framework outperforms previous techniques in terms of accuracy and durability. Furthermore, this study performed extensive tests to investigate the effect of various types of features and feature selection methodologies on the framework’s performance. The system improves the resilience and accuracy of the quality prediction model by including a wide set of image properties. Furthermore, using the Information Gain approach for feature selection minimizes the complexity of the feature space and improves the regression model’s performance. As a whole, the proposed methodology outperforms the competition across many benchmark datasets, the proving its efficacy and robustness in blind image quality evaluation. These findings illustrate the major benefits of the proposed methodology and help to develop the discipline. It is crucial to highlight, however, that the suggested approach has limitations, such as the dependency on certain datasets and the need for future feature extraction technique development. A future study should solve these constraints and investigate the framework’s use in additional image-processing tasks.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. ur Rehman, M.; Nizami, I.F.; Majid, M. DeepRPN-BIQA: Deep architectures with region proposal network for natural-scene and screen-content blind image quality assessment. Displays 2022, 71, 102101. [Google Scholar] [CrossRef]
  2. Qi, K.; Li, H.; Rong, C.; Gong, Y.; Li, C.; Zheng, H.; Wang, S. Blind Image Quality Assessment for MRI with A Deep Three-dimensional content-adaptive Hyper-Network. arXiv 2021, arXiv:2107.06888. [Google Scholar]
  3. Li, G.; Yang, Y.; Qu, X.; Cao, D.; Li, K. A deep learning based image enhancement approach for autonomous driving at night. Knowl.-Based Syst. 2021, 213, 106617. [Google Scholar] [CrossRef]
  4. Rajevenceltha, J.; Gaidhane, V.H. An efficient approach for no-reference image quality assessment based on statistical texture and structural features. Eng. Sci. Technol. Int. J. 2022, 30, 101039. [Google Scholar] [CrossRef]
  5. Li, Q.; Lin, W.; Gu, K.; Zhang, Y.; Fang, Y. Blind image quality assessment based on joint log-contrast statistics. Neurocomputing 2019, 331, 189–198. [Google Scholar] [CrossRef]
  6. Xu, L.; Jiang, X. Blind image quality assessment for anchor-assisted adaptation to practical situations. Multimed. Tools Appl. 2023, 83, 17929–17946. [Google Scholar] [CrossRef]
  7. Zhai, G.; Min, X. Perceptual image quality assessment: A survey. Sci. China Inf. Sci. 2020, 63, 211301. [Google Scholar] [CrossRef]
  8. Nizami, I.F.; Rehman, M.U.; Waqar, A.; Majid, M. Impact of visual saliency on multi-distorted blind image quality assessment using deep neural architecture. Multimed. Tools Appl. 2022, 81, 25283–25300. [Google Scholar] [CrossRef]
  9. Hosu, V.; Lin, H.; Sziranyi, T.; Saupe, D. KonIQ-10k: An ecologically valid database for deep learning of blind image quality assessment. IEEE Trans. Image Process. 2020, 29, 4041–4056. [Google Scholar] [CrossRef] [Green Version]
  10. Nizami, I.F.; Majid, M.; Rehman, M.u.; Anwar, S.M.; Nasim, A.; Khurshid, K. No-reference image quality assessment using bag-of-features with feature selection. Multimed. Tools Appl. 2020, 79, 7811–7836. [Google Scholar] [CrossRef]
  11. Nizami, I.F.; Rehman, M.u.; Majid, M.; Anwar, S.M. Natural scene statistics model independent no-reference image quality assessment using patch based discrete cosine transform. Multimed. Tools Appl. 2020, 79, 26285–26304. [Google Scholar] [CrossRef]
  12. Ribeiro, R.; Trifan, A.; Neves, A.J. Blind Image Quality Assessment with Deep Learning: A Replicability Study and Its Reproducibility in Lifelogging. Appl. Sci. 2023, 13, 59. [Google Scholar] [CrossRef]
  13. Fateh, A.; Fateh, M.; Abolghasemi, V. Multilingual handwritten numeral recognition using a robust deep network joint with transfer learning. Inf. Sci. 2021, 581, 479–494. [Google Scholar] [CrossRef]
  14. Yang, G.; Wang, Y. Deep Superpixel-Based Network For Blind Image Quality Assessment. arXiv 2021, arXiv:2110.06564. [Google Scholar]
  15. Wu, J.; Ma, J.; Liang, F.; Dong, W.; Shi, G.; Lin, W. End-to-end blind image quality prediction with cascaded deep neural network. IEEE Trans. Image Process. 2020, 29, 7414–7426. [Google Scholar] [CrossRef]
  16. Rehman, M.U.; Tayara, H.; Zou, Q.; Chong, K.T. i6mA-Caps: A CapsuleNet-based framework for identifying DNA N6-methyladenine sites. Bioinformatics 2022, 38, 3885–3891. [Google Scholar] [CrossRef]
  17. Rehman, M.U.; Tayara, H.; Chong, K.T. DL-m6A: Identification of N6-methyladenosine Sites in Mammals using deep learning based on different encoding schemes. IEEE/ACM Trans. Comput. Biol. Bioinform. 2022, 20, 904–911. [Google Scholar] [CrossRef]
  18. Rehman, M.U.; Tayara, H.; Chong, K.T. DCNN-4mC: Densely connected neural network based N4-methylcytosine site prediction in multiple species. Comput. Struct. Biotechnol. J. 2021, 19, 6009–6019. [Google Scholar] [CrossRef]
  19. Rakhmatulin, I.; Kamilaris, A.; Andreasen, C. Deep neural networks to detect weeds from crops in agricultural environments in real-time: A review. Remote Sens. 2021, 13, 4486. [Google Scholar] [CrossRef]
  20. Espejo-Garcia, B.; Mylonas, N.; Athanasakos, L.; Fountas, S. Improving weeds identification with a repository of agricultural pre-trained deep neural networks. Comput. Electron. Agric. 2020, 175, 105593. [Google Scholar] [CrossRef]
  21. Rehman, M.U.; Ryu, J.; Nizami, I.F.; Chong, K.T. RAAGR2-Net: A brain tumor segmentation network using parallel processing of multiple spatial frames. Comput. Biol. Med. 2023, 152, 106426. [Google Scholar] [CrossRef]
  22. Rehman, M.U.; Akhtar, S.; Zakwan, M.; Mahmood, M.H. Novel architecture with selected feature vector for effective classification of mitotic and non-mitotic cells in breast cancer histology images. Biomed. Signal Process. Control 2022, 71, 103212. [Google Scholar] [CrossRef]
  23. Chetouani, A.; Quach, M.; Valenzise, G.; Dufaux, F. Combination of Deep Learning-based and Handcrafted Features for Blind Image Quality Assessment. In Proceedings of the 2021 9th European Workshop on Visual Information Processing (EUVIP), Paris, France, 23–25 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–6. [Google Scholar]
  24. Pan, Z.; Zhang, H.; Lei, J.; Fang, Y.; Shao, X.; Ling, N.; Kwong, S. Dacnn: Blind image quality assessment via a distortion-aware convolutional neural network. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 7518–7531. [Google Scholar] [CrossRef]
  25. Ghadiyaram, D.; Bovik, A.C. Perceptual quality prediction on authentically distorted images using a bag of features approach. J. Vis. 2017, 17, 32. [Google Scholar] [CrossRef] [Green Version]
  26. Gu, K.; Zhai, G.; Yang, X.; Zhang, W. Using free energy principle for blind image quality assessment. IEEE Trans. Multimed. 2014, 17, 50–63. [Google Scholar] [CrossRef]
  27. Mittal, A.; Moorthy, A.K.; Bovik, A.C. No-reference image quality assessment in the spatial domain. IEEE Trans. Image Process. 2012, 21, 4695–4708. [Google Scholar] [CrossRef]
  28. Saad, M.A.; Bovik, A.C.; Charrier, C. Blind image quality assessment: A natural scene statistics approach in the DCT domain. IEEE Trans. Image Process. 2012, 21, 3339–3352. [Google Scholar] [CrossRef] [PubMed]
  29. Ye, P.; Kumar, J.; Kang, L.; Doermann, D. Unsupervised feature learning framework for no-reference image quality assessment. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 1098–1105. [Google Scholar]
  30. Moorthy, A.K.; Bovik, A.C. Blind image quality assessment: From natural scene statistics to perceptual quality. IEEE Trans. Image Process. 2011, 20, 3350–3364. [Google Scholar] [CrossRef]
Figure 1. Sample images from the dataset.
Figure 1. Sample images from the dataset.
Applsci 13 07522 g001
Figure 2. Overview of the proposed framework for Blind Image Quality Assessment. The framework includes three main stages: Feature Extraction, Feature Selection, and Quality Prediction.
Figure 2. Overview of the proposed framework for Blind Image Quality Assessment. The framework includes three main stages: Feature Extraction, Feature Selection, and Quality Prediction.
Applsci 13 07522 g002
Figure 3. Visual word occurences ofr different feature vector.
Figure 3. Visual word occurences ofr different feature vector.
Applsci 13 07522 g003
Table 1. Performance Comparision on CSIQ, LIVE, TID2013 datasets. WT: Wavelet Transform, PG: Prewitt and Gaussian, LG: Log and Gaussian, PSG: Prewitt, Sobel and Gaussian.
Table 1. Performance Comparision on CSIQ, LIVE, TID2013 datasets. WT: Wavelet Transform, PG: Prewitt and Gaussian, LG: Log and Gaussian, PSG: Prewitt, Sobel and Gaussian.
MetricsTID2013CSIQLIVE
SSRSCNHFNINIDGBLURJPEGAwgnJPEGJp2kFnoiseJp2kJPEGWnGblurFF
WTsrocc0.90540.95530.93200.90690.90140.91810.93080.93650.90120.92170.89280.92450.92040.97240.95860.9216
lcc0.91110.96540.96540.90780.92010.92190.97790.95170.93480.93560.90110.93760.95020.98190.96340.9253
krocc0.74670.83330.79030.74460.73230.76670.79330.79310.75370.76550.71490.77270.77610.89660.84240.7492
rmse0.81770.20380.27920.26300.66340.49720.33230.05470.11170.11400.10065.7695.11512.97124.34136.3678
PGsrocc0.91690.92920.93630.91200.92010.93310.91590.94290.89870.92530.89210.92400.91670.97730.96550.9231
lcc0.93460.93450.97100.91310.94400.94040.96100.95390.93100.92950.89590.93400.94300.98890.96980.9267
krocc0.75330.78670.80010.76000.77130.79600.78030.76360.74480.75710.76110.76520.76480.89160.86210.7635
rmse0.72100.28650.25800.24980.54650.44600.39360.05480.10530.11970.10516.01875.46052.50903.97176.1869
LGsrocc0.93110.91300.93330.90210.90320.90420.91230.93690.88910.91500.90340.91740.91240.98420.95420.9097
lcc0.94820.91720.97090.90420.92600.90820.96640.94450.93300.92780.90040.92430.93580.99200.96020.9405
krocc0.78670.76560.80010.75790.74010.73790.76000.78330.72910.758860.72870.75760.75590.90640.83250.7635
rmse0.60490.31180.25290.25090.56890.54580.39490.05740.11300.11950.10126.40275.88122.12914.54955.7394
PSGsrocc0.91770.92080.93230.92380.90780.93850.90460.94230.89400.90300.89890.91740.91220.98230.96750.9167
lcc0.93260.92930.96810.92960.94030.94910.94440.95500.91870.92110.89790.92980.94010.99020.97040.9297
krocc0.76670.77800.79330.78000.75230.80000.75670.81280.73890.75860.75170.75380.75460.89160.84730.7538
rmse0.67850.28780.25970.22490.56160.41070.44600.04810.11860.11850.09676.25325.56262.37924.18715.7901
WEKA PGsrocc0.92000.93050.94160.91850.92650.93480.93010.94530.91350.93360.91270.92950.91960.97880.96880.9312
lcc0.92980.93480.97350.91950.94580.94220.97160.95490.94720.93550.92590.94070.94470.99010.97220.9366
krocc0.76670.79230.80470.76670.78000.79860.78670.80790.75620.79540.76210.77270.76370.89660.86740.7712
rmse0.70840.27600.23960.24180.53390.43600.36880.05150.10410.11450.10295.65055.37602.49603.87525.9869
WEKA PSGsrocc0.92220.92670.93830.92550.91040.93780.90850.94580.91160.91630.90120.92510.91480.98360.96780.9178
lcc0.93480.93230.96950.93230.94330.94500.95770.95710.93270.92860.90220.93600.94210.99110.97080.9311
krocc0.77330.78500.79750.78670.75830.79890.75780.81300.74110.76480.76210.77270.76340.90150.86210.7638
rmse0.68240.28380.24720.21760.55870.40970.44230.04780.11490.11430.09435.92525.45262.29053.88235.7641
Table 2. Performance Comparison of Proposed Method with Existing BIQA Techniques.
Table 2. Performance Comparison of Proposed Method with Existing BIQA Techniques.
BIQA ModelsLIVE DatasetTID2013 DatasetCSIQ Dataset
LCCSROCCKROCCRMSELCCSROCCKROCCRMSELCCSROCCKROCCRMSE
FRIQUEE [25]0.94110.93470.78179.20610.76880.69260.51610.79650.90690.88150.70770.1113
NFERM [26]0.94630.94270.80638.80210.74650.67470.49760.83010.86580.82130.63940.1298
BRISQUE [27]0.94820.94360.80058.66050.62130.57390.41490.96680.83110.74030.55900.1442
BLINDS II [28]0.93700.92980.77549.50720.65110.57230.41370.94030.81340.75280.56520.1522
CORNIA [29]0.94730.94520.79538.74780.74510.65420.47700.82470.80440.73250.54640.1554
DIIVINE [30]0.91340.91200.748711.0960.72940.67350.49470.85040.80770.75940.57180.1546
Proposed Model0.95690.94560.81434.67690.94530.92890.78510.40060.94090.92630.78040.0932
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ryu, J. Adaptive Feature Fusion and Kernel-Based Regression Modeling to Improve Blind Image Quality Assessment. Appl. Sci. 2023, 13, 7522. https://doi.org/10.3390/app13137522

AMA Style

Ryu J. Adaptive Feature Fusion and Kernel-Based Regression Modeling to Improve Blind Image Quality Assessment. Applied Sciences. 2023; 13(13):7522. https://doi.org/10.3390/app13137522

Chicago/Turabian Style

Ryu, Jihyoung. 2023. "Adaptive Feature Fusion and Kernel-Based Regression Modeling to Improve Blind Image Quality Assessment" Applied Sciences 13, no. 13: 7522. https://doi.org/10.3390/app13137522

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop