Next Article in Journal
Ground State, Magnetization Process and Bipartite Quantum Entanglement of a Spin-1/2 Ising–Heisenberg Model on Planar Lattices of Interconnected Trigonal Bipyramids
Next Article in Special Issue
Missing Value Imputation Method for Multiclass Matrix Data Based on Closed Itemset
Previous Article in Journal
Similarity Changes Analysis for Heart Rate Fluctuation Regularity as a New Screening Method for Congestive Heart Failure
Previous Article in Special Issue
An Improved Moth-Flame Optimization Algorithm with Adaptation Mechanism to Solve Numerical and Mechanical Engineering Problems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Adaptive Feature Fusion Strategy for Image Retrieval

College of Sciences, North Eastern University, Shenyang 110819, China
*
Author to whom correspondence should be addressed.
Entropy 2021, 23(12), 1670; https://doi.org/10.3390/e23121670
Submission received: 3 November 2021 / Revised: 10 December 2021 / Accepted: 10 December 2021 / Published: 12 December 2021
(This article belongs to the Special Issue Advances in Information Sciences and Applications)

Abstract

:
In the era of big data, it is challenging to efficiently retrieve the required images from the vast amount of data. Therefore, a content-based image retrieval system is an important research direction to address this problem. Furthermore, a multi-feature-based image retrieval system can compensate for the shortage of a single feature to a certain extent, which is essential for improving retrieval system performance. Feature selection and feature fusion strategies are critical in the study of multi-feature fusion image retrieval. This paper proposes a multi-feature fusion image retrieval strategy with adaptive features based on information entropy theory. Firstly, we extract the image features, construct the distance function to calculate the similarity using the information entropy proposed in this paper, and obtain the initial retrieval results. Then, we obtain the precision of single feature retrieval based on the correlation feedback as the retrieval trust and use the retrieval trust to select the effective features automatically. After that, we initialize the weights of selected features using the average weights, construct the probability transfer matrix, and use the PageRank algorithm to update the initialized feature weights to obtain the final weights. Finally, we calculate the comprehensive similarity based on the final weights and output the detection results. This has two advantages: (1) the proposed strategy uses multiple features for image retrieval, which has better performance and more substantial generalization than the retrieval strategy based on a single feature; (2) compared with the fixed-feature retrieval strategy, our method selects the best features for fusion in each query, which takes full advantages of each feature. The experimental results show that our proposed method outperforms other methods. In the datasets of Corel1k, UC Merced Land-Use, and RSSCN7, the top10 retrieval precision is 99.55%, 88.02%, and 88.28%, respectively. In the Holidays dataset, the mean average precision (mAP) was 92.46%.

1. Introduction

Image retrieval aims to find images related to images of interest to users from the image dataset. Over the past 20 years, research in this area has focused on improving the accuracy and efficiency of retrieval [1]. There are two common types of image retrieval systems: text-based image retrieval systems and content-based image retrieval systems. Text-based image retrieval systems require experienced experts to tag images, which is expensive and time-consuming [2]. In contrast, content-based image retrieval uses computers to replace the manual processing of many repetitive tasks, reducing the cost of human, financial, and material resources, so content-based image retrieval is gradually becoming the mainstream method. Content-based image retrieval systems fall into two categories [3]: global features indexed with hashing strategies; local scale-invariant features indexed by a vocabulary tree or a k-d tree. Both approaches have advantages and disadvantages, and their performance complements each other [1,3].
More advanced feature extraction techniques and more appropriate feature selection and fusion strategies are crucial to improve image retrieval performance. In recent years, with the development of deep learning and convolutional neural networks, feature extraction techniques for images have improved substantially compared to the previous ones. Compared with traditional features, CNN features have more powerful representation capabilities. The authors in [4] proposed a fast multi-label feature selection method based on information theory feature ranking, which sorts features according to the importance of features, and then selects the top-ranking features. In feature fusion, fusion at the metric level is widely used, but selecting appropriate features and their corresponding weights for fusion to improve retrieval performance is still an important issue. The authors in [5] used average global weights to fuse Color and Texture features for image retrieval. The authors in [6] traverse the weight domain to select optimal weights for retrieval performance as the final weights.
Unlike previous feature selection and weight processing methods, this paper first selects valid features automatically based on the correlation feedback of single features and gives them the same initial weight; then updates the initial weights using the PageRank algorithm to obtain the final weight of each feature; finally calculates a comprehensive metric based on the final weights to output the retrieved results. For feature selection, we first extract some features and then use their retrieval accuracy as trust, then rank the trust and select some of them with higher trust for fusion. In feature fusion, we first give the same initial weights to the selected features; then construct a probability transfer matrix based on their trust; and finally update the feature weights using the PageRank algorithm, normalizing the final result obtained as the weights of each feature.
The rest of the paper is organized as follows. The second section describes some of the related work involved in this paper. The third section describes the feature extraction and processing process. The fourth section describes the feature selection and fusion strategy of this paper in detail. The fifth section presents the experimental results. The sixth section summarizes the work in this paper and gives suggestions for subsequent work.

2. Related Work

Image retrieval starts with creating an image database, inputting the retrieved image, calculating the similarity to the image in the image database, and finally outputting the retrieval result based on the image similarity.
Feature extraction technology is an essential factor affecting the retrieval effect. Before deep learning became popular, feature extraction methods were mainly based on features of color and texture, such as color histograms, color moments, local binary pattern (LBP) features, or some manual feature extractors such as scale invariant feature transform (SIFT), Gabor filters, generalized search trees (GIST) and histogram of oriented gradients (HOG). The development of deep learning technology has brought a significant breakthrough in image retrieval tasks, especially in feature extraction. CNN can be used to extract more representative depth features in images and fusing multiple depth features can help image retrieval tasks achieve better performance.
There has been a lot of work on fusing multiple features to improve retrieval system performance [7,8]. The relevance feedback algorithm [9] has addressed the semantic gap problem that arises in content-based image retrieval. The results obtained through relevance feedback are very similar to human understanding [10,11]. The main steps of relevance feedback are as follows: first, the retrieval system provides primary retrieval results based on the retrieval keywords provided by the user; then the user makes a judgment on the current retrieval results, which retrieval results are relevant to their retrieval needs and which retrieval results are irrelevant; and then the retrieval system will provide new retrieval results based on the feedback of users. This process is repeated until the search results meet the user’s needs.
The PageRank algorithm is a link analysis algorithm proposed by Google founders Larry Page and Sergey Brin in 1997 and is often used in the field of web search [12,13]. It creates a probability transfer matrix based on the links between web pages and continuously updates the initial probability vector based on the probability transfer matrix to obtain a stable probability vector to represent the probability distribution of users visiting each web page. There have been a lot of improved applications based on the traditional PageRank algorithm. The authors in [14] used the PageRank algorithm to enhance the Peer-Led Team Learning (PLTL) teaching method. The authors in [15] proposed a new algorithm for ranking online shops based on the PageRank algorithm. The authors in [16] proposed a new method based on PageRank to find the most influential individuals in online social networks.
In this paper, we propose a feature selection and fusion strategy based on information entropy and correlation feedback. We first extract features and preprocess the features, then construct similarity using information entropy and obtain the retrieval accuracy of each feature based on relevant feedback and use the accuracy as a trust to select features. Finally, we initialize the feature weights with the mean value, update the weights with the PageRank algorithm to obtain the comprehensive similarity, and obtain the retrieval results based on the comprehensive similarity.

3. Feature Extraction and Preprocess

In this paper, we extract multiple CNN features and traditional color and texture features simultaneously. The specific extraction process is as follows: for CNN features, we use the pre-trained model trained in the ImageNet dataset to extract the fc2 layer features of the original image and grayscale map, respectively. Here, fc2 refers to the features of the second fully connected layer of the network. Of course, in specific experiments, we can also extract features from other appropriate layers. Then, we stitch the extracted two parts of features to get a one-dimensional feature vector. For color and texture features, we extract color histogram features, denoted as color, LBP features, and GIST features for the original image only.
The dimensionality of tandem features is high, and noise points in image feature descriptors may reduce retrieval accuracy. In addition, using the high-dimensional image feature descriptors as a benchmark to calculate the similarity between the query image and the image in the database is time-costly and reduces retrieval efficiency. Based on the above two points, we need to reduce the dimensionality of the data features. Principal component analysis (PCA) is one of the common methods for data dimensionality reduction, which maps high-dimensional data to low-dimensional space with large variance by matrix mapping. It can reduce the dimensionality of data, reduce the computational effort, and avoid the loss of information as much as possible. In the experiments of this paper, we use the classical PCA algorithm to reduce the image features to different dimensions and compare the retrieval results of the features under different dimensions, which are described in Part 5 of the paper.
Since it takes a lot of time to calculate the similarity measure of multiple features, hash coding is needed to reduce the subsequent computation. The authors use binary to encode image features [17] as follows:
a v e ( F i ) = j = 1 m F ( i , j ) m
F ( i , j ) = { 1             F ( i , j ) a v e ( F i ) 0             F ( i , j ) < a v e ( F i )
where i { 1 , 2 , , k }   a v e ( F i ) is the mean of the feature F i , m is the dimension of the feature F i , F ( i , j ) and is the j -th component of the feature F i .

4. Proposed Method

4.1. Method Introduction

In image retrieval methods based on multi-feature fusion, feature selection and weight determination are very important to improve retrieval accuracy. The traditional method first extracts the color, texture, or other effective features of the image and then manually selects several features for fusion. The disadvantage of this approach is that the same feature contributes differently to image retrieval in different datasets. For a given dataset, color information may be more prominent, texture information may be more prominent, or other information may be more prominent. It is difficult to determine the features suitable for the dataset and determine the optimal combination of features in advance in image retrieval work. If we try different combinations for all datasets, we can get better results, but the time consumption is vast, the efficiency is low, and the algorithm is not fixed. In response, this paper proposes a feature selection strategy using relevant feedback based on information entropy. The basic process of the retrieval system is shown in Figure 1.
As shown by Figure 1, for each retrieved image, we first extract the features of the image and process the features using PCA algorithm and hash coding; then combine information entropy and correlation feedback to select the features. Next, we initialize the feature weights using the mean value, update the weights using the PageRank algorithm. Finally, we calculate the combined similarity using selected features and final weights and output the retrieval result based on similarity.

4.2. Feature Selection Strategy

Information entropy is often used as a quantitative indicator of the information content of a system. In statistics, the greater the entropy, the smaller the distribution of discrimination; the smaller the entropy, the greater the discrimination of distribution. In [18], the authors used information entropy to calculate the weights of clusters, which is calculated as follows:
H ( x ) = j = 1 N p ( x ) log 2 p ( x ) 1
Influenced by this idea, this paper proposes a new form of information entropy to measure the differentiation in each dimension of a single feature and assign different weights to each dimension. The smaller the entropy, the greater discrimination of that feature dimension, the greater contribution to the distance calculation and the greater the weight, and vice versa. Then, we construct a weighted distance function to obtain the similarity between images.
In this paper, we use the trust obtained based on entropy and relevant feedback to select these functions. Suppose there are k extracted features, denoted as F i ( p ) , F i ( q ) ( i { 1 , 2 , , k } ) , q is the image to be retrieved, p is the image in the database. Firstly, we calculate the retrieval accuracy of a single feature based on users’ feedback on the retrieval results. Then, the retrieval accuracy is regarded as the trust of a single feature, and the trust is used to select features.
Firstly, under feature F i , calculate the distance between the retrieved image q and all images in the database, and get the similarity S i ( q ) between them based on the distance. The calculation process of similarity S i ( q ) is:
M i ( q ) = j = 1 n w j | F q ( i , j ) F p ( i , j ) |
S i ( q ) = 1 1 q = 1 n M i ( q ) ( M i ( 1 ) , M i ( 2 ) , , M i ( n ) )
where S i ( q ) is the similarity of the retrieved image q with other images in the database calculated based on the feature F i , n is the total number of images in the image database. When calculating distances, we use Manhattan distances with weights. These weights are constructed using information entropy. w j is the weight of the j -th component of feature F i , is calculated as Equations (6) and (7):
w j = e ( 1 - h j ) j = 1 m e ( 1 - h j )
h j = 1 log 2 n i = 1 n ( f i j i = 1 n f i j log 2 f i j i = 1 n f i j )
here, i { 1 , 2 , , n } , j { 1 , 2 , , m } , n is the total number of images in the image library, m is the dimensionality of the feature, f i j is the j -th dimensional feature of the i -th image, and h j is the entropy of the j -th component of F i . When the values of each dimension of the feature are equal, the entropy and weights of each dimension are also equal, and the resulting entropy h j is 1.
Then, the similarity S i ( q ) is sorted, and t samples from the image database with high similarity with the retrieved image q are returned as the retrieval results, denoted as L i = { L 1 i , L 2 i , , L t i } , where t is the predefined number of returned images. The query precision is calculated according to the retrieval results and relevant feedback. The query precision is the trust of a single feature, denoted as P ( i ) , i { 1 , 2 , , k } . In the Holidays dataset, the accuracy is measured by mAP, and in the other datasets, the accuracy is measured by precision. After the single feature trust is obtained, they are ranked from highest to lowest, and the top 5 features are selected for fusion based on trust.

4.3. Post-Processing of Weights by the PageRank Algorithm

Influenced by the PageRank algorithm, this paper takes the probability vector as a feature weight to measure the importance of each feature in feature fusion and uses the PageRank algorithm to obtain the final weights. Firstly, we construct the probability transfer matrix according to the retrieval trust of all selected features as the preference and then update the initial weights so that the features with higher trust can obtain higher fusion weights. Using the PageRank algorithm to update weights has the following two advantages: (1) the weight of the adaptive means that we do not need to design the weight function manually; (2) the time complexity is greatly reduced. Compared with other weight updating algorithms, the PageRank algorithm does not need the help of query results or loss function and only needs to iteratively use the probability transfer matrix, which greatly reduces the calculation amount of weight updating.
Firstly, we calculate the transfer matrix of the degree of retrieval performance preference between every single feature, denoted as T M k k = { T M ( x , y ) } . T M ( x , y ) is the preference of feature x , x { 1 , 2 , , k } to feature y , y { 1 , 2 , , k } , which is constructed as follows:
T M ( x , y ) = [ P ( x ) - P ( y ) ] + 1 2
Then, each column T M k k = { T M ( x , y ) } is normalized by dividing each value in the column by the sum of the values in that column. Finally, using T M k k to obtain the weights of single features. The process is as following Algorithm 1:
Algorithm 1. Post-Processing of Weights by the PageRank Algorithm [13]
Input: the probability transfer matrix T M , hyper-parameter of γ [ 0 , 1 ] and ε , the initial weight w 0 .
Output: the final weight w d .
1: Initialize the weight w 0 = 1 / k .
2: d = 1
3: repeat
4:             w d = γ w d 1 + ( 1 γ ) T M w d 1
5:             w d = w d / s u m ( w d )
6:             d = d + 1
7: until  w d - w d - 1 < ε
8: Return wd
where w 0 is the initial weight, d refers to the number of iterations, ε , γ is the parameter, in this paper ε = 0.005 , γ = 0.4 . After post-processing by the PageRank algorithm, the final weights are used to calculate the integrated similarity measure:
S ( p ) = i = 1 k w d i S i ( q )
Finally, the retrieval results are output according to the integrated similarity metric.

5. Performance Evaluation

5.1. Datasets

To test the effectiveness of the method proposed in this paper, we conducted experiments in four image retrieval datasets: Holidays, Corel-1k, UC Merced Land-Use, and RSSCN7. The details of which are as follows.
Holidays dataset, containing 1491 holiday photos, in 500 categories, where the first image in each category is used as the retrieved image and the remaining images are the corresponding related images. The evaluation index for its retrieval performance is mAP.
Wang (Corel 1K) dataset, which contains 1000 images, is divided into ten categories: flowers, horses, buildings, beaches, etc. Each category contains 100 images. The precision of Top-N is used as the evaluation index for the retrieval system.
The UC Merced Land-Use dataset, a remote sensing dataset describing land use, was released by the University of California and includes 2100 images divided into ten categories. The image size is 256 × 256, and the pixel resolution is 30 cm. Again, we used precision as the evaluation index for the retrieval system.
The RSSCN7 dataset, the remote sensing scene dataset, contains seven categories; each category includes 400 images, each with a pixel size of 400 × 400. These images are sampled at four different scales in different seasons and weather variations. We also used precision as the evaluation index for the retrieval system.

5.2. Evaluation Index

The precision of the first N images (top-N) is calculated as follows:
precision = N r N
N r refers to the number of the first N returned images that belong to the same category as the retrieved image and N refers to the total number of returned images.
The mAP is the mean value of average precision (AP) obtained from multiple queries, which is presented by:
m A P = 1 C k = 1 C A P k
A P = 1 N r i = 1 N i L O C i
where A P k refers to the A P k value obtained by the query of the k-th image, N r the number of the first N returned images that belong to the same category as the retrieved image L O C i refers to the location of the i-th image in the returned image that belongs to the same category as the retrieved image. For example, if five images belonging to the same category as the query image appear at positions 1, 3, 6, 9, and 10, the AP is represented as follows:
A P = ( 1 / 1 + 2 / 3 + 3 / 6 + 4 / 9 + 5 / 10 ) / 5 = 0.62

5.3. Experiment Introduction

In order to prove the effectiveness of the proposed method, we conducted experiments in four datasets: Holiday, Corel-1k, UC Merced Land Use, and RSSCN7, and analyzed the experimental results. First, we obtained and compared the following three kinds of accuracy: (1) the retrieval accuracy using a single feature; (2) the accuracy obtained by selecting five features using the method in this paper and then fusing them using the average value as the weight (AVG); (3) the accuracy obtained using the feature selection and fusion strategy proposed in this paper (Ours). Then, to further prove the effectiveness of the feature selection strategy, we analyze the relationship between the retrieval accuracy of features and the number of times they are selected. After that, to further prove the effectiveness of the PageRank algorithm, we analyzed the relationship between the retrieval accuracy of features and the final weights. Finally, we give a visualization map of the retrieval results for each dataset. All the code used in this article is shared at https://github.com/JNZYBOBO/AZYBOBO.

5.4. Analysis of Experimental Results

5.4.1. Holidays Dataset

In the Holidays dataset, we selected the first image of each category as the retrieved images, 500 images in total, and the others as image database images. Firstly, we obtained the retrieval accuracy of each feature, AVG, and ours under different dimensionality reduction results, and the results are shown in Table 1.
From Table 1, we can see the impact of different dimensions on the retrieval results. The retrieval results of each dimension are basically the same. If only the retrieval accuracy is considered, we can choose to reduce the dimension to about 300 dimensions. If the retrieval efficiency is considered, we can choose to reduce the dimension to a lower level.
The results based on single-feature retrieval show that CNN features achieve better results in single-feature image retrieval, while Color, LBP, Gist, and other color or texture features are less effective. This is because color features cannot describe the local distribution of colors and the spatial location of each color in the image; texture features are easily affected by factors such as illumination and reflection. Meanwhile, CNN features can fully consider the local and global information of the image since they are not easily affected by factors such as location and illumination and are highly resistant to noise.Different features have different effects on image retrieval results and features with stronger expressive power can be selected using the feature selection strategy proposed in this paper. Compared with any single-feature image retrieval, using our feature selection strategy can obtain more than 10 percentage points of mAP improvement. Further, using our PageRank algorithm to update the initial weights also improves the mAP by 1–2 percentage points in different dimensions.
To further prove the rationality of the feature selection strategy proposed in this paper, we counted the ratio of each feature that was selected in 500 retrieved images and compared the relationship between the retrieval accuracy of each feature and the number of times it was selected; the results are shown in Figure 2. In the figure, “Mean mAP” refers to the average value of mAPs for single feature retrieval in different dimensions in Table 1, and “Ratio” refers to the ratio of each feature selected, i.e., the number of times it is selected divided by the number of retrieved images (500). According to Figure 2, The trend of the two broken lines is approximately the same, and the higher the retrieval accuracy of the features, the higher the probability of being selected. Thus, it can be shown that our feature selection method can select features with good results.
In order to further prove the effectiveness of the PageRank algorithm proposed in this paper, we use 20 retrieval images to obtain the mAP of the selected five features and the final weight after updating, a total of 100 mAP values and 100 weight values. The relationship between the two is shown in Figure 3. From the figure, we can see two things: (1) the retrieval accuracy of the selected features is usually high, which further illustrates the rationality of our feature selection strategy; (2) in general, for each feature, the greater the retrieval accuracy, the greater the final weight. Thus, in feature fusion, the PageRank algorithm can give higher weight to the features with higher retrieval accuracy, so that the fused features have a better retrieval effect.
In the multi-feature fusion-based image retrieval system, compared with manual feature selection and design of weight functions, the feature selection strategy proposed in this paper can automatically select the most superior features for different images, while using the PageRank algorithm can automatically update the feature weights and adjust the weight combinations according to the feature trust. Therefore, it has higher retrieval accuracy and robustness. Table 2 shows the comparison between the experimental results in the Holidays dataset and other methods for manual selection of features and fusion weights. The optimal precision of image retrieval based on this method is 0.9246. Compared with the literature [19], retrieval accuracy increased by about 0.0126, compared with the literature [17], retrieval accuracy increased by about 0.0401, compared with the literature [20], retrieval accuracy increased by about 0.0446, compared with the literature [21], retrieval accuracy increased by about 0.0526, compared with the literature [2], retrieval accuracy increased by about 0.0556, compared with the literature [3], and retrieval accuracy increased by about 0.0694.

5.4.2. Corel-1k Dataset

For the Corel-1K dataset, we randomly selected 20% of the images in each category as the images to be retrieved, for a total of 200 images, and the rest as the database images. Firstly, we reduce the dimensionality of each extracted feature by the PCA dimensionality reduction algorithm and then obtain the precision of top10, top20, top30, top40, and top50 in different dimensions. The results are shown in Figure 4. As we can see, image retrieval accuracy is highest when the features drop to 10 dimensions among any number of returned images. The reduction in dimensionality not only improves precision but also reduces computational effort and improves retrieval efficiency.
From the comparison of Table 1 and Table 3, we can see that the same features have different retrieval effects in different datasets, like AlexNet, ResNet50 works well in the Holidays dataset but poorly in the Corel-1k dataset, and NASNetLarge on the contrary. Furthermore, the features required for different images in the same dataset are different. According to the traditional method, we cannot determine the suitable features for different datasets or different images in the same dataset in advance. In contrast, using our feature selection strategy ensures that for each retrieved image, as long as there is a feature to be selected with better effect, it will definitely be selected, and the fused effect will definitely be better than the best retrieval effect of all features. Table 3 compares three kinds of accuracy on the different number of returned images when the feature dimension is 10. It can be seen that in different return numbers, the feature selection strategy in this paper has achieved better detection results. The detection accuracy is nearly 10 percentage points better than that of any single feature. At the same time, the use of the PageRank algorithm also makes the detection accuracy in different top indicators achieve a certain degree of improvement.
Similarly, we also counted the ratio of each feature being selected on 200 retrieved images, and the relationship between these ratios and retrieval accuracy is shown in Figure 5. In the figure, “Mean Precision” refers to the average value of precision for single feature retrieval in different top indicators in Table 3, and “Ratio” still refers to the rate at which each feature is selected. We can see that compared with Figure 2, the trend between the two is more consistent in Figure 5, indicating that the feature selection strategy is more effective in the Corel-1K dataset, which can also be seen from the final experimental results in Table 1 and Table 3. This is because the Holidays dataset has more categories and fewer samples in each category, which are easily affected by randomness and lead to unstable results. In contrast, the Corel-1K dataset has fewer categories and a larger number of images in each category, which is more robust.
Figure 6 shows the optimization effect of the PageRank algorithm on fusion weights. According to Figure 6, it can be seen that the PageRank algorithm optimizes the fusion weights according to the retrieval accuracy, giving higher weights to the excellent features to make the best final retrieval results. Compared with the Holidays dataset, the effect is more obvious in the Corel-1K dataset.
At the same time, we compared the results with the results of other papers in the Corel-1K dataset, as shown in Table 4. The accuracy of our method is 0.9955. Compared with the methods proposed in other papers, our method has obvious advantages.

5.4.3. UC Merced Land Use and RSSCN7 Dataset

Similarly, we conducted experiments in the UC Merced Land Use and RSCN7 datasets. As with the Corel-1K dataset, we also used 20% of each category of images as the retrieved images and the rest as the database images. According to the results after tuning the parameters, we choose to reduce the features extracted from the RSCN7 dataset to 40 dimensions and the features extracted from the UC Merced Land Use dataset to 20 dimensions, after which we obtain the retrieval accuracy under different top metrics, and the results are shown in Table 5 and Table 6. Table 5 shows the precision on different tops in the UC Merced Land Use dataset, and Table 6 shows the precision on different tops in the RSSCN7 dataset. The detection results in both datasets also demonstrate the rationality of the proposed method in this paper. With different evaluation metrics, the detection results are greatly improved by using our method in both cases.
We also demonstrate the effectiveness of the feature selection strategy and weight update strategy in both datasets. In the UC Merced Land Use and RSSCN7 datasets, we selected 10 images in each category to obtain the relationship between the selected ratio and the mean precision of each feature. In addition, we selected 21 images to obtain the relationship between the retrieval accuracy and the final weights, and all the relationships are shown in Figure 7. From Figure 7, we can see that the proposed method still has good results in these two datasets.

5.4.4. Visualization Result

Lastly, in order to make the experimental results clearer, we conducted tests in all datasets to visualize the retrieval results returned, as shown in Table 7. In the “Results and similarity” column, images with green borders are correct results, images with red borders are incorrect results, and the numbers below the images are their similarity to the retrieved images. In the Holidays database, because only two images are correct results of retrieval image, we used mAP to calculate the precision, and the precision was 1.0. In other databases, we used top-10 to calculate the precision; the results were 1.0, 1.0, and 0.9.

6. Conclusions

The multi-feature fusion strategy can improve the accuracy and generalization ability of the retrieval system. Suitable feature selection strategies and feature fusion strategies can help enhance the retrieval performance of the retrieval system ulteriorly. In this paper, we use information entropy and relevant feedback to automatically select features, use the PageRank method to process a single feature’s weight, obtain the comprehensive similarity according to the weight after processing, and finally output the retrieval results. Experimental results in the Holidays, Corel-1k, UC Merced Land-Use, and RSSCN7 show that the proposed method has better retrieval performance than single feature image retrieval systems and image retrieval systems without the PageRank algorithm. The experiments in the Holidays and Corel-1k datasets show that the proposed method is better than the existing methods. The next step is to investigate how to select the appropriate parameters and improve the retrieval performance after the features have been selected.

Author Contributions

Data curation, L.N. and J.W.; formal analysis, Q.C.; methodology, X.L.; software, L.Z. and L.N.; writing—original draft, X.L.; writing—review and editing, L.Z. and Q.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the fundamental research funds for the central universities, N180503017.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data is contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, R.; Wei, S.; Zhao, Y.; Yang, Y. Applications, Indexing of the CNN features for the large scale image search. Multimed. Tools 2018, 77, 32107–32131. [Google Scholar] [CrossRef] [Green Version]
  2. Shi, X.; Guo, Z.; Zhang, D. Efficient Image Retrieval via Feature Fusion and Adaptive Weighting. In Proceedings of the Chinese Conference on Pattern Recognition, Chengdu, China, 5–7 November 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 259–273. [Google Scholar]
  3. Zhou, Y.; Zeng, D.; Zhang, S.; Tian, Q. Augmented feature fusion for image retrieval system. In Proceedings of the 5th ACM on International Conference on Multimedia Retrieval, Shanghai, China, 23–26 June 2015; pp. 447–450. [Google Scholar]
  4. Lee, J.; Kim, D.-W. Fast multi-label feature selection based on information-theoretic feature ranking. Pattern Recognit. 2015, 48, 2761–2771. [Google Scholar] [CrossRef]
  5. Kong, F.-H. Image retrieval using both color and texture features. In Proceedings of the International Conference on Machine Learning and Cybernetics, Baoding, China, 12–15 July 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 2228–2232. [Google Scholar]
  6. Yue, J.; Li, Z.; Liu, L.; Fu, Z. Content-based image retrieval using color and texture fused features. Math. Comput. Model. 2011, 54, 1121–1127. [Google Scholar] [CrossRef]
  7. De Herrera, A.G.S.; Schaer, R.; Markonis, D.; Müller, H. Comparing fusion techniques for the ImageCLEF 2013 medical case retrieval task. Comput. Med. Imaging Graph. 2015, 39, 46–54. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Liu, Z.; Blasch, E.; John, V. Statistical comparison of image fusion algorithms: Recommendations. Inf. Fusion 2017, 36, 251–260. [Google Scholar] [CrossRef]
  9. Zhou, X.S.; Huang, T.S. Relevance feedback in image retrieval: A comprehensive review. Multimed. Syst. 2003, 8, 536–544. [Google Scholar] [CrossRef]
  10. Wang, X.; Wang, Z. A novel method for image retrieval based on structure elements’ descriptor. J. Vis. Commun. Image Represent. 2013, 24, 63–74. [Google Scholar] [CrossRef]
  11. Bian, W.; Tao, D. Biased discriminant Euclidean embedding for content-based image retrieval. IEEE Trans. Image Process. 2009, 19, 545–554. [Google Scholar] [CrossRef] [PubMed]
  12. Xing, W.; Ghorbani, A. Weighted pagerank algorithm. In Proceedings of the 2nd Annual Conference on Communication Networks and Services Research, Fredericton, NB, Canada, 21 May 2004; IEEE: Piscataway, NJ, USA, 2004; pp. 305–314. [Google Scholar]
  13. Page, L.; Brin, S.; Motwani, R.; Winograd, T. The PageRank Citation Ranking: Bringing Order to the Web; Stanford InfoLab: Stanford, CA, USA, 1999; pp. 1–17. [Google Scholar]
  14. Zaki, N. PageRank Algorithm to Improve the Peer-Led Team Learning Pedagogical Approach. In Smart Education and e-Learning 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 225–234. [Google Scholar]
  15. Scholz, M.; Pfeiffer, J.; Rothlauf, F. Using PageRank for non-personalized default rankings in dynamic markets. Eur. J. Oper. Res. 2017, 260, 388–401. [Google Scholar] [CrossRef]
  16. Le, V.V.; Nguyen, H.T.; Snasel, V.; Dao, T.T. Identify influential spreaders in online social networks based on social meta path and PageRank. In International Conference on Computational Social Networks; Springer: Berlin/Heidelberg, Germany, 2016; pp. 51–61. [Google Scholar]
  17. Lu, X.; Wang, J.; Li, X.; Yang, M.; Zhang, X. An adaptive weight method for image retrieval based multi-feature fusion. Entropy 2018, 20, 577. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Zheng, W.; Mo, S.; Duan, P.; Jin, X. An improved pagerank algorithm based on fuzzy C-means clustering and information entropy. In Proceedings of the 3rd IEEE International Conference on Control Science and Systems Engineering (ICCSSE), Beijing, China, 17–19 August 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 615–618. [Google Scholar]
  19. Zhou, J.; Gan, J.; Gao, W.; Liang, A. Image retrieval based on aggregated deep features weighted by regional significance and channel sensitivity. Inf. Sci. 2021, 577, 69–80. [Google Scholar] [CrossRef]
  20. Zheng, L.; Wang, S.; Tian, L.; He, F.; Liu, Z.; Tian, Q. Query-adaptive late fusion for image search and person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1741–1750. [Google Scholar]
  21. Gkelios, S.; Sophokleous, A.; Plakias, S.; Boutalis, Y.; Chatzichristofis, S.A. Deep convolutional features for image retrieval. Expert Syst. Appl. 2021, 177, 114940. [Google Scholar] [CrossRef]
  22. Yu, L.; Liu, N.; Zhou, W.; Dong, S.; Fan, Y.; Abbas, K. Weber’s law based multi-level convolution correlation features for image retrieval. Multimed. Tools Appl. 2021, 80, 19157–19177. [Google Scholar] [CrossRef]
  23. Singhal, A.; Agarwal, M.; Pachori, R.B. Directional local ternary co-occurrence pattern for natural image retrieval. Multimed. Tools Appl. 2021, 80, 15901–15920. [Google Scholar] [CrossRef]
  24. Kayhan, N.; Fekri-Ershad, S. Content based image retrieval based on weighted fusion of texture and color features derived from modified local binary patterns and local neighborhood difference patterns. Multimed. Tools Appl. 2021, 80, 32763–32790. [Google Scholar] [CrossRef]
  25. Xie, G.; Guo, B.; Huang, Z.; Zheng, Y.; Yan, Y. Combination of dominant color descriptor and Hu moments in consistent zone for content based image retrieval. IEEE Access 2020, 8, 146284–146299. [Google Scholar] [CrossRef]
  26. Bhunia, A.K.; Bhattacharyya, A.; Banerjee, P.; Roy, P.P.; Murala, S. A novel feature descriptor for image retrieval by combining modified color histogram and diagonally symmetric co-occurrence texture pattern. Pattern Anal. Appl. 2020, 23, 703–723. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The proposed retrieval system framework.
Figure 1. The proposed retrieval system framework.
Entropy 23 01670 g001
Figure 2. The relationship between retrieval accuracy and selected ratio.
Figure 2. The relationship between retrieval accuracy and selected ratio.
Entropy 23 01670 g002
Figure 3. The relationship between retrieval accuracy and final weights.
Figure 3. The relationship between retrieval accuracy and final weights.
Entropy 23 01670 g003
Figure 4. Precision in the Corel-1k dataset.
Figure 4. Precision in the Corel-1k dataset.
Entropy 23 01670 g004
Figure 5. The relationship between retrieval accuracy and selected ratio.
Figure 5. The relationship between retrieval accuracy and selected ratio.
Entropy 23 01670 g005
Figure 6. The relationship between retrieval accuracy and final weights.
Figure 6. The relationship between retrieval accuracy and final weights.
Entropy 23 01670 g006
Figure 7. Two sets of relations in the last two datasets.
Figure 7. Two sets of relations in the last two datasets.
Entropy 23 01670 g007
Table 1. Retrieval performance comparison in the Holidays dataset.
Table 1. Retrieval performance comparison in the Holidays dataset.
FeaturesDimensions
50040035030025020015010050
Color0.08480.08740.08760.08710.08690.08700.08700.08900.0857
InceptionResNetV20.70920.73420.73060.73220.73240.72380.71210.68420.5762
MobileNetV20.71990.73570.74680.75270.74590.74140.72890.70860.6175
NASNetLarge0.31090.32320.33050.33020.33920.35250.34970.33780.2774
ResNet500.78200.79030.79360.80130.80460.79530.79080.75330.6672
VGG160.71280.73840.74320.75530.75080.73330.72950.69910.6046
VGG190.71070.74440.74890.74950.75670.74060.72440.68330.5987
Xception0.74010.75660.76370.76230.76960.75890.75110.70820.6355
AlexNet0.68080.70070.71160.72410.72120.71720.70070.66810.5805
GIST0.25520.26400.28040.27780.29090.28410.29230.27730.2324
LBP0.32440.32440.32440.32440.32440.32720.31240.29090.2617
DenseNet0.75480.78470.79630.80470.80470.79600.78510.75980.6687
EfficientNetB00.76130.77400.78500.78730.79440.77950.76800.73640.6299
AVG0.90180.90280.91140.91700.91320.91430.91250.90750.8981
Ours0.91940.91880.92460.92450.91780.92390.92160.91500.9050
Table 2. Comparison of retrieval results in the Holidays dataset.
Table 2. Comparison of retrieval results in the Holidays dataset.
MethodOurs[19][17][20][21][2][3]
mAP0.92460.91200.88450.88000.87200.86900.8552
Table 3. Precision for different numbers of returned images in the Corel-1k dataset.
Table 3. Precision for different numbers of returned images in the Corel-1k dataset.
FeaturesTop10Top20Top30Top40Top50
Color0.29950.28400.27880.26860.2613
InceptionResNetV20.91150.90070.88230.87090.8570
MobileNetV20.91750.91170.90400.89490.8795
NASNetLarge0.93350.92550.91150.90610.8814
ResNet500.23100.24820.23630.22260.2315
VGG160.89700.87770.85930.84360.8278
VGG190.87850.86970.54930.83640.8140
Xception0.91550.91350.89930.87750.8495
AlexNet0.28470.28470.28470.28470.2847
GIST0.11620.11620.11620.11620.1162
LBP0.12450.12450.12450.12450.1245
DenseNet0.86420.86420.86420.86420.8642
EfficientNetB00.89570.89570.89570.89570.8957
AVG0.99350.98900.98760.98610.9851
Ours0.99550.99100.98950.98720.9856
Table 4. Comparison of search results in the Corel-1k dataset.
Table 4. Comparison of search results in the Corel-1k dataset.
MethodOurs[17][22][23][24][25][26]
Precision0.99550.97090.86500.85500.82500.80400.8000
Table 5. Precision for different numbers of returned images in the UC Merced Land Use dataset.
Table 5. Precision for different numbers of returned images in the UC Merced Land Use dataset.
FeaturesTop10Top20Top30Top40Top50
Color0.11020.09960.09370.08790.0876
InceptionResNetV20.53690.48250.43720.40320.3744
MobileNetV20.61520.55820.50840.47470.4417
NASNetLarge0.60260.54990.50350.46770.4355
ResNet500.17400.15790.14380.13560.1308
VGG160.55520.50150.46000.43470.4095
VGG190.54800.50080.45860.42920.4035
Xception0.55880.49970.45330.42240.3930
AlexNet0.20380.20380.20380.20380.2038
GIST0.10330.10330.10330.10330.1033
LBP0.12690.12690.12690.12690.1269
DenseNet0.53590.53590.53590.53590.5359
EfficientNetB00.53690.53690.53690.53690.5369
AVG0.86830.81170.77730.73680.6989
Ours0.88020.83340.78790.74470.7067
Table 6. Precision for different numbers of returned images in the RSSCN7 dataset.
Table 6. Precision for different numbers of returned images in the RSSCN7 dataset.
FeaturesTop10Top20Top30Top40Top50
Color0.30190.33230.32530.31130.3026
InceptionResNetV20.53910.50110.47390.45090.4370
MobileNetV20.56230.51890.48770.46490.4495
NASNetLarge0.53690.50250.47630.44610.4114
ResNet500.35660.33900.32850.32260.3015
VGG160.53160.49320.46080.44360.4278
VGG190.54170.50630.48480.46640.4340
Xception0.53530.49840.46670.45750.4495
AlexNet0.25860.25860.25860.25860.2586
GIST0.18290.18290.18290.18920.1892
LBP0.20310.20310.20310.20310.2031
DenseNet0.42250.42250.42250.42250.4225
EfficientNetB00.42440.42440.42440.42440.4244
AVG0.86510.80830.76830.72910.6851
Ours0.88280.83260.78380.73720.6856
Table 7. Examples of image retrieval in different databases.
Table 7. Examples of image retrieval in different databases.
DatabaseHolidaysCorel1kUC Merced Land UseRSSCN7
Retrieval image Entropy 23 01670 i001 Entropy 23 01670 i002 Entropy 23 01670 i003 Entropy 23 01670 i004
Results and similarity Entropy 23 01670 i005
0.99952937
Entropy 23 01670 i006
0.99975951
Entropy 23 01670 i007
0.99985633
Entropy 23 01670 i008
0.9997895
Entropy 23 01670 i009
0.99932636
Entropy 23 01670 i010
0.99974828
Entropy 23 01670 i011
0.99983305
Entropy 23 01670 i012
0.99977838
Entropy 23 01670 i013
0.99932212
Entropy 23 01670 i014
0.99974674
Entropy 23 01670 i015
0.99983272
Entropy 23 01670 i016
0.99976673
Entropy 23 01670 i017
0.99931696
Entropy 23 01670 i018
0.99970117
Entropy 23 01670 i019
0.99982084
Entropy 23 01670 i020
0.99976315
Entropy 23 01670 i021
0.9992626
Entropy 23 01670 i022
0.99969666
Entropy 23 01670 i023
0.99982064
Entropy 23 01670 i024
0.99976234
Entropy 23 01670 i025
0.99917911
Entropy 23 01670 i026
0.99969415
Entropy 23 01670 i027
0.99982037
Entropy 23 01670 i028
0.99975622
Entropy 23 01670 i029
0.99915044
Entropy 23 01670 i030
0.99965442
Entropy 23 01670 i031
0.99982006
Entropy 23 01670 i032
0.99975168
Entropy 23 01670 i033
0.99913248
Entropy 23 01670 i034
0.99965017
Entropy 23 01670 i035
0.99980933
Entropy 23 01670 i036
0.99975145
Entropy 23 01670 i037
0.99911954
Entropy 23 01670 i038
0.99964863
Entropy 23 01670 i039
0.99980879
Entropy 23 01670 i040
0.99975058
Entropy 23 01670 i041
0.99909944
Entropy 23 01670 i042
0.99964755
Entropy 23 01670 i043
0.9998086
Entropy 23 01670 i044
0.99974986
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lu, X.; Zhang, L.; Niu, L.; Chen, Q.; Wang, J. A Novel Adaptive Feature Fusion Strategy for Image Retrieval. Entropy 2021, 23, 1670. https://doi.org/10.3390/e23121670

AMA Style

Lu X, Zhang L, Niu L, Chen Q, Wang J. A Novel Adaptive Feature Fusion Strategy for Image Retrieval. Entropy. 2021; 23(12):1670. https://doi.org/10.3390/e23121670

Chicago/Turabian Style

Lu, Xiaojun, Libo Zhang, Lei Niu, Qing Chen, and Jianping Wang. 2021. "A Novel Adaptive Feature Fusion Strategy for Image Retrieval" Entropy 23, no. 12: 1670. https://doi.org/10.3390/e23121670

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop