Next Article in Journal
Opinion Polarization in Human Communities Can Emerge as a Natural Consequence of Beliefs Being Interrelated
Previous Article in Journal
Experimental Study of the Influence of Different Load Changes in Inlet Gas and Solvent Flow Rate on CO2 Absorption in a Sieve Tray Column
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Efficient Retrieval System Framework for Fabrics Based on Fine-Grained Similarity

School of Textile Science & Engineering, Jiangnan University, No. 1800, Lihu Avenue, Wuxi 214122, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(9), 1319; https://doi.org/10.3390/e24091319
Submission received: 27 June 2022 / Revised: 16 August 2022 / Accepted: 17 August 2022 / Published: 19 September 2022
(This article belongs to the Topic Applications in Image Analysis and Pattern Recognition)

Abstract

:
In the context of “double carbon”, as a traditional high energy consumption industry, the textile industry is facing the severe challenges of energy saving and emission reduction. To improve production efficiency in the textile industry, we propose the use of content-based image retrieval technology to shorten the fabric production cycle. However, fabric retrieval has high requirements for results, which makes it difficult for common retrieval methods to be directly applied to fabric retrieval. This paper presents a novel method for fabric image retrieval. Firstly, we define a fine-grained similarity to measure the similarity between two fabric images. Then, a convolutional neural network with a compact structure and cross-domain connections is designed to narrow the gap between fabric images and similarities. To overcome the problems of probabilistic missing and difficult training in classical hashing, we introduce a variational network module and structural module into the hashing model, which is called DVSH. We employ list-wise learning to perform similarity embedding. The experimental results demonstrate the superiority and efficiency of the proposed hashing model, DVSH.

1. Introduction

In the context of the major strategy of “carbon compliance and carbon neutrality”, as one of the traditional industries with high energy consumption, the textile industry is faced with the severe challenge of energy conservation and emission reduction. Improving the intelligence, digitization and automation level of textile enterprises are effective measures to help enterprises save energy and reduce emissions. In addition, with “small batch, multi-variety, tight delivery” increasingly becoming the main production mode of textile enterprises, enterprises have accumulated large amounts of historical production data. How to quickly locate target data in a large amount of data and use them to guide production has become an urgent problem for textile enterprises. At present, there are two methods of fabric retrieval commonly used by textile enterprises: real sample search and text-based image retrieval (TBIR). The former method stores the real samples of fabrics, and the retrieval is carried out by manual comparison, as shown in Figure 1a, which not only takes up storage space, but also has low efficiency and strong subjectivity in manual comparison and retrieval. In addition, textile fabrics will fade with the extension of storage time, which will affect the retrieval results. As shown in Figure 1b TBIR performs fabric retrieval in a semi-manual manner, using human annotations to index fabric images. Although it overcomes the shortcomings of real sample search to a certain extent, the labeling process still makes the retrieval results subjective, resulting in limited retrieval accuracy, and the retrieval methods that can be provided are relatively simple. Content-based image retrieval (CBIR) uses a specific visual representation method to index image content, as shown in Figure 1c, which can overcome the shortcomings of the above methods, so it has become a research hotspot in this field. In this paper, we propose a novel CBIR method to achieve accurate fabric image retrieval.
When receiving a query image, the CBIR system is expected to output a list with the same visual content as the query. Technically speaking, there are two phases in the CBIR, namely, image representation and feature matching. Image representation vectorizes the input images (including queries and images in the database), and the second phase ranks the images in the database for similarity and outputs similar images. Currently, the most challenging task in CBIR is to associate pixel-based, low-level features with human-perceived, high-level semantic features. In many previous works [1,2,3], some hand-crafted feature descriptors were used to represent the visual content of fabric images, such as SIFT [4], LBP [5] and Color Moment [6]. Even though these pixel-level methods have achieved some success, they rely too much on feature engineering, which leads to their limitations in robustness. Recently, convolutional neural networks (CNNs) have achieved outstanding performance in many vision tasks, such as image classification, person identification [7,8], image segmentation and object detection, which demonstrate its good performance in visual description. Therefore, many researchers adopted CNNs for image retrieval tasks. CBIR has seen a significant breakthrough due to the replacement of earlier low-level feature-based algorithms with an end-to-end framework based on deep learning. Inspired by this trend, we focus on the use of deep CNNs to solve the problem of fabric image retrieval. Krizhevsky et al. [9] directly used the convolutional layer in a CNN as the index for the image, and its excellent retrieval performance demonstrated the superiority of deep CNN for image retrieval. However, the disadvantage of this method is its high computational cost, which resulted in a long retrieval time.
To improve retrieval efficiency, a lot of feature optimization and encoding methods were proposed, of which the most commonly used approach was the approximate nearest neighbor search (ANN). At present, deep hashing [10,11,12], which is designed to automatically learn the optimal hash function and generate image hash codes, has become the most popular ANN method. The deep-hashing-based method searches nearest neighbors by measuring the similarity in Hamming space between the generated hash codes.
Differently from general natural images, the abstraction of semantic features in fabric images is not high, which mainly include color, texture and higher-order features formed by their interaction. The different feature types in images make it difficult for general image retrieval methods to directly apply fabric image retrieval. To achieve efficient fabric image retrieval, many researches represented the visuals of fabric images by using hand-crafted feature descriptors, and achieved good performance. However, the success of hand-crafted methods is limited to small datasets or specific fabric types. Our previous work [13,14] attempted to use classification tasks to guide CNN models to learn fabric image representations. Although the retrieval performance far exceeded those of other low-level feature-based methods, the retrieval accuracy still fell short of the retrieval requirements of textile enterprises. There are two main reasons for this problem: (1) the similarity of fabrics cannot be measured by rough classification; (2) the feature loss is severe in the hashing process. To address the problem, in this paper, we first design a fine-grained similarity to measure the similarity between fabrics, and then introduce the structural network and the variational network in the hashing process to reduce the feature loss. Specifically, to narrow the gap between fabric images and similarities, we design a CNN with a compact structure and cross-shortcut connections, which is regarded as the base network of the hashing model. To overcome the problems of probabilistic missing and difficult training in classical hashing, we introduce a variational network module and structural network module into the hashing model (named DVSH). Then, a fine-grained similarity is defined to measure the similarity between two fabric images. To incorporate the defined fine-grained similarity into hash learning, we employ list-wise learning to complete similarity embeddings in mini-batches.

2. Motivation

The goal of image retrieval is to quickly and accurately retrieve relevant images from the target database. In this paper, we believe that two key issues in this task are: (1) how to define the similarity between two images; (2) How to efficiently retrieve relevant images.
At present, there are generally two ways to define similarity between images, namely, binary pairwise similarity [15,16] and soft similarity [17,18]. The former has two ways of measuring similarity: (1) two images are considered similar if they share at least one label—otherwise, they are dissimilar; (2) if the labels of the two images are completely identical, they are considered similar—otherwise, dissimilar. Due to the loss of too much information, the two definitions of binary pairwise similarity are not suitable for measuring the similarity between fabric images. Regarding soft similarity, it is calculated by the degree of fit or cosine distance between the label matrices, which needs to be established on the basis that the categories are independent of each other. For fabric images with multi-granularity features, which contain features at multiple levels, the above similarity definition methods are not applicable. This paper argues that an ideal measure of fabric similarity should be designed according to some well-designed rules, and the similarity of each dimension should be considered. Therefore, according to the characteristics of fabric images, we designed a fine-grained similarity.
Generally, the dimensionality of the features extracted by a CNN is high, which leads to a large computational cost in feature matching (called the “curse of dimensionality”). To solve this situation, some researchers proposed to use principal component analysis (PCA) [19] for linear dimensionality reduction of high dimensional features. PCA expects that the variance between the information in the projected dimension is the largest, so as to obtain the largest amount of information in fewer data dimensions. The interpretation of the principal components in the features extracted by PCA has a certain ambiguity, and the components with less contributions may be more effective for the representation of the samples, and whether the orthogonal vector space of the eigenvalues is unique remains to be discussed. Approximate nearest neighbor search (ANN) is currently a more efficient method that has made substantial progress in the past decade, especially in visual search applications. Hashing methods are typical in ANN; however, naive hashing methods are difficult to preserve the similarity of input features due to the limitation of code length. With the development of deep learning, deep hashing (DH) methods have achieved rapid development. Deep hashing methods map real vectors in Euclidean space to Hamming space. Efficient searching can be achieved by using the generated binary hash codes, which preserves similarity information. The current commonly used deep hashing network consists of a base network and a hash layer, which is guided by a specific objective function, as shown in Figure 2a. The adopted base network generally consists of some basic types of layers (e.g., convolutional and pooling layer), which are used for abstraction and optimization of features. The hash layer generates binary hash codes. This architecture has two drawbacks: first, although it can learn nonlinear features, such models are not based on probabilistic modeling, which may limit their ability to learn diverse features; second, the hash layer actually is a bottleneck layer, which is difficult to train using the backpropagation algorithm and has serious information loss.
The deep variational generative model [20] integrates data distribution priors into the deep model by combining deep neural networks and probability theory, and its effectiveness has been proven in tasks such as image recognition and image segmentation. Therefore, we believe that variational networks have the potential to improve the performances of deep hashing models. As shown in Figure 2b, the variational network can be viewed as consisting of an encoder and a decoder. The role of the encoder is to sample the output of the fully connected layer. The stochastic layer makes the latent output come from a variational distribution parameterized by a probabilistic model (defined by μ and σ 2 ), which provides the model with strong generalizability. The decoder maps the output of the latent layer into the similarity matrix S. To learn more information during the training process, this paper introduces the structure layer [21] in the hash model, as shown in Figure 2c. The structure layer contains multiple modules, where each module represents multiple binary hash codes, and the structure layer has higher dimensions (the number of nodes) than the hash layer, so more information can be obtained during training. In the testing phase, the output of each node does not correspond to a one-bit hash code, but the outputs of the modules are quantized and spliced together to obtain the final hash code. Figure 2d shows the hash model that combines the two key technologies.

3. Fine-Grained Similarity of Fabric Images

The goal of fabric image retrieval is to search for the most similar image set to the query image from the fabric image database, and the key is how to judge whether two fabric images are similar. In this study, we propose to describe the similarity between fabric images in four different dimensions, namely, coarse-texture, fine-texture, color and tightness. Both coarse-texture and fine-texture are texture features (only the observation scale is different). Color is represented by the composition and distribution of colors in a fabric image. The tightness is indicated by the tightness index of the fabric. The proposed similarity is measured from these four dimensions. This section describes how the fine-grained similarity is calculated.

3.1. Similarity of Textures

In our previous works [13,14], we classified fabric images from coarse-texture and fine-texture perspectives, respectively. Here, we classify coarse textures more delicately, as shown in Figure 3. The coarse texture of the fabric is divided into four major categories: solid color, stripe, plaid and pattern, and each category is subdivided into different numbers of minor categories. Visually, the similarity between fabrics with the same major category but different minor categories is lower than the similarity between fabrics with the same minor category and higher than the similarity between fabrics with different major categories. Therefore, we define three levels for the similarity of coarse textures, namely: similar, approximate and dissimilar. Quantitatively, the similarity between similar fabric images is defined as 1, the similarity between approximate fabric images is 0.5 and the similarity between dissimilar fabrics is 0. The four fabric image samples are shown in Figure 4, in which both (a) and (b) are small lattices, and their similarity in coarse-texture is 1; (a) and (c) are lattice fabrics, but (c) is a pane with different grid sizes, and their similarity in coarse-texture is 0.5; (a) and (d) belong to different categories, and their similarity is 0. This method avoids the drawbacks of the commonly used non-0 or 1 similarity definition method, and makes the similarity exist in a “middle zone”, which is more in line with human’s cognition of the similarity of things.
Visually, the differences between fine-textures are obvious. Therefore, for the similarity of fine-texture, this paper simply adopts the commonly used binary pairwise similarity matrix; that is, the similarity between fabrics in the same category is 1, and the similarity between fabrics in different categories is 0.

3.2. Similarity of Colors

Color is one of the most important characteristics of fabrics. In the image, the color of each pixel is converted into a numerical representation and mapped into the RGB space, and each color can be represented by a three-dimensional vector, so the color is the underlying feature at the pixel level. However, the RGB color model was proposed from a hardware perspective, which has poor uniformity and is difficult to match with the visual features observed by the human eye. Compared with RGB, HSV is closer to the human eye’s perception of color, so this color model is often used for color feature extraction in image processing. The name of HSV consists of the first letters of the three components: H for hue, S for color saturation and V for color brightness. This way of defining colors is more suitable for comparisons between colors. Here, referring to OPENCV, we divide the HSV color space into 10 regions, which are represented by { c 1 c i c 10 } . Except for solid fabrics, other types of fabrics generally contain two or more colors. Before measuring the color similarity of fabric images, the color features need to be quantified first. The quantification process of fabric color is divided into two steps:
(1)
First, convert the fabric image in RGB space to HSV space, and then gather the colors of all pixel points into 10 divided areas according to the classification criteria in Table 1;
(2)
Then, calculate the value of the cluster center of each clustered area and set the pixel point x = { x H , x S , x V } . c i represents the area index where this pixel point is located. Then, the value of the cluster’s center can be represented by the mean of all points in the area class:
x ¯ i = x c i x x c i 1
In addition to the value of the cluster’s center, the proportion of each color needs to be calculated. Here, we directly discard the color whose percentage is less than 1%. Then, the color information in the fabric image can be quantified by C = { ( c i , x ¯ i , p i ) , i = 1 , 2 , N } , where N is the number of colors. Figure 5 shows an example of fabric image quantization. In the final quantization result, white, which accounts for less than 1%, is discarded. In the final quantization result, the fabric image shown in (a) contains four colors.
For the fabric color information obtained by the above quantization algorithm, it is difficult to use the general color distance calculation method. Assuming that C 1 = { ( c i , x ¯ i , p i ) , i = 1 , 2 N 1 } and C 2 = { ( c j , x ¯ j , p j ) , j = 1 , 2 N 2 } represent the color quantization results of two fabric images, respectively, their similarity is calculated by the following equations:
S C ( C 1 , C 2 ) = i = 1 N 1 j = 1 N 2 a i j S i j
a i j = h , s , v ( x ¯ i ) x ¯ j ) 2
S i j = [ 1 | p i p j | ] × min ( p i , p j )
where a i j is the Euclidean distance between the two cluster centers in three color components (HSV). The larger the value of S C ( C 1 , C 2 ) , the more similar the two colors are, and vice versa.

3.3. Similarity of Tightness

For coarse and fine texture, the categories are discrete, whereas the fabric tightness is actually continuous. It is generally believed that the more similar the tightness between the fabrics, the higher the similarity. Here, we use the original fabric tightness information to measure the similarity of tightness. For two fabrics with tightness values w 1 and w 2 , their similarity is defined as follows:
S w ( w 1 , w 2 ) = 1 | w 1 w 2 | α · η
The similarity is evaluated according to the difference between the fabric tightness, α represents the distance between adjacent levels and η represents the number of levels. If w 1 w 2 α · η , it means that there is a big difference between them. In that case, the similarity of the tightness between the two fabrics is regarded as 0. If the difference between w 1 and w 2 is small, the value of S w ( w 1 , w 2 ) is close to 1, indicating that they are very similar. In this paper, α = 2 % and η = 20 are configured so that the similarity between fabrics is continuous. This definition of fabric tightness similarity is more refined than the binary pairwise similarity matrix, and is more suitable for describing the similarity between such sequence features.
This section defines the similarity measure of fabrics from four granularities, in which tightness ( S w ) and texture ( S c t and S f t ) are at the semantic level, and color ( S c ) is at the underlying pixel level. Finally, the overall similarity between the two fabric images is expressed as:
S = S c t + S f t + S c + S w 4

4. Deep Variational and Structural Hashing

Assume I = { I 1 , I n } R d × n is the training set used to train the hash model, where I represents the fabric image, d denotes the channel of the images and n represents the number of images in the training set. The goal of hash learning is to learn a hash code generator that can convert input image into binary hash codes while embedding similarity information S into the generated K-dimensional binary hash codes B. The hashing process can be expressed as:
g : I B { 1 , + 1 } K
where g denotes the hash function and K represents the length of generated hash codes. The framework of the proposed hashing model, which consists of base network module, variational network module and structural network module, is shown in Figure 6 (called DVSH).

4.1. Base Network Module

Most of the features in the fabric image are some global low-level features, such as color and texture, and the middle and high-level features produced by their combinations. Studies [22,23,24] have shown that the last layer of the deep convolutional neural network contains the highest-order features that can be extracted from this model. The output of this layer is the deep features learned after several convolution operations. The visual content is highly abstract and contains rich semantic information, such as the location, size, and category of the target. These features are the abstraction of the output of the previous layer, so the output of the previous layer has a lower degree of abstraction. Therefore, the features extracted by deeper convolutional layers are more abstract. Generally, the output of the first and second layers often contains rich color and texture features. In our previous study [25], we designed a deep convolutional neural network with a compact structure and cross-circuit connections to bridge the “semantic gap.” The architecture of the proposed CNN is shown in Figure 7, which contains five convolutional blocks, two short-circuit connections and one full connection. Comparative experiments demonstrated the effectiveness and superiority of this CNN for fabric image representation. Therefore, this study directly applied the CNN shown in Figure 7 as the base network, whose details can be found in reference [25]. The parameters in the base network are denoted as Θ b a s e ; then the features extracted by the base network can be represented by r i = Φ b a s e ( I i , Θ b a s e ) .

4.2. Variational Network Module

Inspired by the success of variational encoders, we employed a probabilistic interpretation to the hashing network. The output of base network is regarded as the latent representation z i , which is assumed to be represented by its posterior distribution p θ ( z i | r i ) . Under this assumption, the posterior probability approximately follows a normal distribution:
q ( z i | r i ) = N ( z i | μ i , σ i 2 I )
By doing a re-parameterization trick, we can sample z i as:
z i l = μ i + σ n ϵ l
ϵ l N ( 0 , 1 )
where ϵ l means the lth sample of noise and ⊙ denotes element-wise multiplication; μ i and σ i would be the output of the nonlinear projection from the hashing network. Then, the proposal distribution should follow a prior distribution over the latent variable defined as a multivariate Gaussian:
p θ ( z ) = N ( z ; 0 , I )
We can enforce this assumption by using Kullback–Liebler divergence (KLD) derived as:
L K L D = K L D ( q ( z i | x i ) | | p θ ( z ) ) = 1 2 j = 1 J ( 1 + log ( ( σ i ( j ) ) 2 ) ( μ i ( j ) ) 2 ( σ i ( j ) ) 2 )
where j is the jth element of μ and σ . The KLD would act as a regularizer of the proposed distribution.

4.3. Structural Network Module

From the latent variable, we impose structure in the succeeding fc layer by splitting it into M blocks such that Φ s t r u c t u r e is parameterized by Θ s t r u c t u r e = { Θ s t r u c t ( m } m = 1 M . Each block would project the latent sample into a distinct semantic representation. Each block vector is represented as:
U m , i = Φ s t r u c t u r e ( m ) ( z i , Θ s t r u c t ( m ) )
where Φ s t r u c t u r e ( m ) represents the nonlinear projections made on z n for the mth block. Assume U m , i | m = 1 M is a struct block with the length of A. Then, the shared struct layer output would be:
U i = [ Softmax ( U 1 , i ) , Softmax ( U 2 , i ) , , Softmax ( U A , i ) , ]
where U i R 1 × M A . Softmax is adopted for each block output to help maximize the potential of one element on each block, which would prevent approximation loss during encoding.

4.4. Hashing Learning

In fact, the process of converting from continuous variables to binary hash codes is a process of information loss, which can be regarded as the conversion of continuous variables to Hamming distances through lossy information channels. Channel capacity determines the amount of information that can be transferred from continuous variables. Through theoretical derivation, Li et al. [26] proved that when the input continuous variables obey the bi-half distribution, the channel capacity is the largest and the information loss in the hashing process is the smallest. Following its conclusion, we design the loss function in the hashing process:
L q = i = 1 K ( | u i | 1 ) 2 + | i = 1 K u i |
To incorporate the defined fine-grained similarity into hash learning, we use list-wise learning as the learning method. During the training, the model receives a set of n fabric images, X = x 1 , x 2 , , x n . x 1 is regarded as an anchor, and then the similarity matrix S = { S 11 , S 12 , S 1 n } can be obtained by using Equation (6). Based on the similarity matrix, we can obtain the true permutation of the input fabric images, which is represented by P = { P ( 1 ) , P ( 2 ) , P ( n ) } . Then probability of permutation P is expressed by:
P P = Π j = 1 n ϕ ( d P ( j ) ) k = j 5 ϕ ( d P ( k ) )
where d P ( j ) denotes the distance of image at position j of permutation P . Here, we set ϕ ( x ) = exp ( x ) . During training, the true permutation of the input images is known. Then, according to the idea of maximum likelihood estimation, by maximizing the log-likelihood corresponding to the true permutation (or equivalently minimizing the negative log-likelihood), the parameters of the model then can be optimized, and the loss function can be written as:
L P = log ( P P ) = log Π j = 1 n ϕ ( d P ( j ) ) k = j 5 ϕ ( d P ( k ) )
In summary, the total loss function of the model can be expressed as:
L = λ 1 L K L D + λ 2 L q + λ 3 L P
where L is the total loss, L P is the list-wise loss of feature U and L q is the quantitative loss. λ 1 , λ 2 and λ 3 are three weighting parameters to balance the effects of different loss functions. Each sub-item in the objective function is differentiable, so this optimization problem can be regarded as a convex optimization problem. Like other deep learning models, the stochastic gradient descent (SGD) algorithm and the backpropagation (BP) algorithm are used to optimize the parameters in the proposed model.

5. Experiments

5.1. Experimental Detail

As we all know, the CBIR method based on deep learning learns the image representation ability from a certain amount of training data. Therefore, data are the basis for model learning. In our previous studies [13,14], a dataset named WFID has been established for studying fabric image retrieval. WFID consists of 82,073 fabric images, all of which are annotated from four perspectives: coarse texture, fine texture, fabric style and the pattern forming method. In this study, we still used this dataset to demonstrate the effectiveness and superiority of the proposed method for fabric image retrieval. The difference is that only two versions of annotations were used in this study: coarse texture and fine texture. For fair comparison, the proposed method and compared learning-based methods were all trained on the training set with 33,645 fabric images; the performances of all methods were evaluated on the validation set. The validation set consists of 1029 sets of samples, each of which is an image.
The proposed DVSH was implemented by using the Pytorch toolkit. The hardware environment is as follows: CPU = E5 2623V4@2.60GHz, RAM = DDR4 32G, GPU = GeForce RTX 3090(24G) × 2. All compared deep learning-based methods were implemented using the Pytorch toolkit and based on the backbone of VGG-16. Additionally, the other methods, which are based on hand-crafted descriptors, were implemented by using MATLAB 2018b. During the training, the hyper-parameter configuration was as follows: batch_size = 32, weight_decay = 5 × 10 5 , optimizer = ADAM and learning_rate = 1 × 10 3 .
Generally, CBIR methods are evaluated based on precision, recall or the precision–recall curve. In addition, we also adopted m A P and N D C G to more comprehensively evaluate the performance of each method. We computed each evaluation metric by referring to reference [25].

5.2. Parameter Analysis

There are three parameters, λ 1 , λ 2 and λ 3 , to balance different objective functions. L P is the main driver to guide the learning of the model, so its corresponding weight parameter λ 3 was directly set to 1. λ 1 and λ 2 were, respectively, used to adjust the gradients of the corresponding parameters of the objective functions L K L D and L q during training. In this experiment, the influences of different weight parameter configurations on the final retrieval effect of the model were compared to optimize the optimal parameter configuration. We used the control variable method for analysis and comparison—that is, fixing one of the parameters to analyze and discussing the influence of another parameter on the performance of the model.
The retrieval performance comparison results of the models under different parameter configurations are shown in Table 2.
(1)
When the parameter λ 2 is fixed, the performance of the model under each evaluation index shows a similar law; that is, with the increase in λ 1 , the performance of the model shows a trend of first increasing and then decreasing. When its value is close to 1, the guiding effect of the corresponding objective function will be stronger than in other tasks; when its value is lower than 10 5 , the guiding effect of the corresponding objective function is too low and is covered by other tasks. Both cases will affect the feature selection ability of the variational network module, which will lead to the degradation of retrieval performance.
(2)
When λ 1 = 2 is fixed, and the value of λ 2 is 1 and 10 1 , each retrieval evaluation index of the model is very low, and the retrieval effect varies greatly in each coding length. This phenomenon shows that the guiding effect of this objective function completely covers other objective functions under this condition, which leads to the instability of the model performance. In particular, the retrieval performance of the model is the worst in Table 2, and it can be considered that the similarity of the fabric images is not embedded in the generated hash code. When the value of λ 2 is in the range of [ 10 5 , 10 2 ], the retrieval performance of the model is not very different, and all of them can achieve good performance. When λ 2 < 10 6 , the model performance has a certain decline. When the parameters take values in the interval [ 10 5 , 10 2 ], the sensitivity of the model performance is low, and the performance is relatively stable.
In contrast, the DVSH model is more sensitive to λ 2 , because the quantization loss directly affects the quality of the generated hash code. The experimental results show that when λ 1 = 10 3 and λ 2 = 10 3 , the model achieves the best performance, so this configuration was used in subsequent experiments.

5.3. Ablation Study

To gain insight into DVSH, we performed an ablation study. DVSH contains a variational network module, a structural network module and a similarity embedding module. Undoubtedly, the similarity embedding module is an indispensable part of the model, so this module was not analyzed in the ablation study. The other two modules of DVSH are very flexible and can be added or removed. Under the condition that the similarity embedding module is preserved, DVSH produces three variants: DVSH1a (removing the variational network module), DVSH1b (removing the structural network module) and DVSH2a (removing both modules). The comparative experimental results of DVSH and its three variants are shown in Table 3. It is clearly observed that removing any one of the modules in DVSH leads to a decrease in model performance. The performance of DVSH2a with both modules removed is significantly reduced. The experimental results demonstrate that the structural network module and the variational network module can jointly promote hash learning, so that the generated hash codes retain more useful information.

5.4. Time Complexity Analysis

An efficient retrieval system not only needs to output accurate retrieval results, but also must respond quickly to retrieval needs; that is, it must also have high timeliness. To demonstrate the efficiency of DVSH, we first compare the single image encoding times of different methods when the encoding length is 32 or 128 bits, including iterative quantization (ITQ) [27], ITQ+CNN [27], SuBiC [28] and DVSH2a (mentioned in the section of ablation study). Among them, ITQ and ITQ+CNN are two quantization-based dimensionality reduction methods, SuBic is a structured hashing method and DVSH2a is a variant of DVSH. The comparison results of the encoding time of the above methods are shown in Figure 8a, where “CNN Feat” represents the time when the convolutional neural network (Base Network) extracts features. As can be seen in the figure, the hash-based methods, SuBic and DVSH, consumed more time than CNN Feat because the encoding process takes a certain amount of time. Since more nodes will take more quantization time, the DVSH_128 encoding process takes more time than DVSH_32. However, the time consumption of high-dimensional feature extraction accounted for more than 70% of the total time-consuming, and the DVSH encoding time only accounted for less than 30%. DVSH is less than 10% more time-consuming than DVSH2a, indicating that the structural network module and the variational network module do not add too much computation.
Similarity measurement is another time-consuming link in CBIR. The hash method maps the similarity measure to the Hamming space, and then uses a simple and efficient XOR operation to calculate the similarity between binary codes. Quantization-based methods such as DVSQ [29] employ AQD [30] to compute the inner product between features to measure their similarity. SuBiC [28] computes asymmetric distances between features directly using the network output, thereby quickly computing similarities. In the experiment, the parameters of SuBiC and AQD were configured according to reference [28] and reference [29], respectively. The experimental results are shown in Figure 8b, which show that the query speed of AQD was seven times that of Hamming and SuBiC, and the time consumption of Hamming and and that of SuBiC were not very different. In retrieval, SuBiC only needs to perform a certain number of addition operations and use real-valued query codes. Although the time consumption is slightly lower than that of the Hamming, its calculation process takes up a lot of memory. AQD-lookup adopts the AQD of the code table, which uses the pre-calculated M × K code table, and directly uses the code table to query the distance between the two subspaces, which greatly shortens the time for quantizing the product. However, AQD-lookup needs to allocate memory on the code table of each query, increasing memory consumption, and the retrieval speed is lower than Hamming. In conclusion, the overall retrieval efficiency of DVSH is more suitable for efficient fabric image retrieval.

5.5. Comparisons

This section verifies the rationality and superiority of the proposed hashing model through horizontal comparison experiments. Some SOTA (state-of-the-art) deep hashing methods are compared, including: three hashing methods based on triplet loss supervised learning (DRSCH [31], DSRH [32], DNNH [33]), four supervised methods based on pairwise similarity loss (CNNH+ [34], DSH [35], DHN [11], IDHN [36]) and seven methods based on custom similarity guidance (CSQ [37], ISDH [38], DSHSD [39], SuBiC [28], MVL [40], UDHNR [41] and VTDH [42]. SuBiC is similar to DVSH, but it uses a sigmoid and threshold to binarize the output features to generate hash codes.
The comparison results of m A P and N D C G 50 of different hashing methods are shown in Table 4. It can be observed from the results that methods based on triplet loss are generally more stable than methods based on pairwise similarity loss, proving that methods based on triplet loss learn more similarity information. However, the DVSH proposed in this paper not only utilizes the discriminative classification of fabrics to guide hash learning, but also uses list-wise learning to embed more similarity information in the generated hash codes. List-wise learning is guided by the similarity of a batch of fabric images, and its learning objective is more stringent than pairwise and triplet losses. The variational network module and the structural network module are used in the DVSH model to reduce the learning difficulty, and the content of the model’s learning is closer to the ideal fabric similarity arrangement. DVSH far outperformed other methods in N D C G 50 , indicating that the hash code generated by DVSH contained more fabric fine-grain similarity information, and thus achieved a higher ranking score. As shown in Figure 9, the trends of PR curves for different code length models show consistency, illustrating the robust performance of the contrasting methods for fabric retrieval. The area under the PR curve corresponding to DVSH is higher than those of other methods at all encoding lengths, which verifies the superiority of the proposed method. In addition, all experimental results show that the longer the code length, the more fabric similarity information can be inherited, and thus the better the retrieval performance that can be obtained. It is worth noting that methods using soft similarity, such as IDHN and CSQ, achieved better retrieval results than other methods, indicating that more fine-grained similarity measurement methods can improve the retrieval performance of fabric images. Compared with the soft similarity, a more fine-grained measurement method was adopted, and more superior performance was obtained. The experimental results once again proved the validity of the fabric fine-grained similarity defined in this paper. Some retrieval samples are shown in Figure 10. The precise retrieval results show that DVSH has excellent performance for fabric image retrieval.

6. Conclusions

In this paper, a novel method for fabric image retrieval based on variational and structural hashing was proposed. To narrow the gap between fabric images and similarities, we designed a CNN with a compact structure and cross-shortcut connections, which is regarded as the base network of the hashing model. To overcome the problems of probabilistic missing and difficult training in classical hashing, we introduced a variational network module and structural network module into the hashing model (named DVSH). Then, a fine-grained similarity was defined to measure the similarity between two fabric images. To incorporate the defined fine-grained similarity into hash learning, we employed list-wise learning to complete similarity embeddings in mini-batches. The results of ablation experiments showed that the absence of any module will cause the performance of the model to decline, which verified the necessity and effectiveness of each module in DVSH. Through time complexity analysis, the single image encoding time of DVSH was only 16 milliseconds, which verified the real-time performance of the method. The retrieval performances of different hashing methods were compared, and DVSH achieved the best performance in different coding lengths, which verified the superiority of the method for fabric image retrieval. The method proposed in this paper has been successfully applied in cooperative enterprises.

Author Contributions

Conceptualization, R.P.; methodology, J.X.; software, J.X.; validation, J.X. and R.P.; formal analysis, J.X.; investigation, J.X.; resources, W.G.; data curation, J.X.; writing—original draft preparation, J.X.; writing—review and editing, R.P.; visualization, J.X.; supervision, W.G.; project administration, W.G.; funding acquisition, W.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under grant 61976105, in part by Textile Vision Basic Research Program under grant J202006.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, Y.; Zhang, J.; Chen, M.; Lei, H.; Luo, G.; Huang, Y. Shape based local affine invariant texture characteristics for fabric image retrieval. Multimed. Tools Appl. 2019, 78, 15433–15453. [Google Scholar] [CrossRef]
  2. Jing, J.; Li, Q.; Li, P.; Zhang, L. A new method of printed fabric image retrieval based on color moments and gist feature description. Text. Res. J. 2016, 86, 1137–1150. [Google Scholar] [CrossRef]
  3. Li, Y.; Luo, H.; Jiang, G.; Cong, H. Content-based lace fabric image retrieval system using texture and shape features. J. Text. Inst. 2019, 110, 911–915. [Google Scholar] [CrossRef]
  4. Ng, P.C.; Henikoff, S. SIFT: Predicting amino acid changes that affect protein function. Nucleic Acids Res. 2003, 31, 3812–3814. [Google Scholar] [CrossRef]
  5. Zhang, N.; Xiang, J.; Wang, L.; Gao, W.; Pan, R. Image retrieval of wool fabric. Part I: Based on low-level texture features. Text. Res. J. 2019, 89, 4195–4207. [Google Scholar] [CrossRef]
  6. Zhang, N.; Xiang, J.; Wang, L.; Xiong, N.; Gao, W.; Pan, R. Image retrieval of wool fabric. Part II: Based on low-level color features. Text. Res. J. 2020, 90, 797–808. [Google Scholar] [CrossRef]
  7. Wu, Y.; Lin, Y.; Dong, X.; Yan, Y.; Bian, W.; Yang, Y. Progressive learning for person re-identification with one example. IEEE Trans. Image Process. 2019, 28, 2872–2881. [Google Scholar] [CrossRef]
  8. Wu, Y.; Lin, Y.; Dong, X.; Yan, Y.; Ouyang, W.; Yang, Y. Exploit the unknown gradually: One-shot video-based person re-identification by stepwise learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–21 June 2018; pp. 5177–5186. [Google Scholar]
  9. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Harrahs and Harveys, Lake Tahoe, NV, USA, 3–8 December 2012. [Google Scholar]
  10. Erin Liong, V.; Lu, J.; Wang, G.; Moulin, P.; Zhou, J. Deep hashing for compact binary codes learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 8–10 June 2015; pp. 2475–2483. [Google Scholar]
  11. Zhu, H.; Long, M.; Wang, J.; Cao, Y. Deep hashing network for efficient similarity retrieval. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
  12. Venkateswara, H.; Eusebio, J.; Chakraborty, S.; Panchanathan, S. Deep hashing network for unsupervised domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5018–5027. [Google Scholar]
  13. Xiang, J.; Zhang, N.; Pan, R.; Gao, W. Fabric image retrieval system using hierarchical search based on deep convolutional neural network. IEEE Access 2019, 7, 35405–35417. [Google Scholar] [CrossRef]
  14. Xiang, J.; Zhang, N.; Pan, R.; Gao, W. Fabric retrieval based on multi-task learning. IEEE Trans. Image Process. 2020, 30, 1570–1582. [Google Scholar] [CrossRef]
  15. Fred, A.L.; Jain, A.K. Learning pairwise similarity for data clustering. In Proceedings of the 18th International Conference on Pattern Recognition (ICPR’06), Hong Kong, China, 20–24 August 2006; pp. 925–928. [Google Scholar]
  16. Yang, E.; Deng, C.; Liu, W.; Liu, X.; Tao, D.; Gao, X. Pairwise relationship guided deep hashing for cross-modal retrieval. In Proceedings of the AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
  17. Sidorov, G.; Gelbukh, A.; Gómez-Adorno, H.; Pinto, D. Soft similarity and soft cosine measure: Similarity of features in vector space model. Comput. Sist. 2016, 18, 491–504. [Google Scholar] [CrossRef]
  18. Moujahid, D.; Elharrouss, O.; Tairi, H. Visual object tracking via the local soft cosine similarity. Pattern Recognit. Lett. 2018, 110, 79–85. [Google Scholar] [CrossRef]
  19. Abdi, H.; Williams, L.J. Principal component analysis. Wiley Interdiscip. Rev. Comput. Stat. 2010, 2, 433–459. [Google Scholar] [CrossRef]
  20. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  21. Liong, V.E.; Lu, J.; Duan, L.Y.; Tan, Y.P. Deep variational and structural hashing. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 42, 580–595. [Google Scholar] [CrossRef]
  22. Wang, Z.; Turko, R.; Shaikh, O.; Park, H.; Das, N.; Hohman, F. CNN explainer: Learning convolutional neural networks with interactive visualization. IEEE Trans. Vis. Comput. Graph. 2020, 27, 1396–1406. [Google Scholar] [CrossRef]
  23. Qin, Z.; Yu, F.; Liu, C.; Chen, X. How convolutional neural network see the world-A survey of convolutional neural network visualization methods. arXiv 2018, arXiv:1804.11191. [Google Scholar] [CrossRef]
  24. Hänggi, M.; Moser, S.; Pfaffhauser, E.; Moschytz, G.S. Simulation and visualization of CNN dynamics. Int. J. Bifurc. Chaos 1999, 9, 1237–1261. [Google Scholar] [CrossRef]
  25. Xiang, J.; Zhang, N.; Pan, R.; Gao, W. Wool fabric image retrieval based on soft similarity and listwise learning. Text. Res. J. 2022. [Google Scholar] [CrossRef]
  26. Li, Y.; van Gemert, J. Deep unsupervised image hashing by maximizing bit entropy. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; pp. 2002–2010. [Google Scholar]
  27. Gong, Y.; Lazebnik, S.; Gordo, A.; Perronnin, F. Iterative quantization: A procrustean approach to learning binary codes for large-scale image retrieval. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 2916–2929. [Google Scholar] [CrossRef]
  28. Jain, H.; Zepeda, J.; Pérez, P.; Gribonval, R. Subic: A supervised, structured binary code for image search. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 833–842. [Google Scholar]
  29. Cao, Y.; Long, M.; Wang, J.; Liu, S. Deep visual-semantic quantization for efficient image retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1328–1337. [Google Scholar]
  30. Jiang, J.; Lai, Y.-C. Model-free prediction of spatiotemporal dynamical systems with recurrent neural networks: Role of network spectral radius. Phys. Rev. Res. 2019, 1, 033056. [Google Scholar] [CrossRef]
  31. Wang, X.; Hu, P.; Zhen, L.; Peng, D. Drsl: Deep relational similarity learning for cross-modal retrieval. Inf. Sci. 2021, 546, 298–311. [Google Scholar] [CrossRef]
  32. Wang, Y.; Ou, X.; Liang, J.; Sun, Z. Deep semantic reconstruction hashing for similarity retrieval. IEEE Trans. Circuits Syst. Video Technol. 2020, 31, 387–400. [Google Scholar] [CrossRef]
  33. Lai, H.; Pan, Y.; Liu, Y.; Yan, S. Simultaneous feature learning and hash coding with deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–15 June 2015; pp. 3270–3278. [Google Scholar]
  34. Xia, R.; Pan, Y.; Lai, H.; Liu, C.; Yan, S. Supervised hashing for image retrieval via image representation learning. In Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence, Québec City, QC, Canada, 27–31 July 2014; pp. 2156–2162. [Google Scholar]
  35. Liu, H.; Wang, R.; Shan, S.; Chen, X. Deep supervised hashing for fast image retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2064–2072. [Google Scholar]
  36. Zhang, Z.; Zou, Q.; Lin, Y.; Chen, L.; Wang, S. Improved deep hashing with soft pairwise similarity for multi-label image retrieval. IEEE Trans. Multimed. 2019, 22, 540–553. [Google Scholar] [CrossRef]
  37. Yuan, L.; Wang, T.; Zhang, X.; Tay, F.E.; Jie, Z.; Liu, W.; Feng, J. Central similarity quantization for efficient image and video retrieval. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 14–19 June 2020; pp. 3083–3092. [Google Scholar]
  38. Zhang, Z.; Zou, Q.; Wang, Q.; Lin, Y.; Li, Q. Instance similarity deep hashing for multi-label image retrieval. arXiv 2018, arXiv:1803.02987. [Google Scholar]
  39. Wu, L.; Ling, H.; Li, P.; Chen, J.; Fang, Y.; Zhou, F. Deep supervised hashing based on stable distribution. IEEE Access 2019, 7, 36489–36499. [Google Scholar] [CrossRef]
  40. Xiang, J.; Zhang, N.; Pan, R.; Gao, W. Efficient fine-texture image retrieval using deep multi-view hashing. Comput. Graph. 2021, 101, 93–105. [Google Scholar] [CrossRef]
  41. Wang, Y.; Song, J.; Zhou, K.; Liu, Y. Unsupervised deep hashing with node representation for image retrieval. Pattern Recognit. 2021, 112, 107785. [Google Scholar] [CrossRef]
  42. Li, T.; Zhang, Z.; Pei, L.; Gan, Y. HashFormer: Vision Transformer Based Deep Hashing for Image Retrieval. IEEE Signal Process. Lett. 2022, 29, 827–831. [Google Scholar] [CrossRef]
Figure 1. Three kinds of fabric image retrieval methods. (a) Real sample search; (b) text-based image retrieval; (c) search by image.
Figure 1. Three kinds of fabric image retrieval methods. (a) Real sample search; (b) text-based image retrieval; (c) search by image.
Entropy 24 01319 g001
Figure 2. Four different hash network architectures. (a) Typical hashing network; (b) variational network; (c) structural network; (d) variational and structural network.
Figure 2. Four different hash network architectures. (a) Typical hashing network; (b) variational network; (c) structural network; (d) variational and structural network.
Entropy 24 01319 g002
Figure 3. Details about the classification of fabric coarse texture in this study.
Figure 3. Details about the classification of fabric coarse texture in this study.
Entropy 24 01319 g003
Figure 4. Four examples of fabric images. (a,b) Small lattice; (c) window pane; (d) Jacquard (patterns in fabric images are more complex).
Figure 4. Four examples of fabric images. (a,b) Small lattice; (c) window pane; (d) Jacquard (patterns in fabric images are more complex).
Entropy 24 01319 g004
Figure 5. Fabric image color quantification. (a) Original fabric image; (b) percentages of various colors.
Figure 5. Fabric image color quantification. (a) Original fabric image; (b) percentages of various colors.
Entropy 24 01319 g005
Figure 6. Model structure of deep variational and structural hashing.
Figure 6. Model structure of deep variational and structural hashing.
Entropy 24 01319 g006
Figure 7. The architecture of the proposed convolution neural network: a compact structure and cross-circuit connections.
Figure 7. The architecture of the proposed convolution neural network: a compact structure and cross-circuit connections.
Entropy 24 01319 g007
Figure 8. Time complexity analysis. (a) Comparison of encoding time of a single fabric image. (b) Comparison of retrieval time of a single fabric image.
Figure 8. Time complexity analysis. (a) Comparison of encoding time of a single fabric image. (b) Comparison of retrieval time of a single fabric image.
Entropy 24 01319 g008
Figure 9. PR curves of different methods for fabric image retrieval (different code lengths): (a) 32 bits; (b) 64 bits; (c) 128 bits; (d) 256 bits.
Figure 9. PR curves of different methods for fabric image retrieval (different code lengths): (a) 32 bits; (b) 64 bits; (c) 128 bits; (d) 256 bits.
Entropy 24 01319 g009
Figure 10. Several examples of fabric image retrieval. (a) Inlaid stripes; (b) Dobby; (c) Houndstooth; (d) Printed fabric (The pattern on the fabric appears to be “printed”).
Figure 10. Several examples of fabric image retrieval. (a) Inlaid stripes; (b) Dobby; (c) Houndstooth; (d) Printed fabric (The pattern on the fabric appears to be “printed”).
Entropy 24 01319 g010
Table 1. Textile color classification standard.
Table 1. Textile color classification standard.
BlackGreyWhiteRedOrangeYellowGreenCyanBluePurple
h_min000015611263578100125
h_max1801801801018025347799124155
s_min00043434343434343
s_max2554330255255255255255255255
v_min04622146464646464646
v_max46220255255255255255255255255
Table 2. The retrieval performance of DVSH under different parameter configurations.
Table 2. The retrieval performance of DVSH under different parameter configurations.
λ 1 λ 2 mAP NDCG 50
32 Bits64 Bits128 Bits256 Bits32 Bits64 Bits128 Bits256 Bits
fix λ 2 10 6 10 3 0.6620.6720.6980.7110.6090.6120.6280.640
10 5 10 3 0.7040.7450.7610.7730.6480.6930.7080.696
10 4 10 3 0.7410.7950.8150.8240.6670.7160.7420.742
10 3 10 3 0.8250.8530.8760.8790.7430.7850.8060.800
10 2 10 3 0.7530.8070.8270.8350.6930.7510.7440.752
10 1 10 3 0.7150.7630.7840.7930.6580.6940.7210.737
1 10 3 0.6820.7110.7360.7420.6340.6400.6620.683
fix λ 1 10 3 10.4790.5240.6150.6490.4260.4660.5600.591
10 3 10 1 0.6710.7720.8240.8290.5900.6790.7250.738
10 3 10 2 0.7830.8340.8610.8680.6890.7590.7660.773
10 3 10 4 0.8190.8450.8740.8780.7210.7610.7950.799
10 3 10 5 0.8210.8420.8750.8760.7390.7580.7880.788
10 3 10 6 0.7810.8100.8240.8490.6870.7290.7420.764
The optimal value of each column of experimental results in the table is marked with black, and the worst value is marked with underline.
Table 3. Performance comparison of DVSH and its variants.
Table 3. Performance comparison of DVSH and its variants.
Name mAP NDCG 50
32 Bits64 Bits128 Bits256 Bits32 Bits64 Bits128 Bits256 Bits
DVSH1a0.8030.8250.8470.8520.7230.7430.7710.767
DVSH1b0.8140.8320.8540.8630.7330.7570.7860.777
DVSH2a0.7920.7990.8110.8240.7130.7350.7300.750
DVSH0.8250.8530.8760.8790.7430.7850.8060.800
The optimal value of each column of the experimental results in the table is marked with black, and the worst value is marked with an underline.
Table 4. Comparative results of different deep hashing methods for fabric image retrieval.
Table 4. Comparative results of different deep hashing methods for fabric image retrieval.
MethodsmAPNDCG50
32 Bits64 Bits128 Bits256 Bits32 Bits64 Bits128 Bits256 Bits
DRSCH [31]0.6970.7190.7380.7440.5990.6110.6490.647
DSRH [32]0.6890.7250.7390.7420.5930.6240.6280.653
DNNH [33]0.6920.7290.7440.7540.6020.6200.6400.648
CNNH+ [34]0.6610.6750.6910.7080.5750.5810.6010.616
DSH [35]0.6910.7220.7480.7610.6080.6360.6360.670
DHN [11]0.7080.7350.7450.7550.6230.6470.6560.664
IDHN [36]0.7360.7560.7840.7970.6330.6500.6820.693
CSQ [37]0.7420.7670.7820.8030.6530.6670.6880.707
ISDH [38]0.7350.7550.7750.7890.6470.6420.6590.671
DSHSD [39]0.7080.7260.7410.7550.6020.6170.6520.642
SuBiC [28]0.7020.7350.7640.7740.5970.6470.6650.673
MVL [40]0.7510.7710.7950.8090.6870.7120.7390.741
UDHNR [41]0.7680.7860.8140.8220.6920.7170.7320.735
VTDH [42]0.7590.7850.8010.8140.6840.7150.7270.733
DVSH0.8250.8530.8760.8790.7430.7850.8060.800
The optimal value of each column of experimental results in the table is marked with black, and the worst value is marked with an underline.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xiang, J.; Pan, R.; Gao, W. An Efficient Retrieval System Framework for Fabrics Based on Fine-Grained Similarity. Entropy 2022, 24, 1319. https://doi.org/10.3390/e24091319

AMA Style

Xiang J, Pan R, Gao W. An Efficient Retrieval System Framework for Fabrics Based on Fine-Grained Similarity. Entropy. 2022; 24(9):1319. https://doi.org/10.3390/e24091319

Chicago/Turabian Style

Xiang, Jun, Ruru Pan, and Weidong Gao. 2022. "An Efficient Retrieval System Framework for Fabrics Based on Fine-Grained Similarity" Entropy 24, no. 9: 1319. https://doi.org/10.3390/e24091319

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop