Next Article in Journal
Entropy and Information Inequalities
Next Article in Special Issue
Deep Learning for Stock Market Prediction
Previous Article in Journal
Augmentation of Dispersion Entropy for Handling Missing and Outlier Samples in Physiological Signal Monitoring
Previous Article in Special Issue
Emergence of Network Motifs in Deep Neural Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Relative Distribution Entropy Loss Function in CNN Image Retrieval

1
College of Computer Science and Technology, Jilin University, Changchun 130012, China
2
Key Laboratory of Symbolic Computation and Knowledge Engineering of Ministry of Education, Jilin University, Changchun 130012, China
3
School of Mechanical Science and Engineering, Jilin University, Changchun 130025, China
4
College of Software, Jilin University, Changchun 130012, China
5
College of Communication Engineering, Jilin University, Changchun 130012, China
*
Author to whom correspondence should be addressed.
Submission received: 31 January 2020 / Revised: 5 March 2020 / Accepted: 9 March 2020 / Published: 11 March 2020
(This article belongs to the Special Issue Information Transfer in Multilayer/Deep Architectures)

Abstract

:
Convolutional neural networks (CNN) is the most mainstream solution in the field of image retrieval. Deep metric learning is introduced into the field of image retrieval, focusing on the construction of pair-based loss function. However, most pair-based loss functions of metric learning merely take common vector similarity (such as Euclidean distance) of the final image descriptors into consideration, while neglecting other distribution characters of these descriptors. In this work, we propose relative distribution entropy (RDE) to describe the internal distribution attributes of image descriptors. We combine relative distribution entropy with the Euclidean distance to obtain the relative distribution entropy weighted distance (RDE-distance). Moreover, the RDE-distance is fused with the contrastive loss and triplet loss to build the relative distributed entropy loss functions. The experimental results demonstrate that our method attains the state-of-the-art performance on most image retrieval benchmarks.

1. Introduction

In recent years, the newly proposed image retrieval algorithms based on convolutional neural networks [1,2,3,4,5] (CNN) have greatly improved the accuracy and efficiency. This has been the mainstream direction of academic research of image retrieval. In the beginning, CNN could only be applied to image classification tasks [6,7,8]. However, image classification is different from the image retrieval. Krizhevsky [9] flexibly applied a convolutional neural network to image retrieval. AlexNet [9] is designed for image classification and retrieval. Subsequently, Noh [10] proposed a Local Feature descriptor, called DELF (Deep Local Feature), which is suitable for large-scale image retrieval. A large number of studies have illustrated that the output features of the convolutional layer of the neural network have excellent discrimination and scalability. More recently, image retrieval algorithms based on convolutional neural networks emerged one after another. These methods are mainly summarized into three categories: fine-tuned networks, pre-trained networks, and hybrid networks. Among them, hybrid networks are less efficient in image retrieval tasks, and pre-trained networks are widely used. The fine-tuned network initializes the network architecture through the pre-trained image classification model and then adjusts the parameters according to different retrieval tasks. The fine-tuned network usually optimizes the network parameters by training the network architecture of metric learning. Metric learning aims to learn an embedding space, where the embedded vectors of positive samples are encouraged to be closer, while negative samples are pushed apart from each other [11,12,13]. Recently, a lot of deep metric learning methods have been based on pairs of samples such as contrastive loss [14], triplet loss [15], quadruplet loss [16], lifted structured loss [17], N-pairs loss [18], binomial deviance loss [19], histogram loss [20], angular loss [21], distance weighted margin-based loss [22], and hierarchical triplet loss (HTL) [23]. Most of the above-mentioned loss functions take common vector similarity (such as Euclidean distance) as the final image descriptor into consideration. However, it is not accurate enough to measure the similarity between features only by Euclidean distance, which lacks the difference in the internal spatial distribution [24] of the image pair. As illustrated in Figure 1, each rectangle represents a feature descriptor obtained after the convolution of the neural network. The value of Euclidean distance between descriptors for different images may be the equal or small, but the spatial distribution of every descriptor may be greatly different. In the information processing field, entropy [25] is an effective measurement to reflect the distribution information. Relative entropy [26,27,28] is a measure of the distance between two random distributions, which is equivalent to the difference of information entropy between two distributions. Inspired by this, we introduce the idea of relative entropy into image retrieval.
To solve the key problem mentioned above, we propose relative distribution entropy (RDE) to describe the distribution attributes of image descriptors. We combine the relative distribution entropy with the Euclidean distance [29] to build the relative distribution entropy weighted distance (RDE-distance). We fuse the RDE-distance into the contrastive loss and triplet loss to obtain the relative distributed entropy contrastive loss and the relative distributed entropy triplet loss. We call them the relative distributed entropy loss functions. To be more specific, we make three contributions, as follows:
Firstly, we propose a loss function modified by relative distribution entropy, called the relative distribution entropy loss function. Furthermore, the relative distribution entropy loss function contains the following two aspects: (1) Euclidean distance between the descriptors. (2) The differences in internal distribution state between descriptors. The core idea of the algorithm is illustrated in Figure 2. We combine the Euclidean distance with the relative distribution entropy to obtain the relative distribution entropy weighted distance (RDE-distance), which increases the discrimination between the image pair. We replace the Euclidean distance in the original contrastive loss and triplet loss with the relative distribution entropy weighted distance (RDE-distance) to obtain the relative distribution entropy loss function.
Secondly, during the experiment, we use GeM pooling [30] and whitening post-processing [31].
Thirdly, we employ the fine-tuned network [30] to perform our experiments on different datasets to verify the effectiveness of our proposed method.
The organization of this work is as follows. Section 2 introduces the related work. Section 3 mainly introduces our proposed relative distribution entropy loss function. Specific experimental results and analyses will be presented in Section 4. Section 5 is a summary of the contributions and methods of this paper.

2. Related Work

2.1. Deep Metric Learning

Deep metric learning (DML) has become one of the most interesting research areas in machine learning. Metric learning aims to learn an embedding space. In this space, images are converted into embedding vectors. The distance between embedding vectors of positive samples is small, and the distance between embedding vectors of negative samples is large [11,12,13]. Now, deep metric learning has played a vital role in many areas, such as face recognition [32], pedestrian recognition [33], fine-grained retrieval [34], image retrieval [35], target tracking [36], and multimedia retrieval [37]. We will summarize the recent emergence of metric learning methods in the next section. Here, we introduce deep metric learning into image retrieval. Deep metric learning aims to learn a discriminative feature embedding f ( x ) for input image x . In other words, f ( x ) is the descriptor of the image. Formally, we define the Euclidean distance between two descriptors as D x i , x j = f ( x i ) f ( x j )   2 .

2.1.1. Contrastive Loss

The siamese network [14] is a typical pair-based method. Its embedding is obtained through contrastive loss. The purpose of contrastive loss is to reduce the Euclidean distance between positive samples and increase the Euclidean distance between negative samples. The equation of the contrastive loss function is illustrated in (1).
L ( x i , x j ) = γ D x i , x j 2 + ( 1 γ ) m a x ( 0 , m D x i , x j ) 2
γ = 1 if the sample x j is a positive sample, and γ = 0 if the sample x j is a negative sample. m is the margin. This keeps negative distances above a certain threshold.

2.1.2. Triplet Loss

However, the above models only focus on the similarity of intra-class of the samples. To solve this problem, the triplet loss function [15] is proposed. Each triplet comprises a positive sample and a negative sample sharing the query. Triplet loss aims to learn an embedding space. In this space, the distance between the query and the negative sample is greater than the distance between the query and the positive sample. The triplet loss function is illustrated in (2).
L ( x , x + , x ) = m a x ( 0 , D x , x + 2 D x , x 2 + m )
Here, D x , x + = f ( x ) f ( x + ) 2 represents the Euclidean distance between the descriptors of the positive sample and the query. Similarly, D x , x = f ( x ) f ( x ) 2 represents the Euclidean distance between the descriptors of the negative sample and the query. m is the violate margin that requires that the negative distances to be larger than the positive distances.

2.1.3. N-pair Loss

Triplet loss function only compares one negative sample and ignores the negative samples of other classes during the learning stage. As a result, the embedding vector of the query can only be promoted to maintain a large distance between the selected negative samples, but it cannot guarantee to maintain a large distance between the embedding vector and other non-selected negative samples.
N-pair loss [18] has improved the above problems. Unlike triplet loss function, N-pair loss considers the relationship between query samples and other negative samples of different classes within a mini-batch. The equation of the N-pair loss function is illustrated in (3).
L ( x , x + , { x } i = 1 N 1 ) = l o g ( 1 + i = 1 N 1 e x p ( f ( x ) T f ( x i ) f ( x ) T f ( x + ) ) )
Each training tuple of N-pair loss is composed of N + 1 samples: { x , x + , x 1 , x 2 , x N 1 } . x + is the positive sample. { x } i = 1 N 1 are the negative samples.

2.1.4. Lifted Structured Loss

The existing triplet loss methods cannot take full advantage of mini-batch SGD training. Lifted structured loss calculates loss [17] based on all positive and negative sample pairs in the training set (mini-batch). The lifted structured loss function is illustrated in Equation (4).
L = m a x ( 0 , l o g ( ( x i , x j ) ϵ P e x p { m D x i , x j } ) + l o g ( ( x i , x j ) ϵ N e x p { D x i , x j } ) )
where P   is the set of positive samples in the training set, and N is the set of positive samples in the training set. m is the violate margin.
Although these loss functions calculate the distance of descriptors between images, they neglect the difference in internal distribution between the image pair. In this work, we propose a new loss function, which is called the relative distribution entropy loss function. We use the relative distribution entropy to reflect the difference in the descriptor distribution, and add this difference to the loss function. In this way, the new loss function can combine the Euclidean distance with the internal relative difference in the distribution state between the descriptors of the image pair.

2.2. Application of Spatial Information in Image Retrieval

The performance of image retrieval has been greatly improved in recent years through the use of deep feature representations. However, most existing methods aim to retrieve images that are visually similar or semantically relevant to the query, without considering the spatial information. Before that, some researchers attempted to add spatial information into image retrieval algorithms to improve retrieval performance. Mehmood [38] proposed adding a local region and a global histogram to the BoW algorithm and combining them as the final descriptor. Krapac [39] used the BoW descriptor to encode the spatial part of the image, which improved the performance of the image classification. Koniusz [40] used spatial coordinate coding to represent and simplify the spatial pyramid to provide more compact image features. Sanchezet [41] added the spatial position information of features into the descriptor, which overcame the change in the proportion of retrieved objects and the change in the local area of the image. Liu [42] introduced the concept of spatial distribution entropy and combined it with the VLAD algorithm. These methods have fully introduced the spatial information into the image retrieval algorithm, and great effects are obtained. However, these methods all aimed at the improvement of traditional image descriptors. In this work, we attempt to add spatial information to the descriptor by optimizing the loss function of a deep convolutional network.

2.3. Pooling and Normalization

The feature map generated by the deep retrieval framework reflects the color, texture, shape, and other characteristics of the image. Since the convolutional neural network needs to be integrated into a multi-dimensional feature descriptor before retrieval, the feature map of the convolutional layer needs to be further processed. It is required that the processed result retains the main features of the image while reducing the parameters and calculations of the next layer. Babenko [43] proposed the mean-pooling method, which sums the pixel values of the feature map to obtain an N-dimensional feature vector. Razavian [44] used max-pooling or mean-pooling on the feature descriptors and the result reduced MAC [44] descriptors in dimension with a series of normalization and PCA [45] whitening operations. Finally, the region feature vectors are summed to obtain a single image representation. In this work, we used the generalized mean-pooling [30]. We used Z to represent the input to the pooling layer and h to represent the pooling layer output. The mentioned pooling can be expressed as follows:
Max pooling (MAC vector):
h ( m ) = [ h 1 ( m ) h k ( m ) h K ( m ) ] T , h k ( m ) = m a x z Z k   z
Average pooling (SPoC vector [43]):
h ( a ) = [ h 1 ( a ) h k ( a ) h K ( a ) ] T , h k ( a ) = 1 | Z k | z Z k z
Generalized mean pooling (GeM [30]):
h ( g ) = [ h 1 ( g ) h k ( g ) h K ( g ) ] T , h k ( g ) = ( 1 | Z k | z Z k z p k ) 1 p k
where K is the number of feature maps, k means the channel of features and | Z k | is the number of feature values in the k -th channel feature map. The descriptor finally consists of a single value per feature map.   p k is the pooling parameter, which can be manually set or learned. The superscript of h k ( g )   is the pooling method. m , a , g represent the max pooling, average pooling and generalized mean pooling, respectively.
The max-pooling and mean-pooling are special cases of generalized mean-pooling. When   p k = 1 , it is the mean-pooling. On the contrary, when it is positive infinity, it is the max-pooling. Generalized mean-pooling with parameters can better adapt to the network and improve retrieval performance.
In this work, we use L 2 normalization to balance the effect of the range of pixel values as follows:
V L 2 = v i v
where v represents a vector, v represents the norm of the vector, and v i represents the value of the dimension on the vector.

2.4. Whitening

In large-scale image retrieval applications, high-dimensional global images typically require the use of PCA to reduce the dimensions of features for the next step. Jegou and Chum [45] studied the influence of PCA on the BoW and VLAD descriptor representations, and highlighted the use of multiple visual dictionaries for dimensionality reduction, thereby reducing the information loss of the dimensionality reduction process. In this work, whitening is used as a post-processing step. In this paper, the method of whitening is the linear discriminant projection proposed by Mikolajczyk and Matas [31].
The processing steps are divided into two parts. In the first part, the intra-class image feature vector is whitened. The whitening part is the reciprocal of the square root of the intra-class image pair (matched image pair) covariance matrix C S 1 2 .
C S = Y ( x i , x j ) = 1 ( f ( x i ) f ( x j ) ) ( f ( x i ) f ( x j ) ) T
where f ( x i ) and f ( x j ) are the descriptors of the image after pooling, Y ( x i , x j ) = 1 represents the image matched pair, and C s represents the covariance matrix of the matched image pair.
In the second part, the inter-class image feature is rotated. The rotating part is the eigenvector of the covariance matrix C S 1 2 C D C S 1 2 between inter-class image pair (the non-matched image pair)
C D = Y ( x i , x j ) = 0 ( f ( x i ) f ( x j ) ) ( f ( x i ) f ( x j ) ) T
where   f ( x i ) and f ( x j ) are the descriptors of the image after pooling, Y ( x i , x j ) = 0 represents the non-matched image pair, and C D represents the covariance matrix of the non-matched image pair.
Then, we apply the projection P = C S 1 2 e i g ( C S 1 2 C D C S 1 2 ) to P T ( f ( x i ) μ ) , where μ is the mean GeM [30] vector to perform centering.

3. Method Overview

3.1. Calculation of Relative Distribution Entropy

We firstly introduce the concept of relative distribution entropy. Relative distribution entropy can better represent the distribution difference between two descriptors of image samples. The relative distribution entropy is derived from the relative entropy. Relative entropy can be computed as follows:
R ( P Q ) = P ( x ) l o g P ( x ) Q ( x )
P ( x ) and Q ( x ) are the two probability distributions on the random variable X . From this, we can get the equation of relative distribution entropy (RDE).
P ( f ( x i ) m ) = h i s t o g r a m ( f ( x i ) ) [ m ] d i m
R D E ( x i x j ) = m = 1 n ( f ( x i ) m ) l o g P ( f ( x i ) m ) Q ( f ( x j ) m )
where x i , x j represent the images. f ( x i ) represents the descriptor of image x i . It is a normalized vector. R D E ( x i x j ) is the relative distribution entropy of two images. In this work, we use histograms to describe the distribution of image descriptors. n is the number of bins, and is an adjustable parameter. P ( f ( x i ) m )   is the probability distribution in the m t h bin. The equation for Q ( f ( x j ) m ) is the same as for   P ( f ( x i ) m ) . d i m is the dimension of descriptors.

3.2. Relative Distribution Entropy Loss Function

From the above, we introduce the calculation of relative distributed entropy. Next, we show how to add the relative distribution entropy into the loss function. We add the relative distribution entropy to contrastive loss [14] and triplet loss [15] to build relative distribution entropy loss functions.
Firstly, we introduce the fusion process of contrastive loss and relative distribution entropy. The equation of the contrastive loss function is shown in (14).
L ( x i , x j ) = γ D W ( x i , x j ) 2 + ( 1 γ ) m a x ( 0 , m D W ( x i , x j ) ) 2
D W ( x i , x j ) = f ( x i ) f ( x j ) 2
where D W ( x i , x j ) represents the Euclidean distance between the descriptors of the query x i and the sample x j . γ = 1 if the sample x j is a positive sample, and   γ = 0 if the sample x j is a negative sample. m is the margin.
Then, we add the relative distribution entropy to D W to get the D N W :
D N W ( x i , x j ) = f ( x i ) f ( x j ) 2 + α R D E ( x i x j )
where α is the weighting parameter of the relative distribution entropy. As illustrated in Equation (16), we get the new distance metric. We call it the relative distribution entropy weighted distance (RDE-distance). We substitute D N W into (14) to get the new contrastive loss function, as is shown in Equation (17):
L ( x i , x j ) = γ D N W ( x i , x j ) 2 + ( 1 γ ) m a x ( 0 , m D N W ( x i , x j ) ) 2
Similarly, we introduce the fusion process of triplet loss and relative distribution entropy. The equation of the contrastive loss function is shown in (18).
L ( x , x + , x ) = m a x ( 0 , D P ( x , x + ) 2 D N ( x , x ) 2 + m )
D P ( x , x + ) is the Euclidean distance between the descriptors of the positive sample and the query. Similarly, D N ( x , x ) is the Euclidean distance between the descriptors of the negative sample and the query.
Then, we add the relative distribution entropy to D P , D N to get the   D N P ,   D N N .
D N P ( x , x + ) = f ( x ) f ( x + ) 2 + α R D E ( x x + )
D N N ( x , x ) = f ( x ) f ( x ) 2 + α R D E ( x x )
We substitute the new Euclidean distances into Equation (18) and get the new triplet loss.
L ( x , x + , x ) = m a x ( 0 , D N P ( x , x + ) 2 D N N ( x , x ) 2 + m )

3.3. Network Architecture for Relative Distribution Entropy Loss Function

3.3.1. CNN Network Architecture

We construct a CNN neural network to obtain the descriptor of the image. We only use the convolutional layers, discarding the fully connected layer. Convolution layers can extract features of images. The feature map obtained by the convolution layer is vectorized by the GeM pooling operation [30]. If the whitening operation is performed, the whitening is processed following the pooling layer. Whitening can reduce the correlation between features, and it can make the features share the same variance (covariance matrix is 1), which can greatly improve the image retrieval performance. Here, we use the L w   whitening [31] method. The last step is the normalization operation. The purpose of normalization is to make the preprocessed data limited to a certain range (such as [0,1] or [−1,1]), thereby eliminating the adverse effects caused by the singular sample data. The network architecture is pre-trained in ImageNet [46] network architecture.
Furthermore, we adopt network architectures such as ResNet [47] and AlexNet [9], and these two networks are also pre-trained on ImageNet [46].
The CNN network architecture is shown in Figure 3.

3.3.2. Architecture of Training

The training procedure consists of multiple networks sharing the same weight. The architecture of CNN is introduced in the previous part. We added our newly proposed relative distribution entropy to the previous loss function, as shown in Figure 4 and Figure 5:
In Figure 4 and Figure 5, D represents the Euclidean distance between the two descriptors, and RDE represents the relative distribution entropy between the two descriptors. RDE-distance is the relative distribution entropy weighted distance obtained after the fusion of D and RDE.

4. Experiments and Evaluation

In this section, we discuss the implementation details of training and testing. Also, we analyze the experimental results and compare them with previous work.

4.1. Training Datasets

In this work, experimental training data are distilled from Retrieval-SFM-120K [48], which contains 7.4 million images. After clustering [49], we get about 20,000 images as the query seed. The structure-from-motion (SfM) algorithm constructs 1474 3D models from the training datasets. We removed the duplications and retained 713 of them, which contained more than 163,000 different images.
There are 91,642 training images in the dataset, and 98 cluster images that are the same or almost the same as the test dataset. Through the minimum hash and spatial verification methods mentioned in the clustering process, about 20,000 images are selected as query images, 18,1697 pairs of positive images and 551 training clusters, including more than 163,000 clusters [50] from the original dataset. The dataset contains all images from the Oxford 5k [51] and Paris 6k [52] datasets.

4.2. Training Configurations

In the experiments, we use the Pytorch deep learning framework to train the deep network model. We use ResNet [47], VGG [53] and AlexNet [9], which are all pre-trained on ImageNet [46].
In the experiment of relative distribution entropy contrastive loss, ResNet [47] and VGG [53] are trained using Adam learning strategy [54], while AlexNet [9] is trained using SGD. Our initial learning rate for Adam is 10 6 , and the margin for ResNet and VGG are 0.95 and 0.9. We use an initial learning rate equal to 10 3 for SGD, and the margin for AlexNet is 0.75.
In the experiment of triplet loss, we also use ResNet [47], VGG [53], and AlexNet [9] to initialize the network. They are trained using the Adam learning strategy [54]. Our initial learning rate for Adam is 10 6 . The margin for ResNet and VGG are 0.5, and the margin for AlexNet is 0.3. The size of the training image is not more than 362 * 362 while maintaining the aspect ratio of the original image.
The experimental environment is an intel(R) i7-8700 processor, GPU with 12GB of memory, NVIDIA(R) 2080Ti graphics card, driver version 419.**. Operating system is Ubuntu 18.04 LTS, PyTorch version v1.0.0, CUDA version 10.0, CUDNN version 7.5. The time spent in each training cycle trained on our method on VGG, ResNet, and AlexNet is 0.48, 0.72, and 0.22 hours, respectively. During the testing phase, testing VGG, ResNet, and AlexNet networks takes 620, 990, and 277 seconds, respectively. There are subtle differences between different test sets. With the same computing power, the training time of our method is almost the same as that of other methods [30].

4.3. Datasets and Evaluation of Image Retrieval

We conduct our testing experiments on the following benchmark datasets frequently. Herein, we give the details of these datasets.
Oxford5k [51] is a widely used landmark dataset consisting of 5062 building images from the Flickr dataset. It contains 11 famous landmarks in the Oxford area, and each landmark building has 55 query images.
Paris6k [52] contains 6392 images and is also one of the widely used datasets in the field of image retrieval. It collects many landmark buildings in Paris, and most of these images are from tourists. Similar to the Oxford 5k dataset, it also has 55 query images.
In addition, we use 100k interference images to fuse with the Oxford5k and Paris6k datasets to obtain Oxford105k [51] and Paris106k [52].
In the experiments, we use mean average precision (mAP) to measure the performance of image retrieval.

4.4. Results and Analysis

4.4.1. The Adjustment Process of Hyperparameter

In this experiment, two hyperparameters α and β are adjusted to obtain best performance. α is the weight of the relative distribution entropy. As mentioned in Section 3.2, when fusing the Euclidean distance with RDE, the weight α will affect the ratio of Euclidean distance and relative distribution entropy to the finally generated RDE-distance. As mentioned in Section 3.1, β is the total bin amount in the histogram during the calculation of RDE. β will affect the degree of differentiation of internal spatial distribution differences between two descriptors. However, a large β also increases the computational burden. The ability of RDE-distance to distinguish between two descriptors is determined by two factors α and β . The values of these two hyperparameters have a great impact on our experimental results. During our experiments, we adjust them to get the best performance. We use AlexNet and VGG networks for tuning and the GeM pooling [30]. The partially representative results are shown in Table 1.
Here, we take some representative results. In the relative distribution entropy contrastive loss, we take the value of α within 0.5–1, and the results show that the performance is the best when α = 0.9 . Additionally, we take the value of β within 10–100. When the value of β is large ( β > 50 ), the effect of adding relative distribution entropy is not obvious. After a large number of experiments, we can make the following conclusions. When α = 0.90 and β = 30 , our performance achieve the best results on Oxford5k and Pairs6k. The performance achieves 88.00% and 88.12% on VGG. On AlexNet, the performance achieves 68.22% and 80.07%. Therefore, we set β to 30 and α to 0.9 in ResNet and VGG as our final hyperparameters.

4.4.2. Comparison of MAC, SPoC, and GeM

In this section, we combine the relative distribution entropy contrastive loss function with the current most advanced pooling methods, GeM [30], MAC [44] and SPoc [43], for end-to-end training. In this experiment, we use the AlexNet for training. The experimental results are shown in Table 2.
The conclusions drawn from Table 2 are as follows. The results in the table indicate that the experiment results using GeM pooling [30] on AlexNet are superior to the other two pooling methods. We get the results of 60.79%, 68.22%, 75.29% and 80.07%, which are the maximum values on the different datasets. In the next experiments, we will use the GeM pooling [30] method for training.

4.4.3. Comparison of Relative Distribution Entropy Triplet Loss and Triplet Loss

In this section, we present the experimental results of our method in triplet loss and compare them with the previous method [30]. We perform comparison tests on VGG and ResNet. Using the same pooling method, experimental steps, and network model, we compare the relative distribution entropy triplet loss with the traditional triplet loss. The comparison results are shown in Table 3.
From Table 3, the result indicates that when we experiment on VGG, our proposed method obtains the best performance on all these datasets, with 82.39%, 83.07%, 83.61%, and 85.45%. The same conclusion is obtained when we perform the experiments on ResNet. We get the results of 82.88%, 86.54%, 89.33%, and 91.97%, which is the best performance amongst the datasets. In this experience, we combine the relative entropy with the Euclidean distance into relative distributed entropy weighted distance, which is a new metric. We put this new metric into the triplet loss function, and experiments have proven that our method is greatly effective.

4.4.4. Comparison with State-of-Art

In this section, we compare the relative distribution entropy contrastive loss with the latest methods. The performance comparison is shown in Table 4. The results of other methods are given by referring to the results in their papers. From Table 4, it can be learned that our proposed method attains better performance on multiple datasets. As shown in the table, we divide the existing networks into two categories: (1) using fine-tuning networks (yes) and (2) not using fine-tuning networks (no). When using the VGG network, compared with the RMAC [55], relative distribution entropy contrastive loss provides a significant improvement of +4.9% and +1.0% on the Oxford5k and Paris6k datasets, respectively. Compared to the latest release, our method also has performance improvements. When using ResNet, our experimental results achieve +0.6% growth compared to GeM [30] on Oxford 5k. Our method also shows superior performance on large-scale datasets. When using the VGG network, our experimental results achieved +0.2% growth compared to GeM [30] on Oxford 105k. When using the ResNet network, our experimental results achieve +0.3% growth compared to GeM [30] on Oxford105k. Our method shows more obvious performance improvements after adding re-ranking and query expansion. Under the VGG, the gain over GeM + αQE [30] is +0.1% and +0.5% on the Paris 6k dataset, respectively. Under the ResNet, the our method achieved mAP of 91.7%, 89.7%, 96.0%, and 92.1% and offered over 91.0%, 89.5%, 96.7%, and 91.9% gain over the GeM+αQE [37] on Oxford 5k, Oxford 105k, Paris 6k and Paris 106k datasets, respectively.

5. Conclusions

In this paper, we discuss the deficiency of traditional loss functions in spatial distribution differences. To make up for the lack of spatial distribution differences in the descriptors of image pair, the concept of relative distribution entropy (RDE) is presented. The calculation process of the relative distribution entropy is introduced in Section 3. Next, we combine Euclidean distance and relative distribution entropy to obtain a new similarity measurement, called relative distribution entropy weighted distance (RDE-distance). We combine RDE-distance with contrastive loss and triplet loss to obtain relative distribution entropy contrastive loss and relative distribution entropy triplet loss. We train the entire framework in an end-to-end manner, and the results of extensive experiments prove that our new method achieves the state-of-the-art performance.
Our method mainly focuses on how to fuse Euclidean distance and spatial information. Here, we introduce relative distribution entropy to describe spatial information. We would like to focus on other fusion methods instead of the existing linear fusion. In addition, we would concentrate on adding our relative distribution entropy to other loss functions in future work.

Author Contributions

Methodology, P.L.; Project administration, P.L.; Software, L.S., Z.M., and B.J.; Writing—original draft, L.S.; Writing—review & editing, P.L., B.J., and Q.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Nature Science Foundation of China, under Grants 61841602, General Financial Grant from China Postdoctoral Science Foundation, under Grants 2015M571363 and 2015M570272, the Provincial Science and Technology Innovation Special Fund Project of Jilin Province, under Grant 20190302026GX, the Jilin Province Development and Reform Commission Industrial Technology Research and Development Project, under Grant 2019C054-4, and the State Key Laboratory of Applied Optics Open Fund Project, under Grant20173660.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. He, K.; Zhang, X.; Ren, S.; Sun, J. Intelligence, m. Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Yue-Hei Ng, J.; Yang, F.; Davis, L.S. Exploiting local features from deep networks for image retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 53–61. [Google Scholar]
  3. Arandjelovic, R.; Gronat, P.; Torii, A.; Pajdla, T.; Sivic, J. NetVLAD: CNN architecture for weakly supervised place recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NA, USA, 27–30 June 2016; pp. 5297–5307. [Google Scholar] [CrossRef] [Green Version]
  4. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  5. Alzu’bi, A.; Amira, A.; Ramzan, N.J.N. Content-based image retrieval with compact deep convolutional features. Neurocomputing 2017, 249, 95–105. [Google Scholar] [CrossRef] [Green Version]
  6. Howard, A.G. Some improvements on deep convolutional neural network based image classification. arXiv 2013, arXiv:1312.5402. [Google Scholar]
  7. Cireşan, D.; Meier, U.; Masci, J.; Schmidhuber, J. A committee of neural networks for traffic sign classification. In Proceedings of the 2011 International Joint Conference on Neural Networks (IJCNN), San Jose, CA, USA, 31 July–5 August 2011; pp. 1918–1921. [Google Scholar]
  8. Ciresan, D.C.; Meier, U.; Masci, J.; Gambardella, L.M.; Schmidhuber, J. Flexible, high performance convolutional neural networks for image classification. In Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence, Barcelona, Spain, 16–22 July 2011. [Google Scholar]
  9. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems, Lake Tahoe, CA, USA, 3–6 December 2012; pp. 1097–1105. [Google Scholar]
  10. Noh, H.; Araujo, A.; Sim, J.; Weyand, T.; Han, B. Large-scale image retrieval with attentive deep local features. In Proceedings of the IEEE International Conference on Computer Vision(CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3456–3465. [Google Scholar]
  11. Lowe, D.G. Similarity metric learning for a variable-kernel classifier. Neural Comput. 1995, 7, 72–85. [Google Scholar] [CrossRef]
  12. Mika, S.; Ratsch, G.; Weston, J.; Scholkopf, B.; Mullers, K.-R. Fisher discriminant analysis with kernels. In Proceedings of the Neural Networks for Signal Processing IX: Proceedings of the 1999 IEEE Signal Processing Society Workshop (cat. no. 98th8468), Madison, WI, USA, 25 August 1999; pp. 41–48. [Google Scholar]
  13. Xing, E.P.; Jordan, M.I.; Russell, S.J.; Ng, A.Y. Distance metric learning with application to clustering with side-information. In Proceedings of the Advances in neural information processing systems, Vancouver, BC, Canada, 8–13 December 2003; pp. 521–528. [Google Scholar]
  14. Hadsell, R.; Chopra, S.; LeCun, Y. Dimensionality reduction by learning an invariant mapping. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17 June 2006; Volume 2, pp. 1735–1742. [Google Scholar]
  15. Hoffer, E.; Ailon, N. Deep metric learning using triplet network. In Proceedings of the International Workshop on Similarity-Based Pattern Recognition, Copenhagen, Debnark, 12–14 October 2015; pp. 84–92. [Google Scholar]
  16. Law, M.T.; Thome, N.; Cord, M. Quadruplet-wise image similarity learning. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 3–6 December 2013; pp. 249–256. [Google Scholar]
  17. Oh Song, H.; Xiang, Y.; Jegelka, S.; Savarese, S. Deep metric learning via lifted structured feature embedding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NA, USA, 27–30 June 2016; pp. 4004–4012. [Google Scholar]
  18. Sohn, K. Improved deep metric learning with multi-class n-pair loss objective. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 1857–1865. [Google Scholar]
  19. Yi, D.; Lei, Z.; Li, S. Deep metric learning for person re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 24–27 June 2014; pp. 34–39. [Google Scholar]
  20. Ustinova, E.; Lempitsky, V. Learning deep embeddings with histogram loss. In Proceedings of the Advances in Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 4170–4178. [Google Scholar]
  21. Wang, J.; Zhou, F.; Wen, S.; Liu, X.; Lin, Y. Deep metric learning with angular loss. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2593–2601. [Google Scholar]
  22. Wu, C.-Y.; Manmatha, R.; Smola, A.J.; Krahenbuhl, P. Sampling matters in deep embedding learning. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2840–2848. [Google Scholar]
  23. Ge, W. Deep metric learning with hierarchical triplet loss. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 269–285. [Google Scholar]
  24. Rao, A.; Srihari, R.K.; Zhang, Z. Spatial color histograms for content-based image retrieval. In Proceedings of the 11th International Conference on Tools with Artificial Intelligence, Chicago, IL, USA, 9–11 November 1999; pp. 183–186. [Google Scholar]
  25. Shannon, C.E.J.B. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  26. Kullback, S.; Leibler, R. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
  27. Goodfellow, I.; Bengio, Y.; Courville, A.; Bengio, Y. Deep Learning; MIT Press: Cambridge, UK, 2016; Volume 1. [Google Scholar]
  28. Hamza, A.B. Jensen-Rhyi Divergence Measure: Theoretical and Computational Perspectives. IEEE Int. Symp. Inf. Theory 2003. [Google Scholar] [CrossRef]
  29. Lehmann, T.M.; Güld, M.O.; Deselaers, T.; Keysers, D.; Schubert, H.; Spitzer, K.; Ney, H.; Wein, B.B.J.; Graphics, C.M.I. Automatic categorization of medical images for content-based retrieval and data mining. Comput. Med Imag. Graph. 2005, 29, 143–155. [Google Scholar] [CrossRef]
  30. Radenovic, F.; Tolias, G.; Chum, O. Fine-tuning CNN Image Retrieval with No Human Annotation. IEEE Trans. Pattern Anal. 2018. [Google Scholar] [CrossRef] [Green Version]
  31. Mikolajczyk, K.; Matas, J. Improving Descriptors for Fast Tree Matching by Optimal Linear Projection. In Proceedings of the IEEE 11th International Conference on Computer Vision, 2007. ICCV 2007, Rio de Janeiro, Brazil, 14–21 October 2007; pp. 1–8. [Google Scholar]
  32. Huang, R.; Jiang, X. Off-Feature Information Incorporated Metric Learning for Face Recognition. IEEE Signal Process. Lett. 2018, 25, 541–545. [Google Scholar] [CrossRef]
  33. Feng, G.; Liu, W.; Tao, D.; Zhou, Y. Hessian Regularized Distance Metric Learning for People Re-Identification. Neural Process. Lett. 2019, 50, 2087–2100. [Google Scholar] [CrossRef]
  34. Tan, M.; Yu, J.; Yu, Z.; Gao, F.; Rui, Y.; Tao, D. User-Click-Data-Based Fine-Grained Image Recognition via Weakly Supervised Metric Learning. ACM Trans. Multimed. Comput. Commun. Appl. (TOMM) 2018, 14, 1–23. [Google Scholar] [CrossRef]
  35. Cao, R.; Zhang, Q.; Zhu, J.; Li, Q.; Qiu, G. Enhancing remote sensing image retrieval with triplet deep metric learning network. arXiv 2019, arXiv:1902.05818. [Google Scholar] [CrossRef] [Green Version]
  36. Xiang, J.; Zhang, G.; Hou, J.; Sang, N.; Huang, R. Multiple target tracking by learning feature representation and distance metric jointly. arXiv 2018, arXiv:1802.03252. [Google Scholar]
  37. Yang, J.; She, D.; Lai, Y.-K.; Yang, M.-H. Retrieving and classifying affective images via deep metric learning. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
  38. Mehmood, Z.; Anwar, S.M.; Ali, N.; Habib, H.A.; Rashid, M. A Novel Image Retrieval Based on a Combination of Local and Global Histograms of Visual Words. Math. Probl. Eng. 2016, 2016, 8217250. [Google Scholar] [CrossRef]
  39. Krapac, J.; Verbeek, J.; Jurie, F. Modeling Spatial Layout with Fisher Vectors for Image Categorization. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 1487–1494. [Google Scholar]
  40. Koniusz, P.; Mikolajczyk, K. Spatial Coordinate Coding to Reduce Histogram Representations, Dominant Angle and Colour Pyramid Match. In Proceedings of the 2011 18th IEEE International Conference on Image Processing (Icip), Brussels, Belguim, 11–14 September 2011; pp. 661–664. [Google Scholar]
  41. Sanchez, J.; Perronnin, F.; de Campos, T. Modeling the spatial layout of images beyond spatial pyramids. Pattern Recogn. Lett. 2012, 33, 2216–2223. [Google Scholar] [CrossRef]
  42. Liu, P.; Miao, Z.; Guo, H.; Wang, Y.; Ai, N. Adding spatial distribution clue to aggregated vector in image retrieval. EURASIP J. Image Video Process. 2018, 2018, 9. [Google Scholar] [CrossRef] [Green Version]
  43. Babenko, A.; Lempitsky, V. Aggregating deep convolutional features for image retrieval. arXiv 2015, arXiv:1510.07493. [Google Scholar]
  44. Razavian, A.S.; Sullivan, J.; Carlsson, S.; Maki, K. Particular object retrieval with integral max-pooling of CNN activations. arXiv 2015, arXiv:1511.05879. [Google Scholar]
  45. Jegou, H.; Chum, O. Negative Evidences and Co-occurences in Image Retrieval: The Benefit of PCA and Whitening. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Pt Ii 2012. Volume 7573, pp. 774–787. [Google Scholar]
  46. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miani, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  47. He, K.M.; Zhang, X.Y.; Ren, S.Q.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (Cvpr), Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  48. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.H.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  49. Chum, O.; Matas, J. Large-scale discovery of spatially related images. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 32, 371–377. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  50. Radenovic, F.; Schonberger, J.L.; Ji, D.; Frahm, J.-M.; Chum, O.; Matas, J. From dusk till dawn: Modeling in the dark. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 5488–5496. [Google Scholar]
  51. Philbin, J.; Chum, O.; Isard, M.; Sivic, J.; Zisserman, A. Object retrieval with large vocabularies and fast spatial matching. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
  52. Philbin, J.; Chum, O.; Isard, M.; Sivic, J.; Zisserman, A. Lost in quantization: Improving particular object retrieval in large scale image databases. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AL, USA, 24–26 June 2008; pp. 1–8. [Google Scholar]
  53. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  54. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  55. Gordo, A.; Almazán, J.; Revaud, J.; Larlus, D. Deep image retrieval: Learning global representations for image search. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 241–257. [Google Scholar]
  56. Kalantidis, Y.; Mellina, C.; Osindero, S. Cross-dimensional weighting for aggregated deep convolutional features. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 685–701. [Google Scholar]
  57. Mohedano, E.; McGuinness, K.; O’Connor, N.E.; Salvador, A.; Marques, F.; Giro-i-Nieto, X. Bags of local convolutional features for scalable instance search. In Proceedings of the 2016 ACM on International Conference on Multimedia Retrieval, New York, NY, USA, 6–9 June 2016; pp. 327–331. [Google Scholar]
  58. Arandjelovic, R.; Gronat, P.; Torii, A.; Pajdla, T.; Sivic, J. NetVLAD: CNN architecture for weakly supervised place recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 5297–5307. [Google Scholar]
  59. Ong, E.-J.; Husain, S.; Bober, M. Siamese network of deep fisher-vector descriptors for image retrieval. arXiv 2017, arXiv:1702.00338. [Google Scholar]
  60. Gordo, A.; Almazan, J.; Revaud, J.; Larlus, D. End-to-end learning of deep visual representations for image retrieval. Int. J. Comput. Vis. 2017, 124, 237–254. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Each rectangle represents a feature descriptor obtained after the convolution of the neural network. Different colors of small squares represent different feature intensities in the descriptor. The Euclidean distance ( D 1 ) between descriptor n and descriptor q is 0.45 , while the Euclidean distance ( D 2 ) between descriptor p and descriptor q is 0.44 . D1 is approximately equal to D2, but the internal spatial distribution of p and n is obviously different.
Figure 1. Each rectangle represents a feature descriptor obtained after the convolution of the neural network. Different colors of small squares represent different feature intensities in the descriptor. The Euclidean distance ( D 1 ) between descriptor n and descriptor q is 0.45 , while the Euclidean distance ( D 2 ) between descriptor p and descriptor q is 0.44 . D1 is approximately equal to D2, but the internal spatial distribution of p and n is obviously different.
Entropy 22 00321 g001
Figure 2. The core idea of relative distribution entropy (RDE)-loss. D represents the Euclidean distance between two image descriptors. RDE represents the relative distribution entropy between two image descriptors. RDE-distance is a new metric that combines Euclidean distance with relative distributed entropy. We called it the relative distribution entropy weighted distance, which can enhance the discrimination of image descriptors.
Figure 2. The core idea of relative distribution entropy (RDE)-loss. D represents the Euclidean distance between two image descriptors. RDE represents the relative distribution entropy between two image descriptors. RDE-distance is a new metric that combines Euclidean distance with relative distributed entropy. We called it the relative distribution entropy weighted distance, which can enhance the discrimination of image descriptors.
Entropy 22 00321 g002
Figure 3. Convolutional neural network (CNN) network architecture.
Figure 3. Convolutional neural network (CNN) network architecture.
Entropy 22 00321 g003
Figure 4. Training process using relative distribution entropy contrastive loss.
Figure 4. Training process using relative distribution entropy contrastive loss.
Entropy 22 00321 g004
Figure 5. Training process using relative distribution entropy triplet loss.
Figure 5. Training process using relative distribution entropy triplet loss.
Entropy 22 00321 g005
Table 1. Experimental results of hyperparameter comparison in the relative distribution entropy contrastive loss function. The best results would be highlighted in bold.
Table 1. Experimental results of hyperparameter comparison in the relative distribution entropy contrastive loss function. The best results would be highlighted in bold.
Network α β Oxford5kOxford5k(W)Pairs6kPairs6k(W)
AlexNet0.501058.1067.6071.6479.60
0.752060.8767.1975.3379.43
0.852560.7967.9375.6079.59
0.903061.3268.2275.2980.07
1.005060.2767.7274.8880.10
VGG0.853084.6287.8382.4088.01
0.8510076.1883.0481.7187.11
0.903085.0988.0082.6988.12
Table 2. Comparative results of different pooling methods on AlexNet. The best results would be highlighted in bold.
Table 2. Comparative results of different pooling methods on AlexNet. The best results would be highlighted in bold.
NetOxford5kOxford5k(W)Pairs6kPairs6k(W)
SPoC [43]41.8355.3455.4968.61
MAC [44]47.5055.9562.1671.30
GeM [30]60.7968.2275.2980.07
Table 3. Performance comparison of relative distribution entropy triplet loss and triplet loss. The best results would be highlighted in bold.
Table 3. Performance comparison of relative distribution entropy triplet loss and triplet loss. The best results would be highlighted in bold.
LossNetworkOxford5kOxford5k(W)Pairs6kPairs6k(W)
Triplet loss [30]VGG81.4882.8082.7984.78
Ours82.3983.0783.6185.45
Triplet loss [30]ResNet81.4985.3387.7091.11
Ours82.8886.5489.3391.97
Table 4. Comparison of our method with the state-of-art image retrieval methods. The best results would be highlighted in bold.
Table 4. Comparison of our method with the state-of-art image retrieval methods. The best results would be highlighted in bold.
NetMethodF-tunedOxford5kOxford105kPairs6kPairs106k
VGGMAC [44]no56.447.872.358.0
SPoC [43]no68.161.178.268.4
Crow [56]no70.865.379.772.2
R-MAC [52]no66.961.683.075.7
BoW-CNN [57]yes73.959.382.064.8
NetVLAD [58]yes71.6-79.7-
Fisher [59]yes81.576.682.4-
R-MAC [55]yes83.178.687.179.7
GeM [30]yes87.983.387.781.3
oursyes88.083.588.179.9
ResR-MAC [52]no69.463.785.277.8
GeM [30]yes87.884.692.786.9
oursyes88.484.992.786.3
Re-ranking(R) and Query Expansion(QE)
VGGCrow + QE [56]no74.970.684.879.4
R-MAC+R+QE [52]no77.373.286.579.8
BoW-CNN+R+QE [57]no78.865.184.864.1
R-MAC+QE [55]yes89.187.391.286.8
GeM+ α QE [30]yes91.989.691.987.6
oursyes92.089.692.487.3
ResR-MAC+QE [52]no78.975.589.785.3
R-MAC+QE [60]yes90.689.496.093.2
GeM+QE [30]yes91.089.595.591.9
oursyes91.789.796.092.1

Share and Cite

MDPI and ACS Style

Liu, P.; Shi, L.; Miao, Z.; Jin, B.; Zhou, Q. Relative Distribution Entropy Loss Function in CNN Image Retrieval. Entropy 2020, 22, 321. https://0-doi-org.brum.beds.ac.uk/10.3390/e22030321

AMA Style

Liu P, Shi L, Miao Z, Jin B, Zhou Q. Relative Distribution Entropy Loss Function in CNN Image Retrieval. Entropy. 2020; 22(3):321. https://0-doi-org.brum.beds.ac.uk/10.3390/e22030321

Chicago/Turabian Style

Liu, Pingping, Lida Shi, Zhuang Miao, Baixin Jin, and Qiuzhan Zhou. 2020. "Relative Distribution Entropy Loss Function in CNN Image Retrieval" Entropy 22, no. 3: 321. https://0-doi-org.brum.beds.ac.uk/10.3390/e22030321

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop