Next Article in Journal
ARK-BIM: Open-Source Cloud-Based HBIM Platform for Archaeology
Previous Article in Journal
An Enhanced Photonic Quantum Finite Automaton
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Discrete Semantics-Guided Asymmetric Hashing for Large-Scale Multimedia Retrieval

1
School of Computer Science and Engineering, Central South University, Changsha 410083, China
2
Network Resources Management and Trust Evaluation Key Laboratory of Hunan Province, Central South University, Changsha 410083, China
3
Big Data Institute, Central South University, Changsha 410083, China
*
Author to whom correspondence should be addressed.
Submission received: 13 July 2021 / Revised: 10 September 2021 / Accepted: 12 September 2021 / Published: 21 September 2021

Abstract

:
Cross-modal hashing technology is a key technology for real-time retrieval of large-scale multimedia data in real-world applications. Although the existing cross-modal hashing methods have achieved impressive accomplishment, there are still some limitations: (1) some cross-modal hashing methods do not make full consider the rich semantic information and noise information in labels, resulting in a large semantic gap, and (2) some cross-modal hashing methods adopt the relaxation-based or discrete cyclic coordinate descent algorithm to solve the discrete constraint problem, resulting in a large quantization error or time consumption. Therefore, in order to solve these limitations, in this paper, we propose a novel method, named Discrete Semantics-Guided Asymmetric Hashing (DSAH). Specifically, our proposed DSAH leverages both label information and similarity matrix to enhance the semantic information of the learned hash codes, and the 2 , 1 norm is used to increase the sparsity of matrix to solve the problem of the inevitable noise and subjective factors in labels. Meanwhile, an asymmetric hash learning scheme is proposed to efficiently perform hash learning. In addition, a discrete optimization algorithm is proposed to fast solve the hash code directly and discretely. During the optimization process, the hash code learning and the hash function learning interact, i.e., the learned hash codes can guide the learning process of the hash function and the hash function can also guide the hash code generation simultaneously. Extensive experiments performed on two benchmark datasets highlight the superiority of DSAH over several state-of-the-art methods.

1. Introduction

In recent years, due to the rapid development of multimedia Internet of Things technologies, there has been an explosive growth in the amount of multimedia network data. Consequently, the current unimodal search methods can no longer meet the multimedia data retrieval requirements in the complex environment of the new information era. Therefore, cross-modal retrieval methods [1,2,3] have received increasing attention from the information retrieval community and have become a hot research topic in both academia and industry. Specifically, given a query in one modality (such as text), users expect to return its semantically related modality (text) or different modalities (such as images or videos). For decades, as a branch of nearest neighbor search (NNS), the hashing technique has been an active research field in information retrieval community due to the following advantages: (1) Lower storage cost and (2) improved retrieval speed with the hardware-friendly bit-wise XOR and bit-count operations [4]. In the hash code learning process, the learned hash codes should meet a condition, that is, similar instances have similar hash codes in the Hamming space, and vice versa. Among the practical applications are the image retrieval [5,6] and person re-identification [7,8].
According to the learning principle, existing cross-modal hashing methods can be mainly divided into the following categories: Unsupervised cross-modal hashing methods [9,10,11,12,13,14,15]: Unsupervised hashing methods focus on discovering the intra- and inter-relations of multiple heterogeneous modalities without label information to learn the hash codes and corresponding hash functions. However, due to the lack of label information, the retrieval performance of unsupervised cross-modal hashing methods will be affected accordingly, i.e., x f ( x ; ` ) z d ( z ; φ ) x ^ , where f ( · ; θ ) can be considered as either a hash mapping function or an encoder, d ( z ; φ ) can be regarded as a decoder. During the unsupervised learning process, the key design factor of this learning paradigm is the choice of a suitable metric that can measure the distance between x and x ˜ , i.e., the distance between x , x ˜ should be minimized: min θ , φ E | | x x ˜ | | p , where typically p = 2 . Then, the hash codes can be computed by s g n ( z ) . Supervised cross-modal hashing methods [16,17,18,19,20,21,22]: Supervised hashing methods have obtained satisfactory retrieval results by using label information, and have been extensively studied, i.e., x f ( x ; θ ) z c ( z ) L , where f ( · ; θ ) denotes a hash mapping function that selects certain latent representation z , c ( · ) denotes a classifier, sgn ( · ) denotes the element-wise sign operation, L denotes the labels, and B denotes the learned hash codes. Then, the hash codes can be computed by s g n ( z ) .
Although supervised cross-modal hashing methods have achieved significant success, they still have the following limitations: (1) Limited semantic utilization. Converting the label matrix directly to the similarity matrix will lead to a semantic loss, resulting in a large semantic gap, especially when facing multi-label datasets. (2) Inefficient learning strategy. Some cross-modal hashing methods are based on symmetric learning strategies, resulting in a worse retrieval performance than asymmetric learning ones. (3) Flawed optimization strategy. As the optimization process of the hash codes is discrete, the existing optimization strategies are mainly based on two kinds, one is to use the relaxation-based strategy, which will lead to a large quantization error; the other is to use bit-to-bit optimization strategy, such as Discrete Cyclic Coordinate (DCC) descent [23]. Although the problem of quantization error is solved, optimizing the entire hash code requires k iterations, where k is the hash code length, thus the optimization process is very time-consuming.
In order to solve the above limitations, in this paper, we proposed a novel yet simple but effective method, named Discrete Semantics-Guided Asymmetric Hashing (DSAH). Specifically, DSAH handles the nonlinear relations in different modalities with a kernelization technique, then an asymmetric learning scheme is proposed to effectively perform the hash function learning and hash code learning processes; meanwhile, our proposed DSAH considers the following aspects. First, we leverage both label information and similarity matrix to enhance the semantic information of the learned hash codes. Then, we solve the problems of matrix sparsity and outlier processing. In addition, a discrete optimization algorithm is proposed to solve the discrete problems. Our major contributions can be summarized as follows:
1.
A novel supervised cross-modal hashing method, i.e., DSAH, is proposed to learn the discriminative compact hash codes for large-scale retrieval tasks. DSAH takes the label information and similarity matrix into consideration, which can improve the discriminative capability of the learned hash codes, and solves the problems of matrix sparseness and outlier processing.
2.
An asymmetric learning scheme with real-valued embeddings is proposed to effectively learn the hash function and the hash codes.
3.
Comprehensive experiments are conducted on two famous datasets. The experimental results demonstrate that DSAH outperforms some state-of-the-art baselines.
The remainder of this paper is organized as follows. Section 2 briefly reviews the related works. Section 3 introduces the details of the DSAH  model and presents the alternative optimization algorithm. In Section 4, we give the results of experiments performed. Finally, we present the conclusions in Section 5.

2. Related Works

  Cross-modal hashing retrieval has been a widely used technology in the field of information retrieval, machine intelligence, and computer vision. As mentioned above, cross-modal hashing retrieval technology can be divided into supervised and unsupervised categories. However, due to the space limitation, we refer readers to some surveys [4,24] for a more comprehensive coverage of popular hashing methods.

2.1. Unsupervised Hashing

Unsupervised hashing methods do not leverage label information for the training dataset. For example, Local Sensitive Hashing (LSH) [25] and its variants [26,27] use random projections to map instances into a Hamming space. Iterative Quantization (ITQ) [28] optimizes the projection by PCA, and then learns an orthogonal rotation matrix to bridge the quantization gap. Unsupervised semantic deep hashing (USDH) [29] uses semantic information to guide the training of hash mapping function. Unsupervised Deep Video Hashing (UDVH) [30] learns the hash codes in a self-taught manner by jointly integrating discriminative video representation with optimal code learning. Neighborhood Discriminant Hashing (NDH) [31] learns hash function by preserving the neighborhood discriminative information in Hamming space. Collective Matrix Factorization Hashing (CMFH) [32] learns a shared common latent space by a collective matrix factorization algorithm, and then adopts a thresholding operation to generate the hash codes. Fusion Similarity Hashing (FSH) [33] proposes a novel fusion similarity method for hash learning by capturing the latent relations between different heterogeneous modalities.

2.2. Supervised Hashing

Unlike unsupervised hashing methods, supervised ones utilize the labels to improve the retrieval performance. For example, Semantic Correlation Maximization (SCM) [34] preserves the pairwise similarities of the training dataset to learn the hash functions. Semantic Preserving Hashing (SePH) [35] constructs a new semantic affinity matrix into a probability distribution, and then learns the hash codes by using an approximate-based scheme. Discrete Cross-modal Hashing (DCH) [36] learns the hash codes in a bit-by-bit manner by using the discrete cyclic coordinate descent (DCC) algorithm. Label Consistent Matrix Factorization Hashing (LCMFH) [37] directly uses semantic labels to guide the hashing learning procedure. Scalable Discrete Matrix Factorization Hashing (SCRATCH) [38] is a two-step hashing method, which first generates the hash codes, and then learns the hash functions based on the learned hash codes. Nonlinear Robust Discrete Hashing (NRDH) [39] uses a nonlinear model to solve the generalization error caused by kernelization, and generates compact hash codes by constructing an asymmetric hash framework and discrete optimization algorithms. Scalable Deep Asymmetric Hashing (SDAH) [40] builds an asymmetric unequal-dimensional hash learning framework by exploring the information content of different modalities. Nonlinear Supervised Discrete Hashing (NSDH) [41] consists of two parts, the first part is a semantic enhancement descriptor, which is used to extract comprehensive latent representations of heterogeneous multimedia data, and the second part is a fast discrete optimization module, which is used to learn discriminative compact hash codes. Subspace Relation Learning for Cross-modal Hashing (SRLCH) [42] handles relationships of labels in a semantic subspace to make similar instances from different modalities closer in the binary Hamming space.

2.3. Deep Hashing

Recently, with the great success of deep learning in the field of representation learning, many deep hashing methods [13,15,40,43] have been proposed. For example, Deep Semantic-Alignment Hashing (DSAH) [44] is an unsupervised hashing method, which explores the similarity information of different modalities and proposes a semantic-alignment loss to learn the hash codes. Unsupervised Deep Cross-modal Hashing with Virtual Label Regression (UDCH-VLR) [45] proposes a novel unified learning framework to jointly perform deep hash function training, virtual label learning, and regression. Deep Saliency Hashing (DSaH) [46] is a two-step end-to-end model, which mines salient regions and learns semantic-preserving hash codes simultaneously. Supervised Hierarchical Deep Cross-modal Hashing (SHDCH) [47] learns the hash codes by explicitly delving into the hierarchical labels. Deep Semantic cross-modal hashing with Correlation Alignment (DSCA) [48] designs two deep neural networks for image and text modality separately, and learns two hash functions. First, due to the non-smooth property of the discrete optimization causing the problem of unavailable gradient in back-propagation, these methods use the relaxation-based optimization strategies to handle the problem. Even though high retrieval performance is considered to be achieved, these methods still have a large quantization error and can only produce sub-optimal hash codes. Second, they all need large computing resources (e.g., GPUs) and a massive training dataset, which makes them fairly computationally expensive.

3. The Proposed DSAH Framework

  In this section, we introduce our proposed DSAH model. The framework of DSAH is shown in Figure 1, which consists of three main parts: hash function learning, label alignment scheme and asymmetric learning framework. We demonstrate each part in the following section in detail.

3.1. Definitions

Suppose that the multimedia training data contains M modalities, represented by X = { X ( 1 ) , X ( 2 ) , . . . , X ( M ) } , where X ( m ) = { x 1 ( m ) , x 2 ( m ) , . . . , x n ( m ) } R d m × n is the m-th modality data features and d m is the dimension of modality X ( m ) . In this paper, we focus on the supervised hashing paradigm; therefore, label information L { 0 , 1 } c × n is available, c denotes the number of categories, and l i j = 1 indicates the j-th instance belongs to category i, and l i j = 0 otherwise. B R k × n denotes the hash codes, where k is the length of hash codes. f ( · ) denotes the hash function. The main notations used in this paper are listed in Table 1.

3.2. Hash Function Learning

3.2.1. Kernelization

Kernelization is a widely used technique to handle the nonlinear relations in different modalities. Therefore, in this paper, we use Radical Basis Function (RBF) kernel to express the nonlinear correlations among original high dimensional features [49,50,51]. Specifically, we define the RBF function ϕ ( · ) as follows:
ϕ ( x i ) = exp | | x i a 1 | | 2 œ 2 , . . . , exp | | x i a q | | 2 œ 2 . ,
where A = [ a 1 , a 2 , . . . , a q ] denotes the randomly chosen q anchors from the database and σ is the free parameter. Therefore, the complex original feature X ( m ) R d m × n can be converted into a nonlinear relation representation ϕ ( X ( m ) ) R q × n .

3.2.2. Feature Mapping

The aim of DSAH is to project the original features to compact hash codes. In this paper, we adopt two linear projections as the hash mapping functions for image modality X ( 1 ) and text modality X ( 2 ) , respectively.
f 1 ( X ( 1 ) ) = s g n ( P 1 ϕ ( X ( 1 ) ) ) , f 2 ( X ( 2 ) ) = s g n ( P 2 ϕ ( X ( 2 ) ) ) ,
where P 1 R k × q and P 2 R k × q are the hash mapping matrices, which map specific kernel features into Hamming subspace, and f 1 ( · ) and f 2 ( · ) are hash functions for image modality and text modality, respectively.

3.3. Label Alignment Scheme

As described above, labels contain rich semantic information, directly converting the complex label vectors into binary semantic matrix will cause the loss of semantic information. The results of Gui’s work [52] demonstrate that the ordinary least squares regression is sensitive to the boundary contour. Inspired by the work in [53], we consider p , q norm instead of 2 norm to handle the problem, the  p , q norm can be formulated as
min E | | E | | p , q = min E ( i = 1 c ( j = 1 n | E i j | p ) q / p ) 1 / q ,
where E = R B L and R R k × c is the semantic projection matrix. It is easy to find that when p = q = 2 , Equation (3) is a standard Frobenius norm. However, in order to improve the robustness of the model for outliers and the sparsity of the label alignment matrix, we need to redefine the values of p , q . In general, the sparsity of the model can be guaranteed when the constraint conditions satisfy p 2 and 0 q 2 . Therefore, in the paper, we set p = 2 and q = 1 , as if q = 0 , the problem is not convex. Then, we can rewrite the Equation (3) as min E | | E | | 2 , 1 . After some algebraic manipulations, we obtain
min R , B t r ( E DE ) ,
where D R c × c is the diagonal matrix, and the i-th element of D is defined as d i i = 1 2 | | e ( i ) | | 2 , where e ( i ) is the i-th row of E .

3.4. Asymmetric Learning Framework

We briefly review the related work Supervised Hashing with Kernels (KSH) [51], the symmetric learning framework can be formulated as
min B | | B B k S | | F 2 s . t . B { 1 , 1 } k × n ,
where B is the learned hash codes. However, there are two major problems of Equation (5): (1) It is very time-consuming to directly compute S , as the similarity information S is a n × n matrix. (2) Some works [54,55] show that the use of an asymmetric learning framework can not only solve the problem of high time consumption, but also improves retrieval accuracy, because the value range of the asymmetric learning framework is wider than that of symmetric learning. Therefore, in this paper, we construct an asymmetric learning framework to learn the compact hash codes, that is,
min B , P 1 , P 2 | | ( P 1 ϕ ( X ( 1 ) ) ) B 1 k S | | F 2 + | | ( P 2 ϕ ( X ( 2 ) ) ) B 2 k S | | F 2 + α i = 1 2 | | B i P i ϕ ( X ( i ) ) | | F 2 , s . t . B i { 1 , 1 } k × n
The advantages of Equation (6) are as follows:
1.
The learning mode uses an efficient asymmetric learning architecture instead of a time-consuming symmetric one.
2.
The use of the real-valued embeddings instead of the binary embeddings produces a close semantic similarity relation, and the value of the objective function is smaller.
3.
The last term is used to reduce the quantization errors, which leads to a better retrieval performance.
However, there is a limitation of Equation (6) that is for the purpose of cross-modal retrieval tasks, we need to obtain a unified hash codes. Therefore, we need to consider another discrete constraint, i.e.,  min B 1 , B 2 | | B 1 B 2 | | F 2 . In order to make the optimization easy, we set a unified hash code B = B 1 = B 2 instead of minimizing the constraint min B 1 , B 2 | | B 1 B 2 | | F 2 , then Equation (6) can be rewritten as
min B , P 1 , P 2 | | ( P 1 ϕ ( X ( 1 ) ) ) B k S | | F 2 + | | ( P 2 ϕ ( X ( 2 ) ) ) B k S | | F 2 + α | | B 0.5 P 1 ϕ ( X ( 1 ) ) + P 2 ϕ ( X ( 2 ) ) | | F 2 , s . t . B { 1 , 1 } k × n
where α is the balance parameter.

3.5. The Joint Framework

Combining the above constraints and individual objective function, we obtain
min B , R , P 1 , P 2 | | ( P 1 ϕ ( X ( 1 ) ) ) B k S | | F 2 + | | ( P 2 ϕ ( X ( 2 ) ) ) B k S | | F 2 + t r ( E DE ) + α | | B 0.5 P 1 ϕ ( X ( 1 ) ) + P 2 ϕ ( X ( 2 ) ) | | F 2 + γ R e ( P 1 * , P 2 * , R ) , s . t . B { 1 , 1 } k × n
where γ is the trade-off parameter, P i * = P i ϕ ( X ( i ) ) , R e ( · ) = | | · | | F 2 is the Frobenius norm regularization term, which is used to prevent overfitting.

3.6. Optimization

In this part, we use an alternating strategy to solve the four variables B , R , P 1 , P 2 in Equation (8), as the four variables are coupled with each other. The problem is split into four steps as follows.
Fix  R , P 1 , P 2 , update  B . The sub-problem of Equation (8) related to B can be formulated as
min B | | ( P 1 ϕ ( X ( 1 ) ) ) B k S | | F 2 + | | ( P 2 ϕ ( X ( 2 ) ) ) B k S | | F 2 + t r ( E DE ) + α | | B 0.5 P 1 ϕ ( X ( 1 ) ) + P 2 ϕ ( X ( 2 ) ) | | F 2 , s . t . B { 1 , 1 } k × n
In the next step, we need to solve the following problem:
min B t r 2 k SB P 1 ϕ ( X ( 1 ) ) 2 k SB P 2 ϕ ( X ( 2 ) ) + B RDR B 2 B RDL 2 α CB s . t . B { 1 , 1 } k × n
where C = 0.5 ( P 1 ϕ ( X ( 1 ) ) + P 2 ϕ ( X ( 2 ) ) ) . As the B is the discrete value, it is challenging to directly solve the value of B . In this solution, we propose an augmented Lagrangian multiplier (ALM) [39] to compute B . Specifically, we introduce an auxiliary value V { 1 , 1 } k × n to replace the B of second term, i.e.,  B RDR B . Then, we obtain the following formula:
min B t r 2 k SB P 1 ϕ ( X ( 1 ) ) 2 k SB P 2 ϕ ( X ( 2 ) ) 2 α CB + B RDR V 2 B RDL + ξ 2 | | B V + J b ξ | | F 2 , s . t . { B , V } { 1 , 1 } k × n
where J b measures the gap between B and V .
Then, the value of B can be solved with a closed-form solution:
B = s g n ( 2 k P 1 ϕ ( X ( 1 ) ) S + 2 k P 2 ϕ ( X ( 2 ) ) S + 2 α C RDR V + 2 RDL + ξ V J b ) .
However, the computational complexity of P * ϕ ( X ( * ) ) S | * = 1 2 is O ( n 2 ) , which is not suitable for large-scale retrieval tasks. To address this problem, we use the label matrix L R c × n to replace the similarity matrix S R n × n . Specifically, we let L ˜ i j = l i j / | | l i | | 2 , as the element at the i-th row and the j-th column in the matrix L . Then, the similarity matrix S can be rewritten as
S = 2 L ˜ L ˜ 1 n 1 n ,
where 1 n is a vector with all elements being 1. Therefore, we can rewrite the term P * ϕ ( X ( * ) ) S | * = 1 2 as
P * ϕ ( X ( * ) ) S | * = 1 2 = 2 P * ϕ ( X ( * ) ) L ˜ L ˜ P * ϕ ( X ( * ) ) 1 n 1 n | * = 1 2 ,
which consumes O ( ( q + c ) k n ) .
Fix  B , update  V . The sub-problem related to V can be formulated as
min V t r ( B RDR V ) + ξ 2 | | B V + J b ξ | | F 2 . s . t . { B , V } { 1 , 1 } k × n
Then, the value of V can be solved with a closed-form solution,
V = s g n ( RD R B + ξ B + J b ) .
Update  J b . The sub-problem related to J b can be updated as
J b = J b + ξ ( B V ) , ξ = ρ ξ ,
where ρ is a parameter to control the convergence speed.
Fix  B , P 1 , P 2 , update  R . The sub-problem of Equation (8) related to R can be formulated as
min R t r ( E DE ) + γ R e ( R ) .
In the next step, we need to solve the following problem:
min R t r ( B RDR B 2 BL DR ) + γ t r ( RR ) .
Setting the derivative Equation (20) w.r.t R to 0, we obtain
BB RD + γ R = BL D .
We transform Equation (20) into
BB R + γ RD 1 = BL .
Then, it can be seen that Equation (21) is a Sylvester equation. Therefore, the value of R can be easily solved. Due to the space limitation, the detail about the solution is not given here.
Fix  B , R , P 2 , update  P 1 . The sub-problem of Equation (8) related to P 1 can be formulated as
min P 1 | | ( P 1 ϕ ( X ( 1 ) ) ) B k S | | F 2 + α | | B 0.5 P 1 ϕ ( X ( 1 ) ) + P 2 ϕ ( X ( 2 ) ) | | F 2 + γ R e ( P 1 * ) , s . t . B { 1 , 1 } k × n
Setting the derivative Equation (22) w.r.t P 1 to 0, we obtain
2 P 1 ϕ ( X ( 1 ) ) ϕ ( X ( 1 ) ) 2 k BS ϕ ( X ( 1 ) ) 4 α B ϕ ( X ( 1 ) ) + 2 α P 1 ϕ ( X ( 1 ) ) ϕ ( X ( 1 ) ) + 2 α P 2 ϕ ( X ( 2 ) ) ϕ ( X ( 1 ) ) + 2 γ P 1 ( X ( 1 ) ) ϕ ( X ( 1 ) ) = 0
Then, the value of P 1 can be solved with a closed-form solution:
P 1 = ( k BS ϕ ( X ( 1 ) ) + 2 α B ϕ ( X ( 1 ) ) α P 2 ϕ ( X ( 2 ) ) ϕ ( X ( 1 ) ) ) · ( ( 1 + α + γ ) ϕ ( X ( 1 ) ) ϕ ( X ( 1 ) ) ) 1
where S is also transformed using Equation (13); then, we have
BS ϕ ( X ( 1 ) ) = 2 ( B L ˜ ) ( ϕ ( X ( 1 ) ) L ˜ ) ( B 1 n ) ( ϕ ( X ( 1 ) ) 1 n ) ,
which consumes O ( ( q + k ) c n ) .
Fix  B , R , P 1 , update  P 2 . The sub-problem of Equation (8) related to P 2 can be formulated as
min P 2 | | ( P 2 ϕ ( X ( 2 ) ) ) B k S | | F 2 + α | | B 0.5 P 1 ϕ ( X ( 1 ) ) + P 2 ϕ ( X ( 2 ) ) | | F 2 + γ R e ( P 2 * ) , s . t . B { 1 , 1 } k × n
It is easy to find that the optimization of P 2 is almost identical to P 1 -subproblem. Then, the value of P 2 can be solved with a closed-form solution:
P 2 = ( k BS ϕ ( X ( 2 ) ) + 2 α B ϕ ( X ( 2 ) ) α P 1 ϕ ( X ( 1 ) ) ϕ ( X ( 2 ) ) ) · ( ( 1 + α + γ ) ϕ ( X ( 2 ) ) ϕ ( X ( 2 ) ) ) 1
Moreover, the terms of ϕ ( X ( 1 ) ) ϕ ( X ( 1 ) ) and ϕ ( X ( 2 ) ) ϕ ( X ( 2 ) ) are constants and can be computed once before the iterative optimization.
The objective function is solved by iteratively updating four variables until the objective function converges or reaches the preset maximum number of iterations. The iterative optimization for solving the Equation (8) is summarized in Algorithm 1.
Algorithm 1:DSAH
  • Input: Training modalities X = { X ( 1 ) , . . . , X ( M ) } , labels L , hash code length k, parameter α , γ , maximum iteration number T.  
  • Output: Hash mapping functions P 1 and P 2 .  
  • Procedure:
  • 1. Centralize X by means.
  • 2. Computing Kernelized features ϕ ( X ) .
  • 3. Initialize V , B , R , P 1 , P 2 , J b by random initialization.  
  • 4. Repeat
  •  B -Step: Update B according to (12).
  •  V -Step: Update V according to (16).
  •  J b -Step: Update J b according to (17).
  •  R -Step: Update R according to (21).
  •  P 1 -Step: Update P 1 according to (24).
  •  P 2 -Step: Update P 2 according to (27).
  •  Until up to T.  

3.7. Out-of-Sample Extension

In the query phase, the proposed DSAH can easily map the original high-dimensional instances into compact hash codes. Specifically, given a new query x q ( t ) X ( m ) , DSAH learns its corresponding hash codes by
b q ( t ) = s g n ( P t ϕ ( x q ( t ) ) ) ,
where ϕ ( x q ( t ) ) is the nonlinear kernelized embedding of x q ( t ) .

3.8. Complexity Analysis

For each iteration, the time complexity is analyzed as follows. The time computational complexity of B is O ( k 2 c + k c 2 + ( q + c + k ) k n ) , V is O ( k c 2 + k 2 c + k 2 n ) , R is O ( ( k 2 + k c ) n + k 2 c + k c 2 + c 3 ) , P 1 and P 2 are all O ( q 3 + k q 2 + ( k q + k c + q c + q 2 ) n ) . As { k , c , q } n , the training complexity is O ( ( k q + k 2 + k c + q 2 ) n ) . Given the iteration T, the overall training complexity for DSAH is O ( ( k q + k 2 + k c + q 2 ) n T ) , where T { k , c , q } is very small, which is linear to the training set size. Therefore, DSAH is highly scalable for large-scale cross-modal retrieval tasks.

4. Experiments

4.1. Datasets

To evaluate the performance of DSAH, we conducted experiments on two widely used datasets, i.e., MIRFlickr [56] and NUS-WIDE [57] datasets.

4.1.1. MIRFlickr

It contains 25,000 instances collected from open website, which are annotated by at least one of 24 tags. Similar to the work in [38], we ignored the instances that textual tags appear less than 20 times and finally selected 20,015 instances. We randomly selected 2 k instances as the query set and the rest as the retrieval set. Each image is represented as a 512-D GIST feature and each text is represented as a 1386-D bag-of-word (BOW) vector.

4.1.2. NUS-WIDE

It contains 269,648 instances collected from Flick with 5018 unique tags and 81 ground-truth concepts that can be used for evaluation. Similar to the work in [38], we selected the ten most frequent tags and corresponding 186,577 instances. We randomly selected 2000 instances as the query set and the rest as the retrieval set. Each image is represented as a 500-D SIFT feature and each text is represented as a 1000-D bag-of-word (BOW) vector.

4.2. Methodology

To verify the effectiveness of our proposed DSAH method, seven state-of-the-art cross-modal hashing methods are compared. Among them, CMFH [32] and FSH [33] are unsupervised cross-modal hashing methods, and SCM-Seq [34], SePH-km [35], DCH [36], LCMFH [37], and SRLCH [42] are supervised ones. We have briefly introduced the compared baselines in Section 2. For fair comparison, the experimental results with citations are copied from the corresponding works.
To evaluate our proposed DSAH, we conducted two cross-modal retrieval tasks: (1) “Image2Text” using an image query to retrieve texts; (2) “Text2Image” using a text query to retrieve images. In this paper, three widely-used evaluation measures are used to evaluate the retrieval performance, i.e., mean average precision (mAP), precision-recall curves (PR) and precisions w.r.t top-k returned image (P@k).

4.3. Implementation Details

DSAH consists of several parameters, i.e., α and γ . We tune the balance parameters, i.e., α and γ using grid search, and the best performance is achieve when { α = 10 1 , γ = 10 3 } and { α = 10 2 , γ = 10 3 } on MIRFlickr and NUS-WIDE datasets, respectively. q is the number of kernel and optimal performance is obtained when q = 2000 . ξ and ρ are used for ALM algorithm and the best performance is obtained when { ξ = 10 2 , ρ = 1.5 } and { ξ = 10 1 , ρ = 1.5 } on MIRFlickr and NUS-WIDE datasets, respectively. All our experiments are conducted on a workstation with a Intel Xeon Silver 4210 [email protected] GHz of 10 cores and 128 G RAM.

4.4. Results

Table 2 and Table 3 show the mAP scores of different compared cross-modal hashing methods at 8 bits, 16 bits, 32 bits, 64 bits, and 128 bits on MIRFlickr and NUS-WIDE datasets, respectively. Note that the mAP metric is one of the comprehensive evaluation criterions used to measure the effectiveness of the proposed method. From these tables, it can be observed that the mAP scores of DSAH are higher than most compared baselines with different code lengths on the two datasets. In the seven compared baselines, only SRLCH, LCMFH, and DCH obtain satisfactory retrieval performance. The main reason is that they learn the common latent representation across different modalities through matrix factorization operations, thus the common latent representation can be used as a bridge to solve the heterogeneous gap between different modalities. However, they ignore the use of an asymmetric learning framework to enhance the semantic similarity of different modalities and the noises contained in the labels. In contrast, our proposed DSAH leverages both the similarity matrix and label information to enhance the semantic information of the learned hash codes, and solves the problem of noises contained in the labels. Specifically, on the MIRFlickr dataset, compared to the best baselines, i.e., SRLCH, the mAP scores of DSAH have an increase of 2.7% on average, and on the NUS-WIDE dataset, DSAH obtains the highest mAP scores of all compared baselines, which demonstrates the efficacy of DSAH. Meanwhile, by comparing supervised cross-modal hashing methods and unsupervised ones on the two datasets, we find that the supervised hashing methods, i.e., SCM-Seq, SePH-km, DCH, LCMFH, and SRLCH, can always outperform the unsupervised hashing methods, i.e., CMFH and FSH, as the supervised information can improve the ability of hash learning process. In addition, the mAP scores of most cross-modal hashing methods increase as the length of the hash codes becomes longer, revealing that the longer codes can handle more discriminative information. The performance on the T2I task, i.e., the use of text modality to retrieve image modality is better than that on the I2T task, i.e., the use of image modality to retrieve text modality. The reason is that the semantic information in text modality is more than that in image modality.
Figure 2 plots the precision–recall and P@k curves in the cases of 64-bit code length for all compared baselines on two datasets. From the figure, we can draw the following observations.
1.
From the precision–recall curves, we can observe that the area under the precision–recall curves of DSAH is larger than the compared baselines, which shows the effectiveness of DSAH.
2.
From the P@k curves, we can observe that DSAH outperforms the compared baselines in most cases, which further demonstrate its superiority.

4.5. Further Study

4.5.1. Effects of Discrete Optimization

To validate the effects of the proposed discrete optimization strategy, we denote a variant of DSAH, named DSAH-Re. Specifically, we first relax the discrete constraints, then Equation (9) can be solved as
min B α i = 1 2 | | B P i ϕ ( X ( i ) ) | | F 2 + t r ( E DE ) .
Setting the derivative Equation (29) w.r.t. B to 0, and the value of B can be solved with a closed-form solution:
B = ( RDR ) 1 α ( P 1 ϕ ( X ( 1 ) ) + P 2 ϕ ( X ( 2 ) ) ) + RDL .
Then, we obtain the hash codes by mean-thresholding operation. The mAP results of DSAH and DSAH-Re on two datasets are shown in Table 4 and Table 5. From the table, we can observe that the performance of DSAH is better than that of DSAH-Re on two datasets. These results demonstrate that our proposed discrete optimization algorithm performs well in avoiding quantization errors and improving the performance of cross-modal retrieval tasks.

4.5.2. Effects of Kernelization

In this paper, DSAH adopts a kernelization technique to handle the nonlinear relations between different heterogeneous modalities to improve the retrieval accuracy and efficiency. To demonstrate the effects of kernelization, we denote a variant of DSAH, named DSAH-ke, which directly uses the original features to learn the hash codes. We conduct experiments on two datasets with the code length varying from 8 bits to 128 bits to evaluate the performance of DSAH-ke. The mAP results of DSAH-ke are reported in Table 4 and Table 5. From the tables, we can observe that the lack of using kernelization will reduce the retrieval performance.

4.5.3. Effects of 2 , 1 Norm

As shown in Section 3.3, the p , q norm, i.e., 2 , 1 , is used to improve the robustness for outliers. Therefore, in this section, to verify its effectiveness, we denote a variant of DSAH, named DSAH-Nm, which replaced the term | | R B L | | 2 , 1 in Equation (8) with | | R B L | | F 2 . The mAP results on two datasets with the code length varying from 8 bits to 128 bits are reported in Table 4 and Table 5. From the table, we can observe that the 2 , 1 norm is effective to improve the performance of DSAH, the reason may be that the label information often inevitably contains some noises or subjective factors.

4.5.4. Effects of Word Embeddings

In order to verify the impact of different word embeddings on the performance of cross-modal retrieval. We denote a Bidirectional Encoder Representations from Transformers (BERT)-based [58] variant of DSAH, named DSAH-BERT. The BERT-based word embeddings are generated by summing the 786-D features from the last 4-hidden layers of a 12 layers BERT trained in an uncased way (https://github.com/huggingface/transformers, accessed on 13 July 2021). We conduct experiments on NUS-WIDE dataset to evaluate the effects of word embeddings. The mAP scores on NUS-WIDE dataset with the code length varying from 8 bits to 128 bits are reported in Table 6. From the table, we can find that BERT-based word embedding provides a slightly higher average mAP scores than those with bag-of-word embedding.

4.5.5. Effects of Deep Learning Based Representation

More recently, deep neural networks have achieved promising performance in the field of representation learning. To validate the effectiveness of DSAH, we conduct experiments on NUS-WIDE dataset to evaluate the effects of deep learning based representations. The corresponding variant of DSAHis named DSAH-Deep. Specifically, each image is represented as a 4096-D vector extracted by the fc7-layer of VGG-16 net [59]. The mAP scores on NUS-WIDE dataset with the code length varying from 8 bits to 128 bits are shown in Table 7. From the table, we can observe that the use of deep learning based representation for cross-modal retrieval improves the accuracy of retrieving text through images, but reduces the accuracy of retrieving images through text. The reason may be that deep learning-based representation improves the semantics of the image representation, but the difficulty of retrieving images is increased due to the increase of the dimensionality simultaneously.

4.5.6. Effects of Parameters

In this section, we conduct parameter sensitivity analysis experiments to observe the variation of mAP scores under different α and γ . In this experiment, by prefixing the code length as 64 bits, we vary the parameters α and γ in the range of { 10 5 , 10 4 , . . . , 10 4 , 10 5 } . Figure 3 and Figure 4 report the results. From these results, we observe that the performance of our proposed DSAH is relatively stable on a wide range of α and γ values. Specifically, on the MIRFlickr dataset, when α < 10 0 and γ < 10 1 , the retrieval performance becomes stable. On the NUS-WIDE dataset, when α < 10 2 and γ < 10 1 , the scores of mAP have a very small fluctuation. Therefore, our proposed method can be easily tuned for practical implementations.

4.5.7. Convergence Analysis

In order to show the convergence of our proposed DSAH, we conduct the experiments on two datasets with the codes length fixed as 64 bits. Similar results can be obtained on other lengths of hash codes. The results are shown in Figure 5. Note that, in order to visually represent the convergence of the objective function, the value can be normalized by dividing by the maximum value on each dataset. From the figure, we can easily see that the values of objective function can converge very fast, i.e., less than 12 iterations, which demonstrates the efficiently of the closed-form solutions of the optimization algorithm.

4.6. Limitations

The main potential limitation of our proposed DSAH is that the time complexity of constructing the pairwise similarity matrix is O ( n 2 ) . Although the method we proposed uses label matrices instead of a pairwise similarity matrix for matrix decomposition, it cannot effectively solve the large time complexity problem. Therefore, compared with the hash methods that only use label information for learning, the time cost of DSAH is slightly high. In addition, without point-to-point label information, there is no general algorithm to process similarity matrices on all datasets.

5. Conclusions

In this paper, we present a novel cross-modal hashing method, named DSAH, for large-scale cross-modal retrieval. In detail, to enhance the feature representation in the linear model, we handle the nonlinear relations with a kernelization technique. Meanwhile, DSAH incorporates the label information and semantic matrix into the learning process. Therefore, DSAH can obtain more semantic information to improve the discriminative capability of the learned hash codes. However, due to the inevitable noise and subjective factors in labels for large-scale dataset, the 2 , 1 norm is used to sparse the matrix and effectively deal with outliers. In addition, a discrete optimization algorithm is proposed to solve the quantization errors and improve the optimization efficiency. Extensive experiments on two datasets demonstrate the superiority of DSAH on cross-modal retrieval tasks.

Author Contributions

Conceptualization, L.S.; methodology, Z.Y.; software, L.S.; validation, L.S., Z.Y. and J.L.; formal analysis, L.S., Z.Y. and J.L.; investigation, L.S.; resources, L.S. and J.L.; data curation, L.S.; writing—original draft preparation, L.S.; writing—review and editing, Z.Y. and L.H.; visualization, L.S.; supervision, Z.Y.; project administration, J.L.; funding acquisition, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant U2003208, in part by the Science and Technology Plan of Hunan under Grant No. 2016TP1003, and in part by the Key Technology R & D Program of Hunan Province under Grant No. 2018GK2052.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Deng, C.; Chen, Z.; Liu, X.; Gao, X.; Tao, D. Triplet-Based Deep Hashing Network for Cross-Modal Retrieval. IEEE Trans. Image Process. 2018, 27, 3893–3903. [Google Scholar] [CrossRef] [Green Version]
  2. Li, C.; Deng, C.; Li, N.; Liu, W.; Gao, X.; Tao, D. Self-Supervised Adversarial Hashing Networks for Cross-Modal Retrieval. In Proceedings of the 2018 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2018, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4242–4251. [Google Scholar]
  3. Yang, E.; Deng, C.; Liu, W.; Liu, X.; Tao, D.; Gao, X. Pairwise Relationship Guided Deep Hashing for Cross-Modal Retrieval. In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 1618–1625. [Google Scholar]
  4. Wang, J.; Zhang, T.; Song, J.; Sebe, N.; Shen, H.T. A Survey on Learning to Hash. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 769–790. [Google Scholar] [CrossRef] [PubMed]
  5. Wang, G.; Hu, Q.; Cheng, J.; Hou, Z. Semi-supervised Generative Adversarial Hashing for Image Retrieval. In Proceedings of the Computer Vision—ECCV 2018—15th European Conference, Munich, Germany, 8–14 September 2018; pp. 491–507. [Google Scholar]
  6. Xia, R.; Pan, Y.; Lai, H.; Liu, C.; Yan, S. Supervised Hashing for Image Retrieval via Image Representation Learning. In Proceedings of the Twenty-Eighth AAAI Conference on Artificial Intelligence, Québec City, QC, Canada, 27–31 July 2014; pp. 2156–2162. [Google Scholar]
  7. Li, D.; Gong, Y.; Cheng, D.; Shi, W.; Tao, X.; Chang, X. Consistency-Preserving deep hashing for fast person re-identification. Pattern Recognit. 2019, 94, 207–217. [Google Scholar] [CrossRef]
  8. Li, M.; Jiang, Q.; Li, W. Deep Multi-Index Hashing for Person Re-Identification. arXiv preprint 2019, arXiv:abs/1905.10980. [Google Scholar]
  9. Deng, C.; Yang, E.; Liu, T.; Li, J.; Liu, W.; Tao, D. Unsupervised Semantic-Preserving Adversarial Hashing for Image Search. IEEE Trans. Image Process. 2019, 28, 4032–4044. [Google Scholar] [CrossRef]
  10. Song, J.; Yang, Y.; Yang, Y.; Huang, Z.; Shen, H.T. Inter-media hashing for large-scale retrieval from heterogeneous data sources. In Proceedings of the 2013 ACM SIGMOD International Conference on Management of Data, New York, NY, USA, 22–27 June 2013; pp. 785–796. [Google Scholar]
  11. Zhou, J.; Ding, G.; Guo, Y. Latent semantic sparse hashing for cross-modal similarity search. In Proceedings of the 37th International ACM SIGIR Conference on Research & Development in Information Retrieval, Gold Coast, QLD, Australia; 6–11 July 2014, 2014; pp. 415–424. [Google Scholar]
  12. He, K.; Wen, F.; Sun, J. K-Means Hashing: An Affinity-Preserving Quantization Method for Learning Binary Compact Codes. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 2938–2945. [Google Scholar]
  13. Shen, F.; Xu, Y.; Liu, L.; Yang, Y.; Huang, Z.; Shen, H.T. Unsupervised Deep Hashing with Similarity-Adaptive and Discrete Optimization. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 3034–3044. [Google Scholar] [CrossRef]
  14. Weiss, Y.; Torralba, A.; Fergus, R. Spectral Hashing. In Proceedings of the Advances in Neural Information Processing Systems 21, Proceedings of the Twenty-Second Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–11 December 2008; pp. 1753–1760. [Google Scholar]
  15. Zhang, H.; Liu, L.; Long, Y.; Shao, L. Unsupervised Deep Hashing With Pseudo Labels for Scalable Image Retrieval. IEEE Trans. Image Process. 2018, 27, 1626–1638. [Google Scholar] [CrossRef] [Green Version]
  16. Mandal, D.; Chaudhury, K.N.; Biswas, S. Generalized Semantic Preserving Hashing for Cross-Modal Retrieval. TIP 2019, 28, 102–112. [Google Scholar] [CrossRef] [PubMed]
  17. Bronstein, M.M.; Bronstein, A.M.; Michel, F.; Paragios, N. Data fusion through cross-modality metric learning using similarity-sensitive hashing. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 3594–3601. [Google Scholar]
  18. Liu, X.; Nie, X.; Zeng, W.; Cui, C.; Zhu, L.; Yin, Y. Fast Discrete Cross-modal Hashing with Regressing from Semantic Labels. In Proceedings of the 26th ACM international conference on Multimedia, New York, NY, USA, 22–26 October 2018; 2018; pp. 1662–1669. [Google Scholar]
  19. Shen, F.; Shen, C.; Liu, W.; Shen, H.T. Supervised Discrete Hashing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2015, Boston, MA, USA, 7–12 June 2015; pp. 37–45. [Google Scholar]
  20. Luo, X.; Zhang, P.; Wu, Y.; Chen, Z.; Huang, H.; Xu, X. Asymmetric Discrete Cross-Modal Hashing. In Proceedings of the 2018 ACM on International Conference on Multimedia Retrieval, New York, NY, USA, 11–14 June 2018; 2018; pp. 204–212. [Google Scholar]
  21. Liu, H.; Wang, R.; Shan, S.; Chen, X. Deep Supervised Hashing for Fast Image Retrieval. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, 27–30 June 2016; pp. 2064–2072. [Google Scholar]
  22. Yang, Z.; Yang, L.; Huang, W.; Sun, L.; Long, J. Enhanced Deep Discrete Hashing with semantic-visual similarity for image retrieval. Inf. Process. Manag. 2021, 58, 102648. [Google Scholar] [CrossRef]
  23. Yang, R.; Shi, Y.; Xu, X. Discrete Multi-view Hashing for Effective Image Retrieval. In Proceedings of the 2017 ACM on International Conference on Multimedia Retrieval, ICMR 2017, Bucharest, Romania, 6–9 June 2017; Ionescu, B., Sebe, N., Feng, J., Larson, M.A., Lienhart, R., Snoek, C., Eds.; ACM: New York, NY, USA, 2017; pp. 175–183. [Google Scholar]
  24. Wang, J.; Liu, W.; Kumar, S.; Chang, S. Learning to Hash for Indexing Big Data—A Survey. Proc. IEEE 2016, 104, 34–57. [Google Scholar] [CrossRef]
  25. Andoni, A.; Indyk, P. Near-Optimal Hashing Algorithms for Approximate Nearest Neighbor in High Dimensions. In Proceedings of the 47th Annual IEEE Symposium on Foundations of Computer Science (FOCS 2006), Berkeley, CA, USA, 21–24 October 2006; pp. 459–468. [Google Scholar]
  26. Datar, M.; Immorlica, N.; Indyk, P.; Mirrokni, V.S. Locality-sensitive hashing scheme based on p-stable distributions. In Proceedings of the 20th ACM Symposium on Computational Geometry, New York, NY, USA, 8–11 June 2004; pp. 253–262. [Google Scholar]
  27. Huang, Q.; Feng, J.; Fang, Q.; Ng, W.; Wang, W. Query-aware locality-sensitive hashing scheme for lp norm. VLDB J. 2017, 26, 683–708. [Google Scholar] [CrossRef]
  28. Gong, Y.; Lazebnik, S.; Gordo, A.; Perronnin, F. Iterative Quantization: A Procrustean Approach to Learning Binary Codes for Large-Scale Image Retrieval. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2916–2929. [Google Scholar] [CrossRef] [Green Version]
  29. Jin, S.; Yao, H.; Sun, X.; Zhou, S. Unsupervised semantic deep hashing. Neurocomputing 2019, 351, 19–25. [Google Scholar] [CrossRef] [Green Version]
  30. Wu, G.; Han, J.; Guo, Y.; Liu, L.; Ding, G.; Ni, Q.; Shao, L. Unsupervised Deep Video Hashing via Balanced Code for Large-Scale Video Retrieval. IEEE Trans. Image Process. 2019, 28, 1993–2007. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Tang, J.; Li, Z.; Wang, M.; Zhao, R. Neighborhood Discriminant Hashing for Large-Scale Image Retrieval. IEEE Trans. Image Process. 2015, 24, 2827–2840. [Google Scholar] [CrossRef]
  32. Ding, G.; Guo, Y.; Zhou, J.; Gao, Y. Large-Scale Cross-Modality Search via Collective Matrix Factorization Hashing. TIP 2016, 25, 5427–5440. [Google Scholar] [CrossRef] [PubMed]
  33. Liu, H.; Ji, R.; Wu, Y.; Huang, F.; Zhang, B. Cross-Modality Binary Code Learning via Fusion Similarity Hashing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 6345–6353. [Google Scholar]
  34. Zhang, D.; Li, W. Large-Scale Supervised Multimodal Hashing with Semantic Correlation Maximization. In Proceedings of the AAAI Conference on Artificial Intelligence; 2014; pp. 2177–2183. [Google Scholar]
  35. Lin, Z.; Ding, G.; Hu, M.; Wang, J. Semantics-preserving hashing for cross-view retrieval. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3864–3872. [Google Scholar]
  36. Xu, X.; Shen, F.; Yang, Y.; Shen, H.T.; Li, X. Learning Discriminative Binary Codes for Large-scale Cross-modal Retrieval. TIP 2017, 26, 2494–2507. [Google Scholar] [CrossRef] [PubMed]
  37. Wang, D.; Gao, X.; Wang, X.; He, L. Label Consistent Matrix Factorization Hashing for Large-Scale Cross-Modal Similarity Search. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 2466–2479. [Google Scholar] [CrossRef] [PubMed]
  38. Chen, Z.; Li, C.; Luo, X.; Nie, L.; Zhang, W.; Xu, X. SCRATCH: A Scalable Discrete Matrix Factorization Hashing Framework for Cross-Modal Retrieval. IEEE Trans. Circuits Syst. Video Technol. 2020, 30, 2262–2275. [Google Scholar] [CrossRef]
  39. Yang, Z.; Long, J.; Zhu, L.; Huang, W. Nonlinear Robust Discrete Hashing for Cross-Modal Retrieval. In Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, SIGIR 2020, Virtual Event, China, 25–30 July 2020; Huang, J., Chang, Y., Cheng, X., Kamps, J., Murdock, V., Wen, J., Liu, Y., Eds.; ACM: New York, NY, USA, 2020; pp. 1349–1358. [Google Scholar]
  40. Yang, Z.; Raymond, O.I.; Huang, W.; Liao, Z.; Zhu, L.; Long, J. Scalable deep asymmetric hashing via unequal-dimensional embeddings for image similarity search. Neurocomputing 2020, 412, 262–275. [Google Scholar] [CrossRef]
  41. Yang, Z.; Yang, L.; Raymond, O.I.; Zhu, L.; Huang, W.; Liao, Z.; Long, J. NSDH: A Nonlinear Supervised Discrete Hashing framework for large-scale cross-modal retrieval. Knowl. Based Syst. 2021, 217, 106818. [Google Scholar] [CrossRef]
  42. Shen, H.T.; Liu, L.; Yang, Y.; Xu, X.; Huang, Z.; Shen, F.; Hong, R. Exploiting subspace relation in semantic labels for cross-modal hashing. IEEE Trans. Knowl. Data Eng. 2020, 10, 3351–3365. [Google Scholar] [CrossRef]
  43. Jiang, Q.; Li, W. Deep Cross-Modal Hashing. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 3270–3278. [Google Scholar]
  44. Yang, D.; Wu, D.; Zhang, W.; Zhang, H.; Li, B.; Wang, W. Deep Semantic-Alignment Hashing for Unsupervised Cross-Modal Retrieval. In Proceedings of the 2020 on International Conference on Multimedia Retrieval, ICMR 2020, Dublin, Ireland, 8–11 June 2020; Gurrin, C., Jónsson, B.Þ., Kando, N., Schöffmann, K., Chen, Y.P., O’Connor, N.E., Eds.; ACM: New York, NY, USA, 2020; pp. 44–52. [Google Scholar]
  45. Wang, T.; Zhu, L.; Cheng, Z.; Li, J.; Gao, Z. Unsupervised Deep Cross-modal Hashing with Virtual Label Regression. Neurocomputing 2020, 386, 84–96. [Google Scholar] [CrossRef]
  46. Jin, S.; Yao, H.; Sun, X.; Zhou, S.; Zhang, L.; Hua, X. Deep Saliency Hashing for Fine-Grained Retrieval. IEEE Trans. Image Process. 2020, 29, 5336–5351. [Google Scholar] [CrossRef] [PubMed]
  47. Zhan, Y.; Luo, X.; Wang, Y.; Xu, X. Supervised Hierarchical Deep Hashing for Cross-Modal Retrieval. In Proceedings of the MM ’20: The 28th ACM International Conference on Multimedia, Virtual Event/Seattle, WA, USA, 12–16 October 2020; Chen, C.W., Cucchiara, R., Hua, X., Qi, G., Ricci, E., Zhang, Z., Zimmermann, R., Eds.; ACM: New York, NY, USA, 2020; pp. 3386–3394. [Google Scholar]
  48. Zhang, M.; Li, J.; Zhang, H.; Liu, L. Deep semantic cross modal hashing with correlation alignment. Neurocomputing 2020, 381, 240–251. [Google Scholar] [CrossRef]
  49. Kulis, B.; Darrell, T. Learning to Hash with Binary Reconstructive Embeddings. In Proceedings of the Workshop on Learning from Multiple Sources with Applications to Robotics (NIPS), Whistler, BC, Canada, 11 December 2009; pp. 1042–1050. [Google Scholar]
  50. Kulis, B.; Grauman, K. Kernelized locality-sensitive hashing for scalable image search. In Proceedings of the IEEE 12th International Conference on Computer Vision, ICCV 2009, Kyoto, Japan, 27 September–4 October 2009; 2009; pp. 2130–2137. [Google Scholar]
  51. Liu, W.; Wang, J.; Ji, R.; Jiang, Y.; Chang, S. Supervised hashing with kernels. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2074–2081. [Google Scholar]
  52. Gui, J.; Li, P. R 2 SDH: Robust Rotated Supervised Discrete Hashing. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, KDD 2018, London, UK, 19–23 August 2018; Guo, Y., Farooq, F., Eds.; ACM: New York, NY, USA, 2018; pp. 1485–1493. [Google Scholar]
  53. Cheng, M.; Jing, L.; Ng, M.K. Robust Unsupervised Cross-modal Hashing for Multimedia Retrieval. ACM Trans. Inf. Syst. 2020, 38, 1–25. [Google Scholar] [CrossRef]
  54. Jiang, Q.; Li, W. Asymmetric Deep Supervised Hashing. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, (AAAI-18), the 30th innovative Applications of Artificial Intelligence (IAAI-18), and the 8th AAAI Symposium on Educational Advances in Artificial Intelligence (EAAI-18), New Orleans, LA, USA, 2–7 February 2018; pp. 3342–3349. [Google Scholar]
  55. Zhang, Z.; Lai, Z.; Huang, Z.; Wong, W.K.; Xie, G.; Liu, L.; Shao, L. Scalable Supervised Asymmetric Hashing With Semantic and Latent Factor Embedding. IEEE Trans. Image Process. 2019, 28, 4803–4818. [Google Scholar] [CrossRef]
  56. Huiskes, M.J.; Lew, M.S. The MIR flickr retrieval evaluation. In Proceedings of the 1st ACM International Conference on Multimedia Information Retrieval, New York, NY, USA, 30–31 October 2008; 2008; pp. 39–43. [Google Scholar]
  57. Chua, T.; Tang, J.; Hong, R.; Li, H.; Luo, Z.; Zheng, Y. NUS-WIDE: A real-world web image database from National University of Singapore. In Proceedings of the 8th ACM International Conference on Image and Video Retrieval, CIVR 2009, Santorini Island, Greece, 8–10 July 2009. [Google Scholar]
  58. Devlin, J.; Chang, M.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, (Long and Short Papers). Minneapolis, MN, USA, 2–7 June 2019; 2019; Volume 1, pp. 4171–4186. [Google Scholar]
  59. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
Figure 1. The framework of Discrete Semantics-Guided Asymmetric Hashing (DSAH). In our proposed DSAH, we first handle the nonlinear relations in different modalities with the kernelization, then an asymmetric learning scheme is proposed to effectively perform the hash learning process; meanwhile, our proposed method fully considers the label information to enhance the semantic information. In addition, a discrete optimization algorithm is proposed to solve the discrete problems.
Figure 1. The framework of Discrete Semantics-Guided Asymmetric Hashing (DSAH). In our proposed DSAH, we first handle the nonlinear relations in different modalities with the kernelization, then an asymmetric learning scheme is proposed to effectively perform the hash learning process; meanwhile, our proposed method fully considers the label information to enhance the semantic information. In addition, a discrete optimization algorithm is proposed to solve the discrete problems.
Applsci 11 08769 g001
Figure 2. Precision–Recall and P@k curves obtained by different baselines and tested on MIRFlickr and NUS-WIDE datasets.
Figure 2. Precision–Recall and P@k curves obtained by different baselines and tested on MIRFlickr and NUS-WIDE datasets.
Applsci 11 08769 g002

Figure 3. Parameter sensitivity analysis of α and γ on MIRFlickr dataset with 64 bits.
Figure 3. Parameter sensitivity analysis of α and γ on MIRFlickr dataset with 64 bits.
Applsci 11 08769 g003
Figure 4. Parameter sensitivity analysis of α and γ on NUS-WIDE dataset with 64 bits.
Figure 4. Parameter sensitivity analysis of α and γ on NUS-WIDE dataset with 64 bits.
Applsci 11 08769 g004
Figure 5. Convergence curves of DSAH on two datasets with 64 bits.
Figure 5. Convergence curves of DSAH on two datasets with 64 bits.
Applsci 11 08769 g005
Table 1. Notations.
Table 1. Notations.
NotationExplanations
X ( t ) R d m × n Features of heterogeneous modalities
L R c × n Label information
B { 1 , 1 } k × n Hash codes
P t R k × q Hash mapping matrix
D R c × c Projection matrix for label information
V { 1 , 1 } k × n Auxiliary matrix
J b R k × n Auxiliary matrix
d m Dimension of modality X ( m )
nNumber of instances
cNumber of categories
qNumber of Kernelized features
Table 2. Performance comparison on MIRFlickr dataset measured by mAP.
Table 2. Performance comparison on MIRFlickr dataset measured by mAP.
TaskMethod8 Bits16 Bits32 Bits64 Bits128 Bits
I→TCMFH0.55990.56870.56800.56850.5687
FSH0.59110.60160.61490.61940.6242
SCM-seq0.62350.63730.64780.65370.6611
SePH-km0.66410.66850.68180.68300.6873
DCH0.66590.67380.68590.68970.7030
LCMFH0.68210.68120.68870.69090.7034
SRLCH0.70920.71130.72410.72760.7359
DSAH0.71560.72360.74120.74980.7556
T→ICMFH0.56150.56150.56060.56060.5608
FSH0.58690.59790.61140.61860.6251
SCM-seq0.61030.62060.62980.63720.6427
SePH-km0.70330.70760.72120.72930.7348
DCH0.72560.75110.75850.76810.7909
LCMFH0.73510.73080.75440.76890.7806
SRLCH0.74670.76130.77980.78990.8071
DSAH0.75260.77820.80410.81630.8180
Table 3. Performance comparison on NUS-WIDE dataset measured by mAP.
Table 3. Performance comparison on NUS-WIDE dataset measured by mAP.
TaskMethod8 Bits16 Bits32 Bits64 Bits128 Bits
I→TCMFH0.34060.34370.33990.34090.3440
FSH0.36200.37320.38940.40140.4084
SCM-seq0.50130.51200.54220.54880.5483
SePH-km0.52560.55370.56270.56220.5698
DCH0.58400.58080.59070.59320.5843
LCMFH0.59550.61130.62860.63370.6412
SRLCH0.57890.59320.63780.63980.6529
DSAH0.62310.64320.65170.66670.6791
T→ICMFH0.34560.34980.34350.34860.3529
FSH0.36230.37170.38350.39730.4007
SCM-seq0.47090.48360.50670.51410.5161
SePH-km0.61020.64070.65150.66080.6651
DCH0.71060.71030.70980.72600.7223
LCMFH0.67650.71980.73890.76140.7667
SRLCH0.68740.69890.75670.75810.7875
DSAH0.73240.75960.77560.78340.8024
Table 4. mAP scores of different ablated versions of DSAH on MIRFlickr dataset.
Table 4. mAP scores of different ablated versions of DSAH on MIRFlickr dataset.
TaskMethod8 Bits16 Bits32 Bits64 Bits128 Bits
I→TDSAH-Re0.59820.60760.60810.59110.5784
DSAH-Ke0.68320.68720.68980.69130.6906
DSAH-Nm0.69910.72120.72450.72890.7310
DSAH0.71560.72360.74120.74980.7556
T→IDSAH-Re0.55430.55210.57600.57730.5801
DSAH-Ke0.73620.76900.77720.78870.7921
DSAH-Nm0.71110.73040.72970.73580.7439
DSAH0.75260.77820.80410.81630.8180
Table 5. mAP scores of different ablated versions of DSAH on NUS-WIDE dataset.
Table 5. mAP scores of different ablated versions of DSAH on NUS-WIDE dataset.
TaskMethod8 Bits16 Bits32 Bits64 Bits128 Bits
I→TDSAH-Re0.45580.46110.47320.47980.4832
DSAH-Ke0.58420.59110.62870.63340.6499
DSAH-Nm0.60690.60930.64210.64290.6551
DSAH0.62310.64320.65170.66670.6791
T→IDSAH-Re0.45680.46500.47170.48230.4804
DSAH-Ke0.72870.74520.76760.77040.7799
DSAH-Nm0.72250.74070.75890.76210.7697
DSAH0.73240.75960.77560.78340.8024
Table 6. mAP scores on NUS-WIDE dataset using BERT and bag-of-word embeddings of DSAH.
Table 6. mAP scores on NUS-WIDE dataset using BERT and bag-of-word embeddings of DSAH.
TaskMethod8 Bits16 Bits32 Bits64 Bits128 Bits
I→TDSAH0.62310.64320.65170.66670.6791
DSAH-BERT0.62120.64910.65210.66340.6747
T→IDSAH0.73240.75960.77560.78340.8024
DSAH-BERT0.73110.75580.78010.79070.8071
Table 7. mAP scores on NUS-WIDE dataset using deep and shallow representations of DSAH.
Table 7. mAP scores on NUS-WIDE dataset using deep and shallow representations of DSAH.
TaskMethod8 Bits16 Bits32 Bits64 Bits128 Bits
I→TDSAH0.62310.64320.65170.66670.6791
DSAH-Deep0.72320.74210.75190.76300.7793
T→IDSAH0.73240.75960.77560.78340.8024
DSAH-Deep0.65330.69820.71060.72180.7295
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Long, J.; Sun, L.; Hua, L.; Yang, Z. Discrete Semantics-Guided Asymmetric Hashing for Large-Scale Multimedia Retrieval. Appl. Sci. 2021, 11, 8769. https://0-doi-org.brum.beds.ac.uk/10.3390/app11188769

AMA Style

Long J, Sun L, Hua L, Yang Z. Discrete Semantics-Guided Asymmetric Hashing for Large-Scale Multimedia Retrieval. Applied Sciences. 2021; 11(18):8769. https://0-doi-org.brum.beds.ac.uk/10.3390/app11188769

Chicago/Turabian Style

Long, Jun, Longzhi Sun, Liujie Hua, and Zhan Yang. 2021. "Discrete Semantics-Guided Asymmetric Hashing for Large-Scale Multimedia Retrieval" Applied Sciences 11, no. 18: 8769. https://0-doi-org.brum.beds.ac.uk/10.3390/app11188769

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop