Next Article in Journal
Strong Influence of Responses in Training Dialogue Response Generator
Next Article in Special Issue
DOA Estimation in Low SNR Environment through Coprime Antenna Arrays: An Innovative Approach by Applying Flower Pollination Algorithm
Previous Article in Journal
The Wingate Anaerobic Test, a Narrative Review of the Protocol Variables That Affect the Results Obtained
Previous Article in Special Issue
Software Engineering Applications Enabled by Blockchain Technology: A Systematic Mapping Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Top-N Movie Recommendation Framework Based on Deep Neural Network with Heterogeneous Modeling

1
School of Information Science and Engineering, Yanshan University, Qinhuangdao 066004, China
2
The Key Lab for Computer Virtual Technology and System Integration of Hebei Province, Yanshan University, Qinhuangdao 066004, China
3
Key Laboratory for Software Engineering of Hebei Province, Yanshan University, Qinhuangdao 066004, China
4
School of Science, Yanshan University, Qinhuangdao 066004, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Submission received: 28 June 2021 / Revised: 6 August 2021 / Accepted: 6 August 2021 / Published: 12 August 2021

Abstract

:
To provide more accurate and stable recommendations, it is necessary to combine display information with implicit information and to dig out potential information. Existing methods only consider explicit feedback information or implicit feedback information unilaterally and ignore the potential information of explicit feedback information and implicit feedback information, which is also crucial to the accuracy of the recommendation system. However, the traditional Heterogeneous Information Networks (HIN) recommendation ignores the attribute information in the meta-path and the interaction between the user and the item and, instead, only considers the linear characteristics of the user-object often ignoring its non-linear characteristics. Aiming at the potential information acquisition problem from assorted feedback, we propose a new top-N recommendation method MFDNN for Heterogeneous Information Networks (HINs). First, we consider explicit and implicit feedback information to determine the potential preferences of users and the potential features of the product. Then, matrix factorization (MF) and a deep neural network (DNN) are fused to learn independent feature embeddings through MF and DNN, and fully considering the linear and non-linear characteristics of the user-object. MFDNN was tested on several real data sets, such as Movie-Lens, and compared with benchmark experiments. MFDNN significantly improved the hit ratio (HR) and normalized discounted cumulative gain (NDCG). Further research showed that the meta-path bias had an excellent effect on the gain of potential information mining and the fusion of explicit and implicit information in the accuracy and stability of user interest classification.

1. Introduction

A Recommender System (RS), a program that attempts to recommend the most suitable products/services to a user, aims at providing personalized services by retrieving the most relevant information and services from the big data generated on open, private, social, and IoT (Internet of Things) data islands [1]. With the rapid increase in the amount of information, when many users are looking for information about learning [2], movies [3], music [4], popular events [5], and other fields, how to quickly and accurately obtain the information they need most has become a key problem that needs to be solved in the current development of big data. The emergence of the recommendation system provides an opportunity to alleviate this problem [6].
With the development of data-mining algorithms, recommendation systems are used in information retrieval (e.g., Google and Baidu), news feeds (e.g., Toutiao and Google News), e-commerce [7] (e.g., Amazon, Taobao, and Alibaba), and social networks (e.g., Facebook, Tencent, and Twitter) have achieved great success in various fields, effectively alleviating the contradiction between information and users. Recommendation systems due to their multi-domain applicability are among the main topics of scientific interest in recent years [8].
Today, almost every organization leverages Recommender Systems to better understand their customers and to suggest products and services [1]. For example, in the field of e-commerce, recommendation systems are used to personalize products recommended to users; and, in the field of short videos, recommendation systems are designed to personalize and recommend short videos that users will love [9].
The recommendation system mines user interests from big data, captures interest changes in real time [10], quickly feeds back user needs, helps customers complete data access work with simple operation procedures and comprehensive data analysis [11], and formulates or adjusts user recommendation information in a targeted manner. The security and user experience the efficiency of production and life are greatly improved, the process of information interaction, commodity circulation, and industrial asset circulation are accelerated, and social development and the improvement of people’s living standards are effectively promoted.The research by Ricci et al. [9] showed that recommendation systems have been ubiquitous in various fields, including movies, music, travel, video, news, books, and general products.
There are three main combination strategies for hybrid recommendation [12]: pre-fusion, middle fusion, and post-fusion. (1) Pre-fusion refers to the fusion of multiple recommendation algorithms in the process of constructing a recommendation model, combining them into a unified model, performing a feature extraction training model, and then generating recommendation results based on the fusion model. (2) Middle fusion is based on one recommendation algorithm as the framework while fusing another recommendation algorithm. (3) Post-fusion means that each recommendation algorithm is trained separately to generate recommendation results, and finally a combination strategy is adopted to fuse the recommendation results of each recommendation model. Combination strategies that can be adopted include simple voting, linear combination, etc.
Recommendation systems rely on user feedback to evaluate attitudes toward items viewed by users. According to the nature of user feedback [6], this can be divided into explicit user feedback (for example, ratings, likes, and dislikes) or implicit feedback (for example, clicks, plays, and views), that is, display feedback and implicit feedback [7]. Explicit feedback is that the user’s preferences can be directly expressed and exist in a way that makes it easy to obtain their preferences. Implicit feedback refers to the user’s preference behavior information expressed in an indirect way rather than directly [13].
Explicit feedback data has the ability to express user preferences and behaviors more accurately; however, in real life, it is difficult to obtain representative and sufficient amount of explicit feedback information based on users [14]. At the extreme, feedback data can be very scarce and not easy to obtain in many application scenarios. Implicit information data is easy to obtain, and the amount of information is relatively large; however, its information cannot accurately express user preferences. Therefore, if we make full use of the advantages of the two types of data, we can achieve a good recommendation effect [8].
For the first time, ROber and Yen [15] proposed the theoretical idea of combining explicit feedback data with implicit feedback data. Nathan N. Liu et al. [16] proposed a matrix factorization model, which used different weights for implicit feedback data and explicit feedback data for learning modeling. The main idea of the matrix factorization model is to treat purchased and viewed commodities as implicit feedback data and mark them as “1”, to mark other types of commodities as “0” for processing, and then combine them with display feedback data, before combining explicit feedback and implicit feedback.
Weike [17] first clustered user sets and item sets to propose a learning model. GaiLim [18] proposed a personalized ranking model that combined explicit and implicit feedback, which was implemented by optimizing the evaluation index ERR (Expected Reciprocal Rank).
With the rapid development of deep learning, the application of deep learning in recommendation models is gradually increasing. Ding et al. [19] proposed a friend recommendation model based on a Bayesian ranking deep neural network, which converted the recommendation problem into a ranking problem. The recommendation method based on deep learning has been successfully applied to label recommendation [20] and POI [21] recommendation, and different neural network structures have been proposed, such as multi-layer perceptron (MLP), convolutional neural networks (CNN), and recurrent neural networks (RNN) [6].
A deep neural network can effectively simulate nonlinearity in the data through nonlinear activation. Some are also used to transform recommendation problems into classification problems. Although deep learning has been widely used in recommendation methods and recommendation systems, the research on recommendation methods based on deep learning is still in the development stage [10].
Recently, some researchers have realized the importance of heterogeneous information for recommendation. Heterogeneous Information Networks (HINs) effectively integrate more information and form a new trend in the development of data mining. A large amount of user information can be obtained to make the content of the recommendation system more diverse, including academics, commodities, friends, music, services, etc. [22]. In addition to traditional recommendation methods, a large number of new recommendation methods have also been generated, such as social network-based recommendation methods, context awareness recommended methods of the internet [23], and location-based recommendations [24].
However, these recommendation methods based on heterogeneous information also have some challenges: (1) Massive amounts of HIN data hide the objects’ comprehensive and detailed information. Hence, mining and analyzing valuable information for HIN recommendations is a key challenge. (2) The rapid expansion of HINs generates increasing amounts of data, such as a wide variety of user features. How to take advantage of these features to build a unified top-N recommendation model is a substantial problem. (3) It is difficult to combine and measure all of the features of objects to produce HIN recommendations. Considering all of the features may require a significant amount of time and cause an over-fitting problem. (4) Transmission delays, energy saving issues, data redundancy, and inaccuracy of data transmission during data transmission are also issues that need to be resolved [25]. As a result, selecting the most relevant one from the recommendation results among all the features of objects in HINs is challenging.
Based on user–item history information, rating prediction models predict the specific rating for an item given by a user [26]. In practical applications, merchants are concerned about whether users will buy an item, which means telling whether a user will watch a movie is more consequential than predicting the rating value that the user may give after watching the movie.
This study mainly considers a bipartite network, a special type of heterogeneous information network, for generating top-N recommendations [27]. Existing user–item data recommendation methods mostly consider user–item implicit feedback and ignore the user preference characteristics behind the explicit data. Here, the explicit feedback data refers to the user–item rating information, and the implicit feedback refers to whether user–item interaction information exists or not.
To obtain the user’s preference information more comprehensively, this study considers both the explicit and implicit feedback from user–item interactions for mining the users’ potential preferences and the underlying features of items [28]. In order to improve the performance of the recommendation algorithm of the heterogeneous information network, the recommendation model is constructed by fusing matrix factorization (MF) [29] and a deep neural network (DNN) [19], which, respectively, obtain the explicit and implicit feedback prediction results.
The explicit and implicit feedback prediction results are combined to generate the top-N recommendations [30,31,32]. Two bias factors are introduced to consider the characteristics of the user–item data in explicit feedback information. More specifically, we first consider both explicit and implicit feedback data; the explicit and implicit feedback information is separately trained as input to better mine the potential information behind the user–item rating meta-path information. Here, explicit feedback data refers to user–item ratings, and includes both meta-path attribute information and the characteristics of objects; implicit feedback refers to the user–item relationship data.
Then, the MF and DNN models are merged to form the relationship between the user–item rating meta-path and the attribute value information. As they learn embedded features independently, MF and DNN can fully consider the linear and nonlinear user–item features, and respectively train the explicit and implicit feedback data in order to obtain the corresponding output. Subsequently, by combining the explicit and implicit feedback prediction results, the top-N items are recommended to the target user.
Finally, we use the MovieLens dataset to verify the model and apply leave-one-out to further evaluate the model. The performance of the method is evaluated using the hit ratio (HR) and normalized discounted cumulative gain (NDCG) evaluation metrics. The proposed method was found to outperform the traditional recommendation model and state-of-the-art recommendation methods.
The contributions of this paper are summarized as follows.
  • We exploit both explicit and implicit feedback information to obtain the user’s preference information and the underlying characteristics of the item based on the meta-path selection results. Additionally, in order to obtain explicit feedback information, two bias factors are introduced according to the individual characteristics of the user–item information.
  • We fuse MF and DNN to mine the potential features of users and items from both linear and nonlinear perspectives. MF and DNN learning are independently embedded to better capture user preference information and the potential feature information of items.
  • Using the leave-one-out evaluation method, we combine explicit and implicit feedback results to obtain the top-N recommendation list for target users and adopt the HR and NDCG metrics to evaluate the proposed model.
The remainder of this paper is organized as follows: We briefly outline the related work in Section 2. We provide the problem definition and explain the proposed architecture in Section 3. Section 4 shows and discusses the experimental results that validate our model. Finally, we conclude this paper in Section 5.
The notations used in this paper are summarized in Table 1:

2. Related Work

An increasing number of researchers have focused on HIN recommendations with different types of objects or relations [33]. Since HINs were first proposed in [34], many HIN recommendation methods have been proposed. In these works, similarity measurements are vitally important and fundamental, and the most popular method is path-based. For example, a meta-path associated with top-N similarity measurement was proposed in [34,35] proposed a recommendation method based on personalized semantics to predict users’ ratings of items, and [36] proposed symmetric measurements on arbitrary meta-paths. Random-walk-based methods are usually used to mine the paths, weigh the paths, and compute the closeness or relevance between two nodes in a HIN [37].
Random walks in the connected components of the graph assume the properties of Markov Chains (steady-state distribution, irreducibility, etc.) [38,39]. However, these traditional HIN techniques ignore the value of link attributes; as a result, the meta-path cannot accurately capture the relationship between objects [40]. More recently, other strategies have been proposed to alleviate this shortcoming. A unified and flexible personalized sorting framework, MFPR, was proposed in [41]; this framework combines explicit feedback with multiple implicit feedback. In [42], a unified model fusing generalized matrix factorization and multilayer perception was proposed. In [43], a collaborative filtering recommendation method in view of heterogeneous relations was proposed.
Among the works described above, we must compare MFDNN against the model proposed in [44], because it is not only a state-of-the-art HIN recommendation method but also very similar to our model. DeepMF [44] performs click-through rate (CTR) prediction by combining the recommendation ability of factorization machines with the feature learning ability of deep learning, and simultaneously learns the low-order and high-order feature interactions from the original features of the input.
Compared with DeepMF, MFDNN has three main differences: (1) In MFDNN, MF and DNN learn embedded features separately, while DeepMF shares the same raw input feature vector. (2) The input layer of MFDNN combines user–item explicit and implicit feedback, while the input layer of DeepMF is a one-shot encoding of each feature field (e.g., gender, and location). (3) The output of MFDNN is a top-N recommendation list, while DeepMF aims to predict the value of ratings.
There are many other HIN recommendation algorithms, including collaborative filtering [45] and content-based recommendation [46]. These traditional methods were extensively used in the early phases of HIN. However, the former is not applicable to high-dimensional data, and there is a cold-start problem [47], while the latter takes fewer attributes into consideration.
Deep neural networks (DNN) have demonstrated breakthroughs in data mining, e.g., voice recognition [48], image labeling [49,50], and text classification [51,52,53]. Deep learning-based methods, which can learn a large-scale nonlinear network structure and obtain deep feature representations of users and items, have proven effective in recommendation tasks [54,55,56]. Convolutional neural networks have a powerful ability to learn feature representations and have the potential to learn sophisticated feature interactions [57,58].
BayDNN was proposed in [19] as a Bayesian personalized ranking deep neural network model for social network friend recommendations; in this model, the recommendation problem is regarded as a ranking problem. The method described in [59] adopts a simple pre-training strategy using a four-layer neural network for link prediction. In [21], a deep content-aware point-of-interest (POI) recommendation (DCPR) algorithm was proposed; broad learning from multiple sources of information is utilized to solve the problem. Based on the above studies, we found that deep learning-based recommendation methods are still in their infancy, and MFDNN effectively improves the accuracy of HIN recommendation.
In the past few decades, numerous researchers have focused on designing and implementing top-N recommendation methods; however, these methods only consider the direct relations between pairs of items to compute the similarities needed for constructing recommendation frameworks. In fact, a high-order information and neighborhood-based method was proposed to merge high-order information earlier in the process; however, it did not significantly improve performance.
The sparse linear method (SLIM) was proposed in [60]; this method aggregates users’ purchase/rating profiles to generate recommendation results. However, it can only model the relationship between items co-purchased by at least one user. To address the limitations of SLIM, LorSLIM [61], which introduces a low-rank structure, was proposed. Low-rank assumptions are usually driven by factor models. HOSLIM was proposed in [62], which revisited the problem of using higher-order information rather than low-rank information.

3. Our Approach: MFDNN

3.1. Problem Definition

In this subsection, we first provide the related preliminary definition and then provide a formal problem definition.
Definition 1.
Heterogeneous information network (HIN). HINs were first defined in [34]. A directed graph G = V , E is defined to present an information network, where V is the set of objects, E is the set of relations, the object-type mapping function is ϕ : V A , and the relation-type mapping function is ψ : E R . A network is called a heterogeneous information network when the types of objects | A | > 1 or the types of relations | R > 1 | ; otherwise, it is a homogeneous information network.
A bibliographic information network [63] is a typical HIN that contains three types of objects: author, venue, and paper, and two types of relations: publish and write. Other examples of HINs are shown in [64]. We mainly focus on the bipartite network—a special HIN that has two types of objects.
Definition 2.
Bipartite Network. A bipartite network is a special HIN that has two types of objects.
Problem 1.
U = u 1 , u 2 , , u m is a user set of size m, and I = i 1 , i 2 , , i n is an item set of size n. We first analyze and select a reasonable meta-path that can help find the most similar user or item according to the meta-path. An example of a selected meta-path is shown in Figure 1.
In this study, we consider the meta-path link information as well as the attribute information of the user and item. According to the U I rating meta-path, we define the user–item interaction matrix Y R m × n and the user–item rating matrix Y + R m × n according to the historical rating record as Equations (1) and (2):
y u i = 1 if interaction ( user u , item i ) is observed ( score > 2 ) ; 0 otherwise .
y u i + = Y u i + the rating ( user u , item i ) ; 0 otherwise .
Here, a value of 1 for y u d + denotes that u and i have an interaction; however, this does not necessarily mean that u actually likes i. Similarly, a value of 0 for y u d also does not indicate that user u dislikes item i; perhaps user u is not aware of item i at all. In other words, observed relations reflect the users’ preferences on items, while unobserved relations can result from missing data. For example, when shopping online, the rating value of an item is affected by factors other than the item itself, such as delivery speed and service attitude.
In such cases, the final rating may not indicate whether the user likes the item. However, if the user buys the item, it is certain that one of the characteristics of the item attracts the user. Therefore, we simply conclude that the level of the rating reflects the preferences of the user. It becomes a challenge to learn users’ intentions from the historical rating record since it contains various noisy data indicating users’ preferences. We often cannot obtain explicit feedback information directly, and the data is sparse.
In contrast, we can easily obtain implicit feedback information, and the data covers most users and objects; thus, it can mitigate the problem of sparse data to some extent [65,66,67]. We obtain a top-N recommendation list via modeling with a recommendation algorithm according to historic explicit and implicit feedback information.

3.2. Top-N Recommendation Architecture

In general, a user’s preferences will not change significantly over a relatively short period of time. Therefore, our goal is to combine the explicit and implicit feedback information by fusing MF and DNN to predict the missing user–item interaction rating value y ^ u i and sort the rating values to obtain the top-N recommendation list. The main framework is shown in Figure 2.
As shown in Figure 2, we first construct a user–item rating matrix and a user–item relation matrix, here, for the explicit information matrix construction, that is, the user–item rating matrix, we fill in the corresponding position according to the historical scoring record. For those without scoring, we fill in with 0; For the construction of the implicit information matrix, that is, the user–item relation matrix, according to the historical scoring record, we fill in the position corresponding to the user’s score greater than 2 with 1, and the other positions with 0; the user–item relation matrix and user–item rating matrix are made in different ways.
Then, we run the MFDNN model to obtain the explicit and implicit feedback prediction results; finally, we combine the explicit and implicit results to obtain the top-N recommendation list for target users. Here, we can choose the weighted average method (WAM) or simply sum the explicit and implicit prediction results. For explicit feedback prediction, we choose the parameters by minimizing the value of the cross-entropy loss between y u i + and y ^ u i + , which is expressed by the formula in Equation (3):
L + = ( u , i ) Y l o g ( y ^ u i + + b u + b i ) ( u , i ) Y l o g ( 1 y ^ u i + b u b i ) .
This is the same as implicit feedback—the only difference is that explicit feedback information considers two individual bias factors b u and b i , where y ^ u i + denotes the explicit prediction results. By minimizing Equation (3), we can obtain the best recommendation list according to the explicit feedback. Additionally, we can obtain the results of b u and b i according to Equations (4) and (5):
b u = y ^ u i + r ¯ u # i ,
b i = y ^ u i + r ¯ i # u .
where y ^ u i + is the rating value that user u has given to item i, r ¯ u is the average of the ratings given by user u, and # u is the number of items that u has rated. Similarly, r ¯ i is the average rating of item i, and # d is the number of users that have rated i. By adding a regularization term to optimize the target loss function, the target loss function of the regular term is introduced as Equation (6):
L + = ( u , i ) Y l o g ( y ^ u i + + b u + b i ) ( u , d ) Y l o g ( 1 y ^ u i + b u b i ) + λ 2 ( y ^ u i + 2 + b u 2 + b i 2 ) .
where λ is the regularization parameter.

3.3. Framework of MFDNN

In this section, we describe the design of MFDNN, a recommendation architecture based on MF and DNN. MF can fully consider the linear relation between users and items, on the other hand, DNN can fully consider the nonlinear features between users and items. The framework of MFDNN is shown in Figure 3.
The input data includes user–item explicit and implicit feedback, as shown in Figure 3. The explicit feedback is the user–item rating matrix constructed based on the meta-path, and the implicit feedback implies a user–item relation matrix. They are trained in order to obtain the corresponding results: Y ^ and Y ^ + . The embedding layers are independently trained for MF and DNN. For MF, the user is embedded as p u F , and the item is embedded as q i F , For DNN, the user is embedded as p u I , and the item is embedded as q i I .
The subsequent user and item embedding can be viewed as a potential vector for describing users and items in the context of a latent factor model. The embedding layer is a fully connected layer that maps the coefficient representation of the input layer to a dense vector. The MF model and DNN model separately train the result and, finally, fuse the results by an activation function. This is shown as Equation (7):
y ^ u i = σ ( y ^ u i , M F + y ^ u i , D N N ) .
Here, we select the sigmoid function as the activation function because of the probability of y ^ u i [0,1]. We note that y ^ u i , M F and y ^ u i , D N N are trained independently in the model.

3.4. Implementation of MFDNN

A linear combination of potential features of a user and an item can be learned by matrix factorization, and p u and q i are used to represent the potential vectors of u and i, respectively. The matrix factorization estimates the inner product of p u and q i as the prediction function value, as shown in Equation (8):
y ^ u i = p u T q i = k = 1 K p u k q i k .
where K denotes the dimensions of latent space. However, using a simple inner product to estimate complex user–item interactions in the low-dimensional latent space limits the expression of MF and affects the generalization ability of the model. Thus, we define the mapping function of the first layer of MF as Equation (9):
ϕ ( p u F , q i F ) = p u F q i F .
where ⊙ denotes the element-wise product of vectors; the output of MF is given by Equation (10):
y ^ u i , M F = α o u t ( h T ϕ ( p u F , q i F ) ) .
where α o u t is an activation function. In consideration of convergence speed, we used the ReLU (rectified linear unit) function as the activation function, which is simply defined as max(0, x); h T is the weight vector. For DNN, we first obtain the first layer results by processing the embedding layer using Equation (11):
f 1 = ϕ 1 ( p u I , q i I ) .
In the same manner, the second layer results are obtained using Equation (12):
f 2 = α 2 ( W 2 T f 1 + b 2 ) .
where W 2 T and b 2 are the weight matrix and biased vector, respectively, and α 2 is the activation function. The results of the N-th layer are obtained using Equation (13):
f N = α N ( W N T f N 1 + b N ) .
According to Equations (11)–(13), we obtain the final DNN prediction result using Equation (14):
y ^ u i , D N N = α ( W | H | + 1 · p u I , q i I T + b | H | + 1 ) .
where α is the activation function, H is the number of hidden layers, and W and b are the weight matrix and biased vectors, respectively. Here, we chose the ReLU function as the activation function. The sigmoid activation function maps the output of each neuron to the (0,1) interval. This may hamper the performance of the model, and it is likely to cause an over-fitting problem; that is, when the output approaches 0 or 1, the neuron stops learning.
Although Tanh mitigates the problem of the sigmoid to a certain extent, the result is a scaled version of the sigmoid function. Therefore, the ReLU activation function was selected for the model. The ReLU activation function avoids over-fitting and supports sparse data so that the model does not overfit. The explicit and implicit prediction results are generated by MFDNN. After obtaining the Y ^ + and Y ^ by executing MFDNN, we can obtain the final user–item prediction results according to Equation (15):
Y ^ = ω 1 Y ^ + ω 2 Y ^ + ( 0 ω 1 1 , 0 ω 2 1 ) .
where ω 1 + ω 2 = 1 , ω 1 is the weight of implicit feedback and ω 2 is the weight of explicit feedback. As for the best recommendation list for target users, we find the optimal weights by minimizing the objective function using Equation (16):
L = ( u , i ) Y l o g y ^ u i ( u , i ) Y l o g ( 1 y ^ u i ) + λ 2 ω 1 2 + ω 2 2 = ( u , i ) Y Y y u i l o g y ^ u i + ( 1 y u i ) l o g ( 1 y ^ u i ) + λ 2 ω 1 2 + ω 2 2 .
In the network structure, each layer employs fewer neurons in succession. By using a small number of hidden units at the upper layer, more abstract features can be learned from the data. For higher layers, the scale is reduced compared with the previous layer.
In addition, we utilized the dropout technique to alleviate the over-fitting problem. We chose the Adam algorithm [68] to train the model from scratch; this yielded faster convergence than SGD, which was important because we were unable to pay more attention to tuning the learning rate. The main steps of MFDNN are shown in Table 2:

4. Experiments

4.1. Experimental Setup

4.1.1. Datasets

In this study, we used MovieLens 1m and Netflix, two benchmark datasets commonly used for testing recommendation systems, to evaluate the proposed model. These datasets do not require additional processing. We obtained the last interaction for every user, and then randomly selected 100 movies that the user had not interacted with. Table 3 shows the statistics for these two datasets.
According to the definition of HIN and the content of [35,69], MovieLens 1m and Netflix are typical examples of HINs; the network pattern of these datasets are shown in Figure 4. Figure 4 shows that there are four types of objects: users, movies, actors, and directors. In this study, we only considered users and movies, which have the relations “rating” and “rated by.” If two users often view the same movies, we can simply conclude that they have similar interests or preferences. Using the implicit and explicit feedback based on user–movie relations, we evaluated the performance of the proposed model.

4.1.2. Evaluation Metrics

We adopted the HR (hit ratio) and NDCG (normalized discounted cumulative gain) metrics to evaluate the performance of MFDNN. HR: The probability that the user clicks or browses the recommended item. NDCG: Measures the quality of the ranking, which considers the ranking of the ratings; it is defined as
N D C G = D C G I D C G .
The NDCG value of the first k ratings is defined as
N D C G = D C G k I D C G k .
where DCG is the discount cumulative gain. The DCG value of the first k ratings is defined as:
D C G k = i = 1 k 2 r e l i 1 l o g 2 ( i + 1 ) ) .
where r e l i denotes the ith rating. IDCG denotes the ideal DCG, that is, the recommendation list sorted according to the value of ratings from high to low.

4.1.3. Baseline Methods

In this section, we aim to explain how our proposed MFDNN outperformed the existing top-N recommendation methods. We compare the MFDNN with the following representative methods in addition to two state-of-the-art recommendation methods (DMF [29] and NCF [42]) and three HIN-based methods (HeteCF [43], HeteMF [70], and CMF [71]).
DMF (Deep Matrix Factorization): A new matrix decomposition model based on a neural network structure. It uses the user–item explicit feedback matrix as input and learns a common low-dimensional space of objects via a deep learning framework.
NCF (Neural Collaborative Filtering): NCF can be used to express and generalize matrix decomposition under its framework. In order to use nonlinear enhanced NCF modeling, a multilayer perceptron is used to learn user–item interactions. NCF learning emphasizes the probability model of the binary properties of implicit data. It unifies the linear modeling advantages of MF and the nonlinear advantages of MLP to model the potential structure of user-projects.
HeteCF (Heterogeneous network Embedding based approach for Recemendation): the HeteCF method is based on a social collaborative filtering algorithm using heterogeneous relations.
HeteMF (Dual Similarity Regularization): the HeteMF method is based on the HIN recommendation method through combining user ratings and item similarity matrices.
CMF (Dual Similarity Regularization): The CMF method is based on the coupled matrix factorization recommendation method integrating user couplings and item couplings into the basic MF model.

4.2. Parameters Analysis

In order to determine the best learning rate, we evaluated the MFNN model using learning rates of 0.0001, 0.0005, 0.001, and 0.005. The results are shown in Figure 5. Figure 5 shows that there were fewer differences when the learning rate was 0.0005, 0.001, or 0.005. We further analyzed the HR and NDCG values to select the best rate. According to HR, the performance was better when the learning rate was 0.001 rather than 0.005. It is also clear that the value was higher when the learning rate was 0.001 rather than 0.0005 during early training. The trends of NDCG values were similar to those of HR. Thus, we concluded that a learning rate of 0.001 was best in terms of the experimental results.
We used similar methods to determine that the best batch size was 256. We also considered how the number of hidden layers impacts the recommendation performance, in order to determine whether deeper was actually better. We trained the DNN model with 1, 2, 3, and 4 deep layers. The HR and NDCG values achieved with the different numbers of deep layers are shown in Figure 6. As shown in Figure 6, the DNN performed best when the number of hidden layers was 3. In line with these results, we reached the following conclusion: it is not correct to assume that the greater the number of hidden layers, the better the performance is, or vice versa.
Thus, we selected the most reasonable and best number of hidden layers to implement the MFDNN. In addition, the number of embedding factors for MF affected the recommendation performance. We conducted tests using different numbers of embedding factors in order to find the optimal number; the results are shown in Figure 7. Figure 7 shows that, when the number of embedding factors was 32, the MF performance was the best according to HR and NDCG; thus, we set the number of factors to 32.
Dropout, a technique for addressing overfitting, refers to the probability that a neuron is kept in the network [72]. We set the dropout rate to be 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, and 0.9. As shown in Figure 8, MFDNN was able to achieve its best performance when the dropout rate was 0.6 according to the HR and NDCG metrics. The results illustrate that the robustness of MFDNN was strengthened by adding reasonable randomness.

4.3. Performance and Comparison

The MFDNN recommendation model combines user–item explicit and implicit feedback. The recommendation model combines MF and DNN, which learn embedded features independently; however, the model only merges them in the final output layer through the activation function. In this experiment, we compared and analyzed results in all these aspects. The selected top-N value was taken as N = 10, and the dataset was MovieLens 1m.
(1) Explicit and implicit feedback information
In order to measure the recommendation performance achievable with explicit and implicit feedback information, the explicit and implicit feedback data were separated out for experiments. Specifically, we first removed the implicit feedback information, denoted by MFDNN+, and then we removed the explicit feedback information, denoted by MFDNN-. The experimental results for MFDNN+, MFDNN-, and MFDNN are shown in Figure 9. As shown in Figure 9, the most accurate recommendations were generated using the explicit and implicit feedback data.
According to the HR value, MFDNN provided the highest performance. In the small batch of data before training, the performance of MFDNN- was better than that of MFDNN+, and subsequently the performance of MFDNN+ was better. According to the NDCG value, MFDNN had the best performance, followed by MFDNN- and, finally, MFDNN+. These experiments verified that the combination of explicit and implicit feedback data provided better recommendation performance.
These results also illustrate the shortcomings of simply considering explicit and implicit feedback data individually. Explicit and implicit feedback data reflect the user’s preferences or item feature information from a certain aspect. In order to generate accurate recommendations, it is necessary to fully consider the user’s interests or preferences and the potential features of the item itself.
The performance was further verified by considering two bias factors in the explicit feedback information; the results were compared with those from the method that does not consider these two factors (MFDNN–). The experimental results are shown in Figure 10. As shown in Figure 10, when considering two bias factors, the performance of the recommendation model was improved. Although the performance improvement was not very large, the overall performance improved to some extent. Further, in practical applications, different users have different score preferences.
For users who tend to give positive reviews even when they are not satisfied with an item, the score values will not be too low, and the ratings from this type of user are generally high. Conversely, users who are more stringent may give a lower rating even if they are somewhat satisfied with an item; thus, their score values will not be particularly high. In addition, if an item is inexpensive, its overall rating will be higher, and if the item is of poor quality, its overall rating will be lower.
Based on the above analysis, it is meaningful to consider the explicit feedback score preference information, and in theory this can improve the performance of the recommendation algorithm. Although the model considers the bias factor, its performance improvement is not obvious. In subsequent research, it will be necessary to learn more suitable bias factors to improve the recommendation performance [28,73].
(2) Recommendation model based on MF and DNN
MF is a linear model that can mine linear user–item correlation features. On the other hand, DNN is a nonlinear model that can mine the potential nonlinear relationship characteristics of user–item data. In order to verify the recommendation performance of MFDNN, MF and DNN were separately trained as recommendation models. The HR and NDCG values for MF, DNN, and MFDNN are shown in Figure 11. As shown in Figure 11, MFDNN demonstrated the best performance. According to the HR, MFDNN had the best performance, MF and DNN were similar, and their trends were consistent. According to the NDCG value, MFDNN performed best, followed by MF and, finally, DNN. The results of the comparison experiments verify that MFDNN provided the best recommendation performance.
In order to evaluate the recommendation performance of MF and DNN, which learn embedded features independently, separate training processes were employed to facilitate the sharing of the best MF and DNN embedding layers; these are, respectively, denoted as MFDNN: (share MF) and MFDNN: (share DNN). The HR and NDCG values for MFDNN: (share MF), MFDNN: (share DNN), and MFDNN are shown in Figure 12. Figure 12 shows that, although there are small batches of data indicating that the performance was the best when embedding layers were shared, the overall trend shows that MFDNN performed the best.
According to the HR, MFDNN provided the best overall performance, while MFDNN: (share MF) performed better than MFDNN: (share DNN) at the early epochs; subsequently, the performance of MFDNN: (share MF) decreased. According to the NDCG, the overall trend is consistent with HR. Optimal performance was achieved when MF and DNN learned embedding independently.
(3) Comparison with the baseline methods
In order to evaluate the performance of MFDNN over other recommendation models, we trained two state-of-the-art recommendation models (DMF and NCF) and three HIN-based methods (HeteCF, HeteCF, and CMF) separately. The HR and NDCG values for MFDNN and the two baseline methods are shown in Figure 13 (MovieLens 1m) and Figure 14 (Netflix). The HR and NDCG values for MFDNN and the three HIN-based baseline methods are shown in Figure 15 (MovieLens 1m) and Figure 16 (Netflix). The best performance results of each baseline method are shown in Figure 17.
Figure 13, Figure 14, Figure 15 and Figure 16 show that MFDNN achieved the highest performance. According to the HR, MFDNN performed the best, followed by the two baseline methods NCF, DMF, and the HIN-based baseline methods, HeteCF, HeteMF, and CMF. According to the NDCG value, MFDNN performed the best, followed by the two baseline methods, NCF, DMF, and the HIN-based baseline methods, HeteCF, HeteMF, and CMF. Further analysis indicates that the performance of NCF was better than that of DMF. As NCF uses implicit feedback data and DMF uses explicit feedback data, the results are consistent with the experimental results shown in Figure 9. Table 4 shows the best HR and NDCG values for the MFDNN and baseline methods.
Table 4 clearly shows that MFDNN outperformed the two baseline methods. In terms of HR, MFDNN achieved an average 5.4% improvement over DMF, an average 2.3% improvement over NCF on the MovieLens 1m datasets. In terms of HR, the two compared methods underperformed MFDNN by an average of 2.75% in terms of HR on the Netflix datasets. The five methods underperformed MFDNN by an average of 3.75% in terms of NDCG. The performance improvements provided by MFDNN are statistically significant according to these results.
(4) Selection of N of Top-N
In the above experiment, the value of N was 10; however, with different N values, the HR and NDCG values will also be different. We selected N values of 5, 10, and 15 to train the model and chose the best performance in the training epochs. The HR and NDCG values for different top-N values are shown in Table 5:
Table 5 clearly shows that MFDNN performed the best when N was 15; however, we cannot conclude that the larger the value of N, the better the performance. HR relates to whether a test item is in the recommendation list; therefore, for this metric, the larger the value of N, the better the performance. The NDCG relates to the order of the test items in the recommendation list.

4.4. Discussions

In this section, we further analyze the architecture of MFDNN and discuss the experimental results to illustrate the performance of MFDNN.
(1) Table 4 clearly shows that MFDNN outperformed the other baseline methods, which indicates that MFDNN improved the top-N recommendation performance of HINs to an extent.
(2) We see that, by combining explicit and implicit feedback information, we can improve the recommendation performance significantly (+3.0% and +2.5%, respectively, for MFDNN+ and MFDNN- in terms of HR, and +1.9% and +1.0%, respectively, for MFDNN+ and MFDNN- in terms of NDCG).
(3) We also found that configuring MF and DNN to learn embedding factors independently improved the recommendation performance (+0.6% and +0.7%, respectively, for MFDNN+ and MFDNN- in terms of HR and +0.8% and +0.9%, respectively, for MFDNN: (share MF) and MFDNN: (share DNN) in terms of NDCG). Although it is not obvious, learning embedding factors independently can improve the performance to a certain extent.
Although MFDNN provided significantly improved performance compared with baseline methods, there is room for further improvement in terms of the metrics. We will continue our best efforts to improve the performance of MFDNN.

5. Conclusions

In this work, we explored the information behind the meta-path in the binary network. We designed a new framework MDFNN. The model considers both the explicit feedback information and implicit feedback information of the user-object. It fully captures the preference information of the object based on the meta-path and merges the obtained information into the MFDNN to mine the user–item linear and non-linear characteristics. We proved the rationality and effectiveness of MFDNN through a large number of experiments on various data sets and achieved improvements to existing models. In our comparative experiments, MFDNN was superior to the five models in terms of HR and NDCG.
Although MFDNN improved the recommendation performance, there are still other factors that we should consider. On the one hand, we excavated certain potential features, and there are other available features that have not been excavated, such as other semantic information. On the other hand, we still need to improve the operating efficiency. This takes longer to run on a data set with a large amount of data. This work explores the potential of using explicit and implicit information to mine the potential information of the meta-path in recommendation.
In addition to the meta-path information used in this article, there is other potential information in the real scenes, such as the mining of semantic information of gender, age, and time; another exciting direction is to apply the model to other realities in the scene or use other structural information of the real scene, such as social networks and project context. In social networks, through the combination of social network information, we can also investigate how social influence affects recommendations.

Author Contributions

Conceptualization, J.G., Y.S., and X.Z.; Project administration, X.Z. and J.G.; methodology, X.Z.; software, X.Z.; validation, Q.L., C.W. and Z.Z.; formal analysis, Q.L.; investigation, Y.S.; resources, S.W.; data curation, X.Z.; writing—original draft preparation, Y.S.; writing—review and editing, X.Z.; visualization, Z.Z. and S.W.; supervision, Y.S.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Beheshti, A.; Yakhchi, S.; Mousaeirad, S.; Ghafari, S.; Goluguri, S.; Edrisi, M. Towards cognitive recommender systems. Algorithms 2020, 13, 176. [Google Scholar] [CrossRef]
  2. Jing, X.; Tang, J. Guess You Like: Course Recommendation in MOOCs. In Proceedings of the IEEE/WIC/ACM International Conferences on Web Intelligence, Leipzig, Germany, 23 August 2017; pp. 783–789. [Google Scholar]
  3. Diao, Q.; Qiu, M.; Wu, C.Y.; Smola, A.J.; Jiang, J.; Wang, C. Jointly modeling aspects, ratings and sentiments for movie recommendation (JMARS). In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, New York, NY, USA, 24 August 2014; pp. 193–202. [Google Scholar]
  4. Su, J.H.; Chang, W.Y.; Tseng, V.S. Integrated Mining of Social and Collaborative Information for Music Recommendation. Data Sci. Pattern Recognit. 2017, 1, 13–30. [Google Scholar]
  5. Liu, Y.; Peng, H.; Guo, J.; He, T.; Li, X.; Song, Y.; Li, J. Event detection and evolution based on knowledge base. Proc. Kbcom 2018, 2018, 1–7. [Google Scholar]
  6. Rizkallah, S.; Atiya, A.; Shaheen, S. New Vector-Space Embeddings for Recommender Systems. Appl. Sci. 2021, 11, 6477. [Google Scholar] [CrossRef]
  7. Zdziebko, T.; Sulikowski, P. Monitoring Human Website Interactions for Online Stores. In New Contributions in Information Systems and Technologies; Advances in Intelligent Systems and Computing; Springer: Cham, Switzerland, 2015; Volume 354, pp. 375–384. [Google Scholar]
  8. Sulikowski, P.; Zdziebko, T. Horizontal vs. Vertical Recommendation Zones Evaluation Using Behavior Tracking. Appl. Sci. 2021, 11, 56. [Google Scholar] [CrossRef]
  9. Bin, S.; Sun, G. Matrix Factorization Recommendation Algorithm Based on Multiple Social Relationships. Math. Probl. Eng. 2021, 2021, 6610645. [Google Scholar] [CrossRef]
  10. Shu, J.; Shen, X.; Liu, H.; Yi, B.; Zhang, Z. A content-based recommendation algorithm for learning resources. Multimed. Syst. 2018, 24, 163–173. [Google Scholar] [CrossRef]
  11. Peng, H.; Li, J.; Gong, Q.; Song, Y.; Ning, Y.; Lai, K.; Yu, P. Fine-grained Event Categorization with Heterogeneous Graph Convolutional Networks. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, IJCAI-19, Macao, China, 13 July 2019; pp. 3238–3245. [Google Scholar]
  12. Liang, H.; Xu, Y.; Li, Y.; Nayak, R.; Tao, X. Connecting users and items with weighted tags for personalized item recommendations. In Proceedings of the ACM conference on Hypertext and hypermedia, Toronto, ON, Canada, 13–16 June 2010; pp. 51–60. [Google Scholar]
  13. Liu, Y.; Peng, H.; Li, J.; Song, Y.; Li, X. Event detection and evolution in multi-lingual social streams. Front. Comput. Sci. 2020, 5, 1–15. [Google Scholar] [CrossRef]
  14. Sulikowski, P. Evaluation of Varying Visual Intensity and Position of a Recommendation in a Recommending Interface Towards Reducing Habituation and Improving Sales. In International Conference on e-Business Engineering; Springer: Cham, Switzerland, 2019; pp. 208–218. [Google Scholar]
  15. Bell, R.; Koren, Y. Scalable collaborative filtering with jointly derived neighborhood interpolation weights. In Proceedings of the Seventh IEEE International Conference on Data Mining (ICDM 2007), Omaha, NE, USA, 28–31 October 2007; pp. 43–52. [Google Scholar]
  16. Liu, N.; Xiang, E.; Zhao, M.; Yang, Q. Unifying explicit and implicit feedback for collaboratie filtering. In Proceedings of the 19th ACM International Conference on Information and Knowledge Management, Toronto, ON, Canada, 26–30 October 2010; pp. 1445–1448. [Google Scholar]
  17. Pan, W.; Liu, Z.; Ming, Z.; Zhong, H.; Wang, X.; Xu, C. Compressed knowledge transfer via factorization machine for heterogeneous collaborative recommendation. Knowl.-Based Syst. 2015, 85, 234–244. [Google Scholar] [CrossRef]
  18. Li, G.; Chen, Q. Exploiting explicit and implicit feedback for personalized ranking. Math. Probl. Eng. 2016, 2016, 2535329. [Google Scholar] [CrossRef] [Green Version]
  19. Ding, D.; Zhang, M.; Li, S.Y.; Tang, J.; Chen, X.; Zhou, Z.H. BayDNN: Friend Recommendation with Bayesian Personalized Ranking Deep Neural Network. In Proceedings of the Twenty-Sixth Conference on Information and Knowledge Management (CIKM’ 17), California, CA, USA, 6–10 November 2017; pp. 1479–1488. [Google Scholar]
  20. Nguyen, H.T.; Wistuba, M.; Grabocka, J.; Drumond, L.R.; Drumond, L.R.; Schmidt-Thieme, L. Personalized Deep Learning for Tag Recommendation. Knowl. Discov. Data Min. 2017, 6, 186–197. [Google Scholar]
  21. Wang, F.; Qu, Y.; Zheng, L.; Lu, C.T.; Philip, S.Y. Deep and Broad Learning on Content-aware POI Recommendation. In Proceedings of the IEEE 3rd International Conference on Collaboration and Internet Computing, FUTO, Nigeria, 14 June 2017; pp. 369–378. [Google Scholar]
  22. Kelen, D.; Daróczy, B.; Ayala-Gómez, F.; Ország, A.; Benczúr, A. Session Recommendation via Recurrent Neural Networks over Fisher Embedding Vectors. Sensors 2019, 19, 3498. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Liang, T.; He, L.; Lu, C.T.; Chen, L.; Philip, S.Y.; Wu, J.; Chen, L.; Philip, S.Y.; Wu, J. A Broad Learning Approach for Context-Aware Mobile Application Recommendation. ICDM 2017, 5, 955–960. [Google Scholar]
  24. Hakan, B.; Pinar, K. Context-aware friend recommendation for location based social networks using random walk. In Proceedings of the International Conference on World Wide Web, New York, NY, USA, 11–15, April 2016; pp. 531–536. [Google Scholar]
  25. Zheng, J.; Bhuiyan, M.Z.A.; Liang, S.; Xing, X.; Wang, G. Auction-based adaptive sensor activation algorithm for target tracking in wireless sensor networks. Future Gener. Comput. Syst. 2014, 39, 88–99. [Google Scholar] [CrossRef]
  26. Wang, Z.J.; Chen, K.M.; He, L. AsySIM: Modeling Asymmetric Social Influence for Rating Prediction. Data Sci. Pattern Recognit. 2018, 2, 25–40. [Google Scholar]
  27. Bhuiyan, M.Z.A.; Wang, G.; Vasilakos, A.V. Local Area Prediction-Based Mobile Target Tracking in Wireless Sensor Networks. IEEE Trans. Comput. 2015, 64, 1968–1982. [Google Scholar] [CrossRef]
  28. Sheng, Y.; Wu, T.; Wang, X. Incorporating term definitions for taxo-nomic relation identification. In Proceedings of the 9th Joint International Semantic Technology Conference (JIST), Hangzhou, China, 25–27 November 2019; Springer: Berlin/Heidelberg, Germany, 2019. [Google Scholar]
  29. Xue, H.J.; Dai, X.; Zhang, J.; Huang, S.; Chen, J. Deep Matrix Factorization Models for Recommender Systems. Int. Jt. Conf. Artif. Intell. 2017, 17, 3203–3209. [Google Scholar]
  30. Mao, Q.; Li, J.; Wang, S.; Zhang, Y.; Peng, H.; He, M.; Wang, L. Aspect-based sentiment classification with attentive neural turing machines. In Proceedings of the IJCAI, Macao, China, 10–16 August 2019; pp. 5139–5145. [Google Scholar]
  31. Peng, H.; Bao, M.; Li, J.; Bhuiyan, M.Z.A.; Liu, Y.; He, Y.; Yang, E. Incremental term representation learning for social network analysis. FGCS 2018, 86, 1503–1512. [Google Scholar] [CrossRef]
  32. Peng, H.; Li, J.; Wang, S.; Wang, L.; Gong, Q.; Yang, R.; Li, B.; He, L.; Yu, P.S. Hierarchical taxonomy-aware and attentional graph capsule rcnns for large-scale multi-label text classification. IEEE Trans. Knowl. Data Eng. 2020, 33, 2505–2519. [Google Scholar] [CrossRef] [Green Version]
  33. Cheng, C.Y.; Lin, I.C.; Wu, H.J. Recommendation System to Identify Collusive Users in Online Auctions Using the Pollution Diffusion Method. J. Internet Technol. 2019, 20, 353–358. [Google Scholar]
  34. Sun, Y.Z.; Han, J.W.; Yan, X.F.; Yu, P.S.; Wu, T. Pathsim: Meta path-based top-k similarity search in heterogeneous information networks. VLDB Endow. 2011, 4, 992–1003. [Google Scholar] [CrossRef]
  35. Shi, C.; Zhang, Z.Q.; Luo, P.; Yu, P.S.; Yue, Y.; Wu, B. Semantic Path based Personalized Recommendation on Weighted Heterogeneous Information Networks. In Proceedings of the 24th ACM International on Conference on Information and Knowledge Management (CIKM’ 15), Melbourne, Australia, 18–23 October 2015; pp. 453–462. [Google Scholar]
  36. Shi, C.; Kong, X.; Huang, Y.; Philip, S.Y.; Wu, B. A general framework for relevance measure in heterogeneous networks. IEEE Trans. Knowl. Data Eng. 2014, 26, 2479–2492. [Google Scholar] [CrossRef] [Green Version]
  37. Jiang, Z.; Liu, H.; Fu, B.; Wu, Z.; Zhang, T. Recommendation in heterogeneous information networks based on generalized random walk model and bayesian personalized ranking. In Proceedings of the Eleventh ACM International Conference on Web Search and Data Mining, Marina Del Rey, CA, USA, 5–9 February 2018; pp. 288–296. [Google Scholar]
  38. Alexandridis, G.; Siolas, G.; Stafylopatis, A. Accuracy versus novelty and diversity in recommender systems: A nonuniform random walk approach. In Recommendation and Search in Social Networks; Springer: Cham, Switzerland, 2015; pp. 41–57. [Google Scholar]
  39. Alexandridis, G.; Siolas, G.; Stafylopatis, A. A biased random walk recommender based on Rejection Sampling. In Proceedings of the 2013 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining, Niagara, ON, Canada, 25–28 August 2013; pp. 648–652. [Google Scholar]
  40. Wang, Z.; Liu, H.; Du, Y.; Wu, Z.; Zhang, X. Unified embedding model over heterogeneous information network for personalized recommendation. In Proceedings of the 28th International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 3813–3819. [Google Scholar]
  41. Liu, J.; Shi, C.; Hu, B.B.; Liu, S.; Philip, S.Y. Personalized Ranking Recommendation via Integrating Multiple Feedbacks. In Proceedings of the Knowledge Discovery and Data Mining, Halifax, Canada, 13–17 August 2017; pp. 131–143. [Google Scholar]
  42. He, X.N.; Liao, L.Z.; Zhang, H.W.; Nie, L.; Hu, X.; Chua, T.S. Neural Collaborative Filtering. In Proceedings of the 26th International Conference on World Wide Web (WWW’ 17), Perth, Australia, 3–7 April 2017; pp. 173–182. [Google Scholar]
  43. Luo, C.; Pang, W.; Wang, Z.; Lin, C. Social-based collaborative filtering recommendation using heterogeneous relations. In Proceedings of the International Conference on Data Mining (ICDM), Washington, WA, USA, 14–17 December 2014; pp. 917–922. [Google Scholar]
  44. Guo, H.; Tang, R.; Ye, Y.; Li, Z.; He, X. DeepFM: A factorization-machine based neural network for CTR prediction. arXiv 2017, arXiv:1703.04247. [Google Scholar]
  45. Liu, J.; Tang, M.; Zheng, Z.; Liu, X.; Lyu, S. Location-aware and personalized collaborative filtering for web service recommendation. IEEE Trans. Serv. Comput. 2016, 9, 686–699. [Google Scholar] [CrossRef]
  46. Benedikt, L.; Katja, H.; Jürgen, Z. Blended recommending: Integrating interactive information filtering and algorithmic recommender techniques. In Proceedings of the 33rd Annual ACM Conference on Human Factors in Computing Systems, Seoul, Korea, 18–23 April 2015; pp. 975–984. [Google Scholar]
  47. Wei, J.; He, J.H.; Chen, K.; Zhou, Y.; Tang, Z. Collaborative filtering and deep learning based recommendation system for cold start items. Expert Syst. Appl. 2017, 69, 29–39. [Google Scholar] [CrossRef] [Green Version]
  48. Suraj, S.; Babu, R.V. Deep Learning in Neural Networks: An Overview. Comput. Sci. 2015, 61, 85–117. [Google Scholar]
  49. Sun, Z.Q.; Li, F.; Huang, H.F. Large Scale Image Classification Based on CNN and Parallel SVM. In Proceedings of the International Conference on Neural Information Processing, California, CA, USA, 4 December 2017; pp. 545–555. [Google Scholar]
  50. Huang, K.W.; Lin, C.C.; Lee, Y.M.; Wu, Z.X. A Deep Learning and Image Recognition System for Image Recognition. Data Sci. Pattern Recognit. 2019, 3, 1–11. [Google Scholar]
  51. He, Y.; Li, J.; Song, Y.; He, M.; Peng, H. Time-evolving Text Classification with Deep Neural Networks. Int. Jt. Conf. Artif. Intell. 2018, 18, 2241–2247. [Google Scholar]
  52. Arif, M.H.; Li, J.; Iqbal, M.; Peng, H. Optimizing XCSR for text classification. IEEE Symp. Serv. Oriented Syst. Eng. (Sose) 2017, 8, 86–95. [Google Scholar]
  53. Peng, H.; Li, J.; He, Y.; Liu, Y.; Bao, M.; Wang, L.; Song, Y.; Yang, Q. Large-scale hierarchical text classification with recursively regularized deep graph-cnn. In Proceedings of the 2018 World Wide Web Conference, Lyon, France, 23–27 April 2018; pp. 1063–1072. [Google Scholar]
  54. Cheng, H.T.; Koc, L.; Harmsen, J.; Shaked, T.; Chandra, T.; Aradhye, H.; Anderson, G.; Corrado, G.; Chai, W.; Ispir, M.; et al. Wide & deep learning for recommender systems. In Proceedings of the 1st Workshop on Deep Learning for Recommender Systems, Boston, MA, USA, 15 September 2016; pp. 7–10. [Google Scholar]
  55. Xie, Z.; Zeng, Z.; Zhou, G.; Wang, W. Topic enhanced deep structured semantic models for knowledge base question answering. Sci. China Inf. Sci. 2017, 60, 1–15. [Google Scholar] [CrossRef]
  56. Zhao, Y.Y.; Qin, B.; Liu, T. Encoding syntactic representations with a neural network for sentiment collocation extraction. Sci. China Inf. Sci. 2017, 60, 110101. [Google Scholar] [CrossRef] [Green Version]
  57. Qu, Y.R.; Cai, H.; Ren, K.; Zhang, W.; Yu, Y.; Wen, Y.; Wang, J. Product-based neural networks for user response prediction. In Proceedings of the IEEE 16th International Conference on Data Mining (ICDM), Barcelona, Spain, 12–15 December 2016; pp. 1149–1154. [Google Scholar]
  58. Zhang, W.N.; Du, T.M.; Wang, J. Deep learning over multi-field categorical data. A case study on user response prediction. In Proceedings of the 38th European Conference on Information Retrieval Research (ECIR), Padua, Italy, 20–23 March 2016; pp. 45–57. [Google Scholar]
  59. Wang, C.; Liu, J.; Luo, F.; Tan, Y.; Deng, Z.; Hu, Q.N. Pairwise input neural network for target-ligand interaction prediction. In Proceedings of the IEEE International Conference on Bioinformatics and Biomedicine, Belfast, UK, 2–5 November 2014; pp. 67–70. [Google Scholar]
  60. Ning, X.; Karypis, G. Slim: Sparse linear methods for top-n recommender systems. In Proceedings of the 2011 IEEE 11th International Conference on Data Mining(ICDM), Vancouver, BC, Canada, 11 December 2011. [Google Scholar]
  61. Cheng, Y.; Yin, L.; Yu, Y. LorSLIM: Low rank sparse linear methods for top-N recommendations. In Proceedings of the International Conference on Data Mining (ICDM), Washington, DC, USA, 14–17 December 2014. [Google Scholar]
  62. Christakopoulou, E.; Karypis, G. HOSLIM: Higher-order sparse linear method for top-N recommender systems. In Proceedings of the Conference on Knowledge Discovery and Data Mining (PAKDD), New York, NY, USA, 24–27 August 2014. [Google Scholar]
  63. Shi, C.; Li, Y.; Zhang, J.; Sun, Y.; Philip, S.Y. A Survey of Heterogeneous Information Network Analysis. IEEE Trans. Knowl. Data Eng. 2017, 29, 17–37. [Google Scholar] [CrossRef]
  64. Zhao, H.; Yao, Q.; Li, J.; Song, Y.; Lee, D.L. Meta-Graph Based Recommendation Fusion over Heterogeneous Information Networks. In Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Halifax, NS, Canada, 13–17 August 2017; pp. 635–644. [Google Scholar]
  65. Zhene, Z.; Hao, P.; Lin, L.; Guixi, X.; Du, B.; Bhuiyan, M.Z.A.; Li, D. Deep Convolutional Mesh RNN for Urban Traffic Passenger Flows Prediction. In Proceedings of the 2018 IEEE SmartWorld, Ubiquitous Intelligence & Computing, Advanced&Trusted Computing, Scalable Computing & Communications, Cloud & Big Data Computing, Internet of People and Smart City Innovation (SmartWorld/SCALCOM/UIC/ATC/CBDCom/IOP/SCI), Guangzhou, China, 8–12 October 2018. [Google Scholar]
  66. Du, B.; Peng, H.; Wang, S.; Bhuiyan, M.Z.A.; Wang, L.; Gong, Q.; Liu, L.; Li, J. Deep irregular convolutional residual lstm for urban traffic passenger flows prediction. IEEE Trans. Intell. Transp. Syst. 2020, 21, 275–285. [Google Scholar] [CrossRef]
  67. Xu, X.; Wang, J.; Peng, H.; Wu, R. Prediction of academic performance associated with internet usage behaviors using machine learning algorithms. Comput. Hum. Behav. 2019, 98, 166–173. [Google Scholar] [CrossRef]
  68. Diederik, P.K.; Jimmy, B. Adam: A Method for Stochastic Optimization. In Proceedings of the ICLR, Santiago, MN, USA, 7–9 May 2015; pp. 1–15. [Google Scholar]
  69. Harper, F.M.; Joseph, A.K. The MovieLens Datasets: History and Context. ACM Trans. Interact. Intell. Syst. 2015, 5, 1–19. [Google Scholar] [CrossRef]
  70. Yu, X.; Ren, X.; Gu, Q.; Sun, Y.; Han, J. Collaborative filtering with entity similarity regularization in Heterogeneous information networks. In Proceedings of the IJCAI-HINA Workshop, IJCAI, Beijing, China, 2–3 August 2013. [Google Scholar]
  71. Li, F.; Xu, G.; Cao, L. Coupled matrix factorization within non-iid context. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining, Ho Chi Minh, Vietnam, 1–22 May 2015; Springer: Cham, Switzerland, 2015; pp. 707–719. [Google Scholar]
  72. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A Simple Way to Prevent Neural Networks from Overfitting. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
  73. Sheng, Y.; Xu, Z.; Wang, Y.; Melo, G.D. Murex: Multi-document semantic relation extraction for news analytics. WWW J. 2020, 23, 2043–2077. [Google Scholar]
Figure 1. The selected meta-path.
Figure 1. The selected meta-path.
Applsci 11 07418 g001
Figure 2. Architecture of the proposed method.
Figure 2. Architecture of the proposed method.
Applsci 11 07418 g002
Figure 3. Framework of MFDNN.
Figure 3. Framework of MFDNN.
Applsci 11 07418 g003
Figure 4. The network pattern of MovieLens 1m and Netflix.
Figure 4. The network pattern of MovieLens 1m and Netflix.
Applsci 11 07418 g004
Figure 5. The HR and NDCG values for different learning rates.
Figure 5. The HR and NDCG values for different learning rates.
Applsci 11 07418 g005
Figure 6. The MAP and NDCG values for different numbers of hidden layers.
Figure 6. The MAP and NDCG values for different numbers of hidden layers.
Applsci 11 07418 g006
Figure 7. The HR and NDCG values for different numbers of embedding factors.
Figure 7. The HR and NDCG values for different numbers of embedding factors.
Applsci 11 07418 g007
Figure 8. The HR and NDCG values for different dropout rates.
Figure 8. The HR and NDCG values for different dropout rates.
Applsci 11 07418 g008
Figure 9. The HR and NDCG values achieved with MFDNN+, MFDNN-, and MFDNN.
Figure 9. The HR and NDCG values achieved with MFDNN+, MFDNN-, and MFDNN.
Applsci 11 07418 g009
Figure 10. HR and NDCG values for MFDNN- - and MFDNN.
Figure 10. HR and NDCG values for MFDNN- - and MFDNN.
Applsci 11 07418 g010
Figure 11. HR and NDCG values for MF, DNN, and MFDNN.
Figure 11. HR and NDCG values for MF, DNN, and MFDNN.
Applsci 11 07418 g011
Figure 12. The HR and NDCG values for MFDNN: (share MF), MFDNN: (share DNN), and MFDNN.
Figure 12. The HR and NDCG values for MFDNN: (share MF), MFDNN: (share DNN), and MFDNN.
Applsci 11 07418 g012
Figure 13. The HR and NDCG values for the MFDNN and baseline methods.
Figure 13. The HR and NDCG values for the MFDNN and baseline methods.
Applsci 11 07418 g013
Figure 14. HR and NDCG values for MFDNN and baseline methods.
Figure 14. HR and NDCG values for MFDNN and baseline methods.
Applsci 11 07418 g014
Figure 15. The HR and NDCG values for the MFDNN and HIN-based baseline methods.
Figure 15. The HR and NDCG values for the MFDNN and HIN-based baseline methods.
Applsci 11 07418 g015
Figure 16. The HR and NDCG values for the MFDNN and HIN-based baseline methods.
Figure 16. The HR and NDCG values for the MFDNN and HIN-based baseline methods.
Applsci 11 07418 g016
Figure 17. The results of the MFDNN and baseline methods.
Figure 17. The results of the MFDNN and baseline methods.
Applsci 11 07418 g017
Table 1. Notations.
Table 1. Notations.
SymbolDescription
YUser–item interaction matrix
USet of users
ISet of items
Y ^ Final prediction results
Y ^ Implicit feedback prediction results
Y ^ + Explicit feedback prediction results
Y User–item relation matrix
Y + User–item rating matrix
Y ^ u i , M F Final prediction results of MF
Y ^ u i , D N N Final prediction results of DNN
p u F User embedding vector of MF
q i F Item embedding vector of MF
p u I User embedding vector of DNN
q i I Item embedding vector of DNN
Table 2. The MFDNN algorithm.
Table 2. The MFDNN algorithm.
Algorithm MFDNN algorithm
Y ˜ User–item relation matrix;
Y ˜ + User–item rating matrix;
λ Parameter of regularization term;
Learning rate⇐0.001;
epochs ⇐ Number of iterations;
p u F User embedding vector of MF;
q i F Item embedding vector of MF;
p u I User embedding vector of DNN;
q i I Item embedding vector of DNN;
epochs Calculate y ^ u i , M F + Equations (8)–(10)
Calculate y ^ u i , D N N + Equations (11)–(14)
Update MFDNN with Adam
Calculate y ^ u i + Equation (7)
Calculate y ^ u i at the same way
Calculate Y ^ Equation (15)
Top-N recommendation list
Table 3. Descriptions of the MovieLens 1m and Netflix datasets.
Table 3. Descriptions of the MovieLens 1m and Netflix datasets.
AspectMovieLens 1mNetflix
#users604048,018
#movies370617,770
#ratings1,000,20911,160,900
Rating Density0.044680.01308
Table 4. The experimental results compared with the results from the baseline methods.
Table 4. The experimental results compared with the results from the baseline methods.
MovieLens 1m Netflix
MethodHR@10NDCG@10HR@10NDCG@10
MFDNN0.72780.43190.68280.4214
DMF0.67350.39750.57760.3459
NCF0.70480.42520.62450.4000
HeteCF0.70970.42680.66010.4013
HeteMF0.71230.42710.66090.4062
CMF0.72350.43080.64450.3893
Table 5. The HR and NDCG values for different top-N values.
Table 5. The HR and NDCG values for different top-N values.
MovieLens 1m Netfilx
Top-NHRNDCGHRNDCG
N=50.53030.36720.45830.3058
N=100.72790.43190.68280.4214
N=150.78690.44430.75420.4386
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gong, J.; Zhang, X.; Li, Q.; Wang, C.; Song, Y.; Zhao, Z.; Wang, S. A Top-N Movie Recommendation Framework Based on Deep Neural Network with Heterogeneous Modeling. Appl. Sci. 2021, 11, 7418. https://0-doi-org.brum.beds.ac.uk/10.3390/app11167418

AMA Style

Gong J, Zhang X, Li Q, Wang C, Song Y, Zhao Z, Wang S. A Top-N Movie Recommendation Framework Based on Deep Neural Network with Heterogeneous Modeling. Applied Sciences. 2021; 11(16):7418. https://0-doi-org.brum.beds.ac.uk/10.3390/app11167418

Chicago/Turabian Style

Gong, Jibing, Xinghao Zhang, Qing Li, Cheng Wang, Yaxi Song, Zhiyong Zhao, and Shuli Wang. 2021. "A Top-N Movie Recommendation Framework Based on Deep Neural Network with Heterogeneous Modeling" Applied Sciences 11, no. 16: 7418. https://0-doi-org.brum.beds.ac.uk/10.3390/app11167418

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop