Next Article in Journal
Dynamic Temperature Compensation of Pressure Sensors in Migratory Bird Biologging Applications
Next Article in Special Issue
Comparison of Selected Machine Learning Algorithms in the Analysis of Mental Health Indicators
Previous Article in Journal
Learning-Based Collaborative Computation Offloading in UAV-Assisted Multi-Access Edge Computing
Previous Article in Special Issue
Machine Learning for Energy-Efficient Fluid Bed Dryer Pharmaceutical Machines
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multiscale Neighbor-Aware Attention Network for Collaborative Filtering

1
Institute of Intelligent Information Processing, Shanxi University, Taiyuan 030006, China
2
North Automatic Control Technology Institute, Taiyuan 030006, China
3
School of Computer and Information Technology, Shanxi University, Taiyuan 030006, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(20), 4372; https://doi.org/10.3390/electronics12204372
Submission received: 25 September 2023 / Revised: 11 October 2023 / Accepted: 17 October 2023 / Published: 22 October 2023
(This article belongs to the Special Issue Advances in Intelligent Data Analysis and Its Applications)

Abstract

:
Most recommender systems rely on user and item attributes or their interaction records to find similar neighbors for collaborative filtering. Existing methods focus on developing collaborative signals from only one type of neighbors and ignore the unique contributions of different types of neighbor views. This paper proposes a multiscale neighbor-aware attention network for collaborative filtering (MSNAN). First, attribute-view neighbor embedding is modeled to extract the features of different types of neighbors with co-occurrence attributes, and interaction-view neighbor embedding is leveraged to describe the fine-grained neighborhood behaviors of ratings. Then, a matched attention network is used to identify different contributions of multiscale neighbors and capture multiple types of collaborative signals for overcoming sparse recommendations. Finally, we make the rating prediction by a joint learning of multi-task loss and verify the positive effect of the proposed MSNAN on three datasets. Compared with traditional methods, the experimental results of the proposed MSNAN not only improve the accuracy in MAE and RMSE indexes, but also solve the problem of poor performance for recommendation in sparse data scenarios.

1. Introduction

Collaborative filtering has become a fundamental technology in e-commerce platforms, which has made remarkable achievements. Collaborative filtering assumes that similar users have similar interests, and makes recommendation services for the target user based on similar users’ interests. In most e-commerce scenarios, the interaction between users and items reflects users’ interest preferences, which is often used to find similar neighbors to model collaborative representations for users and items [1,2].
However, the e-commerce platform generates a large number of new users and product items every day. Some users have not rated or purchased new products, which leads to fewer interactive records. Thus, it is difficult to learn high-quality embedding representations for users and items. As a result, collaborative filtering still faces the problem of interaction data sparsity in recommender systems [3]. Aiming to solve the sparse recommendation problem, most recommender systems leverage auxiliary attribute information of users and items to establish users’ preferences and mine similar neighbors for collaborative filtering [4]. However, similar neighbors mainly come from an attribute preference of uniform scale. In fact, different users have various social attributes, and different items have multiple product descriptions. Different combinations of attributes are conducive to generating various types of neighbors to describe semantic characteristics of different granularity for nodes. That is, when some users have multiple attributes and the other users have a few attributes, we can make collaborative recommendations according to their multiscale attribute neighbors. In addition, the interaction rating of users reflects their differentiated preference for items. We can leverage differentiated rating preference via different rating neighbors to explore the fine-grained preference motivation. Thus, how to deal with collaborative filtering by combining multiscale attribute neighbors with rating neighbors is an important task. Multiscale node embedding describes the fine-grained semantic representations from multiple perspectives, and can effectively solve the problem of poor performance of sparse recommendation, which is of great significance for industrial applications.
In the e-commerce recommender systems, multiscale attribute combinations can produce single-attribute-view neighbors and multi-attribute-view neighbors. For example, in movie recommender systems, users with the same gender and age have more similar interest behaviors than users of the same gender. As a result, different types of attribute-view neighbors can be constructed in various attribute combination spaces. In addition, different interaction-view neighbors can be obtained according to the types of interaction ratings, such as 1–5. We model the interaction-view neighbor embedding of nodes on various interactive views. The attention mechanism [5] is used to focus on specific input features, analyze the importance of all aspects of input features, and improve the expression ability of the model, which has been widely applied in the fields of natural language processing and image processing. Inspired by [6], this paper captures a different type of attribute neighbor embedding and interactive neighbor embedding and mine their collaborative signals with the attention mechanism to model multiscale node embedding. The multiscale node embedding can effectively capture diverse semantics of nodes from different types of neighbors to enhance sparse recommendation.
To summarize, the main contributions of this paper are as follows.
  • Various neighbor graphs of attribute tag and rating tag are designed to learn attribute neighbor embedding and interaction neighbor embedding, which capture embedding signals of various neighbors at different levels.
  • An attention network is developed to refine the collaborative semantics of multiscale neighbors, which is utilized to filter the irrelevance signals of various types of neighbors.
  • A joint learning of multiscale neighbor embedding is proposed for rating prediction, which solves the problem of poor accuracy in the context of sparse recommendation.
The rest of this paper is structured as follows. Section 2 outlines related work, including the state-of-the-art of neighbor-based recommendation and attention mechanism. In Section 3, we present the framework of the proposed multiscale neighbor-aware attention network. Section 4 provides the methodology of the MSNAN recommendation. Section 5 describes the experimental setup and evaluation. Section 6 gives the experimental results and analysis. Finally, Section 7 concludes this work.

2. Related Work

In this section, the related work on neighbor-based recommendations of collaborative filtering and attention mechanisms are briefly reviewed.

2.1. Traditional Collaborative Filtering

Traditional collaborative filtering utilizes the nearest neighbors of a user or an item to generate recommendation results. Most popular methods leverage matrix factorization to learn the latent factors of users and items in terms of their historical information, such as BiasedMF [7], PMF [8], SVD++ [9], and LLORMA [10]. The decomposed low-dimensional factor vectors can be used to predict the user’s rating preference for items. These methods mainly depend on the rating interactions between the user and item and have low performance in the case of sparse data. Thus, some side information is merged into matrix factorization to alleviate the sparsity of interaction records [11,12]. SSLIM [13] develops a sparse aggregation coefficient matrix by considering the user–item profiles and side information of items. The auto-encoder and -decoder techniques are used to learn the latent factors of nodes for collaborative filtering [14]. Park et al. [15] developed a group recommender system to select the suited recommendation items for store product placement. In recent years, some matrix decomposition models of deep learning have been studied [16,17]. DeepFM learns low-order and high-order interaction features of compressed interaction neighbors through a neural network [18]. Cai et al. [19] leveraged various multi-grained sentiment features and latent factors of matrix factorization to obtain sufficient representations of users and items to make rating predictions. Although these models have the ability to handle the sparsity recommendation problem, they leverage raw neighbors to learn their high-order interaction features and have limitations on processing different contributions of fine-grained neighbors.

2.2. Deep Learning-Based Collaborative Filtering

Deep learning-based methods can utilize the interaction neighbors between users and items to extract their latent vectors, which achieves excellent performance. Deep neural networks take advantage of both side information and feedback information to model linear and nonlinear features. For example, NFM [20] encodes interaction neighbor IDs of users, items, and their features into different vectors. Wide and Deep learning [21] combines the linear model and the deep neural network to implement efficient recommendations for scenarios with sparse data. Some works transfer diverse interactions of neighbors to learn rich features of nodes [1]. MCCF decomposes and recombines the latent components of user–item interaction graph to capture fine-grained user preference [22]. Aiming to solve the cold-start problem, Magron et al. [23] considered content information of acoustic features to learn the interaction between users and songs and proposed a neural content-aware collaborative filtering framework for music recommendation. Graph neural networks can learn the interaction characteristics of multi-order neighbors. MBGCN leverages multiple types of user-to-item interactions and the similarity item-to-item to propagate neighbor semantics [24]. Multiple neighbors on the path are used to capture the diversity of user interests and improve the accuracy of personalized recommendations [25]. User–item neighbor interaction and item–item neighbor relevance were leveraged to model a two-hop paths-based deep network to improve user engagement [26]. Tai et al. [27] designed a user-centric two-level path network in terms of entities of knowledge graph to generate user portfolio information. Duan et al. [28] learned the features of nodes from different dimensions of time and position and investigated the consistency of two representations for sequential recommendation. Most of neighbor embedding algorithms exploit interactive neighbors and do not distinguish the representations of different types of neighbors.

2.3. Attention Network for Collaborative Filtering

An attention mechanism is applied in collaborative filtering recommendations by identifying different importances of neighbors [5,29]. DGCF identifies the importance of diverse user–item interaction neighbors and models fine-grained intent-aware graph collaborative filtering [30]. By integrating content-based with collaborative filtering, ACCM considers the importance of end-to-end features and traditional attributes adaptively via an attention mechanism to handle the cold-start problem [3]. Chen et al. designed both item-level and component-level attention networks for multimedia recommendation [31]. Different auxiliary information can be learned to enhance important signals for neural networks. By leveraging intra-entity interaction and inter-entity interaction, AKUPM explores the relationships between users and other entity neighbors for alleviating the sparsity recommendation problem [32]. The attentional mechanism can provide different functions of multi-type entities on the knowledge graph and help obtain important neighbor information on different paths. KGAT learns the importance of higher-order neighbors of nodes in knowledge graphs by considering various features of entities [33]. A knowledge-aware attention mechanism is adopted to discriminate the contributions of different collaborative neighbors for recommender systems [34]. There are considerable advantages to applying the attention model in graph-based neural networks. A neural co-attention model utilizes auxiliary information of meta-based neighbors for top-n recommendation of heterogeneous information networks [35]. By leveraging the higher-order friends in the social network, Xiao et al. [36] designed a social explorative attention network to make personal interest recommendations. Ye et al. [37] utilized both influence graph and the preference graph to fuse different user and item embeddings to make rating predictions. However, most attentional models deal with the role of the same type of neighbors, which limits the discriminative contributions of various neighbors to the user’s overall decision making.

3. Framework of Multiscale Neighbor-Aware Attention Network

Figure 1 shows the framework of the proposed multiscale neighbor-aware attention network. The framework comprises four components: (1) attribute-view neighbor node embedding, (2) interaction-view neighbor node embedding, (3) attentional multiscale neighbor node embedding, and (4) rating prediction. In the framework, attribute-view neighbor node embedding is used to learn the different roles that similar neighbors with different attribute sets play in collaborative filtering. The interaction-view neighbor node embedding learns the role of similar neighbors with the same rating behavior in collaborative filtering. Neighbors with different attribute sets can form multiscale similar neighbors. Considering that users with different attributes have different rating behavior habits, we model the preferences of attribute sets for rating behavior through a multiscale neighbor-aware attention network. The attention network measures the collaborative contribution of coarse-scale attribute neighbors and fine-scale attribute neighbors to the rating decisions of target users.
In a nutshell, the framework works as follows. For the attribute-view neighbor node embedding, we first construct an attribute-view neighbor graph according to the association of a node on an attribute set such as {a} or {a,b}. Attribute sets of different scales induce multiple types of attribute neighbors. Based on various attribute-view neighbors, graph neural networks are used to obtain attribute-view neighbor node embedding.
For the interaction-view neighbor node embedding, we divide different rating tag spaces according to different rating grades. Under different rating tag spaces, we form similar neighbors with different rating behaviors. Then, graph neural networks are leveraged on different user–item interaction neighbor graphs to obtain various interaction-view neighbor node embedding.
Then, an attention mechanism is utilized to estimate the interactive contributions between various attribute neighbors and interaction neighbors. For the attribute-view and interaction-view neighbors, we calculate the global attribute neighbor collaborative signal and interactive neighbor collaborative signal, respectively. Meanwhile, we match the local collaborative signals between different-scale attribute neighbors and interactive neighbors. Considering the global and local semantic information provided by different neighbors, multiscale neighbor node embedding is computed to capture rich collaborative signals.
Finally, based on multiscale user embedding and item embedding, the inner product can be used to predict the rating score of the user to the item.

4. Methodology

4.1. Attribute-View Neighbor Embedding

In e-commerce networks, some attribute descriptions describe characteristics of users or products, which are helpful to discover various types of similar users or similar items. In this subsection, we utilize various types of attribute sets to calculate similar neighbors of different scales. Then, we learn the nodes’ attribute-view neighbor embedding in terms of different-scale neighbor graphs.
Usually, users have various kinds of attributes, such as gender, age, occupation, and so on, which reflects users’ interest preference to a certain extent. For example, users with the same gender can form a neighbor graph with a coarse-grained perspective, while users with the same gender and age can build a neighbor graph with a fine-grained attribute space. A coarse-grained neighbor can provide robust interest preferences for cold-start recommendation. A fine-grained neighbor helps discover refined similar preferences and model accurate collaborative recommendation. Based on this assumption, we can construct different views of attribute neighbor graphs to incorporate signals of various neighbors for modeling the embedding of nodes.
Given an attribute a, we can define a neighbor set N u a of user u on an attribute a as follows:
N u a = { u | f a ( u ) = f a ( u ) }
where f a ( u ) is the attribute value of the user u on attribute a. N u a describes the collaborative neighbors with the same attributes as user u. Considering all attribute value types in the set A, we can construct the user-attribute relation matrix as M U × A . Then, based on the user distribution of the set A, we can establish the neighbor relationship matrix of users by M M T , labeled as M U × U . Here, various attribute-view neighbors reflect multiscale collaborative preference, which can affect the decision-making tendency of the target user.
Based on the idea in [33,38], we can define the collaborative signal of a first-order neighbor u for user u on the attribute a as follows.
m u u a = u + ( u u ) | N u a | | N u a |
Here, m u u a represents the influence of similar neighbors in terms of attribute a on the target user u. u is the initialized embedding vector. Thus, considering all the first-order neighbors, attribute a-view collaborative signal for user u can be defined as e u a ( 1 ) = u N u a m u u a ( 1 ) . As is known, neighbors with the same attributes tend to spread their preferences in the social network. Considering the spread contributions of k − 1 hop neighbors, the recursive collaborative signal of neighbor u for user u on the attribute a can be formulated as follows [33,38].
m u u a ( k ) = u ( k 1 ) + ( u ( k 1 ) u ( k 1 ) ) | N u a | | N u a |
Further, attribute a-view recursive collaborative signal for user u can be defined as e u a ( k ) = u N u a m u u a ( k ) . We adopt the average pooling to obtain the attribute a-view neighbor-aware node embedding for user u as e u A t t _ a = a g g ( e u a ( 1 ) , , e u a ( k ) ) . Here, some aggregation strategies can be used to fuse different orders of neighbor embedding.
Different types of attributes can induce different similar neighbors. Considering various types of neighbors in other attribute views, we can obtain attribute A-view neighbor-aware user embedding as e u A t t _ A = a v g ( e u A ( 1 ) , , e u A ( k ) ) . Similarly, given an item v, we adopt different types of item neighbors to obtain the attribute-view neighbor-aware item embedding.
Different users have special behavioral perceptions of rating labels, which can be used to explore users’ behavioral preferences in fine granularity. Thus, according to the types of rating labels, we also divide different interaction view spaces with rating labels and construct various rating interaction graphs for learning user and item embeddings. For example, we can think of item groups with the same rating as neighbors of a user with the same scale preference. Based on different rating labels, we model interaction-view neighbor-aware embedding with different rating neighbors, which can be defined as { e u I n t 1 , , e u I n t r } . Then, interaction-view neighbor-aware item embedding of an item v can be defined as { e v I n t 1 , , e v I n t r } .

4.2. Cross Attention-Based Multiscale Neighbor Embedding

In the homogeneous attribute view, the multiscale neighbor-aware embedding signals with different granularity can provide diversified collaborative signals for node representation. In addition, various neighbor-aware embeddings of attribute view and interactive view can be used to model the heterogeneous collaborative signals, which can enrich and enhance the representation ability of node embedding.
According to neighbor-aware node embedding on different attribute-view spaces, we can model global attribute neighbor-aware node embedding for user u as follows.
q i = h i e u A t t _ i
Equation (4) describes the influence of user embedding in the attribute i space on the global attribute neighbor-aware user embedding. h i is the parameter vector. Considering different user embeddings of m spatial types, the normalized weights are defined in Equation (5).
α i = e q i s { 1 , , m } e q s
Then, the global attribute neighbor-aware node embedding for user u is defined as follows.
e u A t t _ g = i { 1 , , m } α i e u A t t _ i
In Equation (5), different neighbor-aware user embedding representations depict discriminative semantic information from various attribute-induced neighbors, which provides collaborative signals of neighbors comprehensively for target user’s decisions. The global collaborative embedding e u A t t _ g of various neighbors can improve the semantic ability for recommender systems. Similarly, we can obtain the global collaborative embedding of interaction view spaces as e u I n t _ g .
Given the item v, we can also obtain global attribute-neighbor-aware item embedding and interaction-neighbor-aware item embedding as e v A t t _ g and e v I n t _ g .
Further, taking into account the collaborative signals matched by different neighbor embedding from the attribute and interactive views, we utilize cross-attention to model matched neighbor embedding. Based on the global collaborative embedding of interaction view, we can compute matched neighbor embedding of attribute view for user u as follows.
β i = e u I n t _ g e u A t t _ i
Here, Equation (7) defines the preferential influence of user embedding in attribute i’s view on the user’s rating decision. Then, we normalize this influence using Equation (8).
γ i = e β i s { 1 , , m } e β s
The normalized weight γ i reflects the influence of multiscale attribute neighbor embedding for the user’s interaction rating. Furthermore, the attribute embedding by incorporating user rating behavior preference can be updated as in Equation (9).
e u A t t = i { 1 , , m } γ i e u A t t _ i
The multi-type matched signals based on attribute neighbors and interactive neighbors can enhance the embedding representation of nodes. Similarly, we can calculate the matched neighbor embedding of the interactive view as below.
w j = e u A t t _ g e u I n t _ j
Here, w j is the dependency influence of user embedding in rating tag j’s view on user’s attributes. In terms of different rating levels, this dependence effect can be normalized using Equation (11).
g j = e w j p { 1 , , t } e w p
e u I n t = p { 1 , , t } g j e u I n t _ j
In Equation (12), e u I n t represents the user’s interaction embedding incorporating the dependency of user attributes. Considering the matched neighbor embedding in the attribute view and neighbor embedding in the interactive view, we can model fused multiscale neighbor embedding with the concatenation operator as follows.
e u = e u A t t | | e u I n t

4.3. Rating Prediction

Based on the fused multiscale neighbor embedding of users and items, the dot product is used to predict the rating of the user u on an item v, which is shown as follows:
y ^ = e u · e v + b g + b u + b v
Here, the parameters b g , b u , and b v are global bias, user bias, and item bias. To preserve the preference information of user attributes over item attributes, we define rating prediction of user u for the item v with their attribute-view neighbor embeddings as follows:
y ^ A t t = e u A t t · e v A t t + b g + b u + b v
Similarly, we can compute the rating with the interaction-view neighbor embeddings in interactive space, which is shown as below.
y ^ I n t = e u I n t · e v I n t + b g + b u + b v
In the process of model optimization, to observe the influence of attribute neighbors and interaction neighbors on rating prediction, we define the joint root mean squared error (RMSE) loss function as follows.
L = λ R M S E y ^ , y + λ A t t R M S E y ^ A t t , y + λ I n t R M S E y ^ I n t , y
The joint loss function considers the global and local important neighbors to predict the user’s rating of the item, which not only captures the user’s attribute preference for the item from various attribute-view neighbors, but also retains the behavioral preference of collaborative fine-grained interaction neighbors.

5. Experiments

In this section, we verify the performance of the proposed model with the aim of answering the following three questions:
  • RQ1: How does MSNAN perform compared with state-of-the-art neighbor-based collaborative filtering methods?
  • RQ2: How does the multiscale neighbor node embedding perform for sparsity recommendation?
  • RQ3: How do different types of neighbor embedding affect the performance of the model?

5.1. Dataset

We ran the proposed MSNAN model and baselines on three public datasets: Movielens-100kr (ML-100kr) (https://grouplens.org/datasets/movielens/ (accessed on 24 September 2023)), Book-Crossing-10croe (BK-10C) (http://bookcrossing.com (accessed on 24 September 2023)), and Douban (https://movie.douban.com/ (accessed on 24 September 2023)) datasets to verify their effectiveness. The ML-100kr contains the interaction ratings of 943 users on 1682 movies. Users have 5 attributes, and movies have 19 attributes. The rating score is on the scale 1–5. For the BK-10C dataset, we selected users who had rated at least 10 books and books that have been rated by at least 10 users, which involved 1820 users and 2030 books. The rating score uses the range of 1–10. The Douban dataset contains ratings of 6971 movies from 3022 users with rating values of 1–5 [39]. On all datasets, the higher the rating is, the more the user likes the movie/book. Statistical information of experimental datasets is shown in Table 1. During the experiment, we use the MAE and RMSE indexes for performance evaluation. The smaller MAE and RMSE values indicate better performance. All the datasets were divided into training set, validation set, and testing set with a proportion of 8:1:1. The rating prediction performance of the model is evaluated on the testing set.

5.2. Baseline Models

To test the contribution of MSNAN method, we employ traditional methods, deep learning-based recommendations, and GCN-based models to compare the performance. Several baselines are as follows.
(1)
NCF [17]. A neural network method is used to learn the interaction information between users and items for collaborative filtering.
(2)
Wide and Deep [20]. A method combining a generalized linear model and deep neural network is designed to improve the performance of recommender systems.
(3)
NGCF [33]. A collaborative filtering method based on a graph neural network learns the embedding representations of users and items with a user–item interaction graph.
(4)
GCN [40]. A graph convolutional neural network leverages the information of multi-order neighbors by superimposing several convolutional layers for recommendation.
(5)
LightGCN [38]. The embedding representations of user and item are learned by aggregating linear neighbor information of nodes.
(6)
GAT [41]. A graph convolutional neural network method together with attention mechanism learns weighted node embedding representation for recommendation.
(7)
ACCM [3]. The attention mechanism is used to integrate a content-based method with collaborative filtering for rating prediction.
(8)
AFM [5]. An attentional network factorization method learns the interactive importance of different features for prediction.
(9)
TANP [42] A task-adaptive neural network is constructed to learn the relevance of different tasks for user cold-start recommendations.

5.3. Parameter Settings

In the experiment, we adopt a grid search to obtain the parameters for optimizing the performance of model. For the attention mechanism, we tune its dimension with the values {32, 64, 128, 256}. To prevent the model overfitting, we adjust the dropout values {0.1, 0.2, 0.3, 0.4, 0.5}. We utilize stochastic gradient descent to optimize the model with a learning rate of 0.01. In order to obtain a better comparison, we use inner product in the prediction layer for NGCF, GCN, LightGCN, and GAT methods to make rating predictions.

5.4. Experiment Settings

For the proposed multiscale neighbor method, the NGCF graph convolution model is used in Equations (2) and (3) to capture the cooperative signal propagation between nodes. We can also use other signal propagation methods to model the initial representation of nodes in a certain view, and then observe the effect of multiscale neighbors. During the experiment, for the multiscale neighbor graph, we adopted different graph neural network models, such as NGCF, GCN, LightGCN, and GAT, to learn various neighbor embedding and observe the robust effect of multiscale neighbor signals. In the experiment, the hyperparameters λ , λ A t t , and λ I n t are set to 0.01.

6. Experimental Results

In this subsection, we compare the proposed MSNAN model with the benchmark models in terms of MAE and RMSE metrics on three datasets. The experimental results are shown in Table 2. In Table 2, we have the following observations from the experimental results.
Through the comparison results, the performance of MSNAN shows excellent improvements on three datasets. For example, compared with the ACCM method of collaborative filtering, the MAE and RMSE values of the MSNAN + LightGCN method increase by 2.74%, 2.35%, 8.18%, 5.27%, 4.70%, and 4.08% on three datasets, respectively. In addition, the MAE and RMSE of the MSNAN + NGCF model have achievements of 3.29%, 2.05%, 1.09%, 1.26%, 1.90%, and 0.94% over the best baseline on three datasets, respectively. Meanwhile, the improvements of the MSNAN + LightGCN model to the best baseline are 2.74% and 2.35% for MAE and RMSE on the ML-100kr dataset, 1.33% and 1.41% on the BK-10C dataset, and 2.23% and 1.86% on the Douban dataset, respectively. This shows that the proposed MSNAN model can better learn the embedding representation of users and items and effectively improve the accuracy of rating prediction, which verifies the significance of collaborative semantics of multiscale neighbors.
Compared with these graph neural network baselines, the proposed MSNAN based on multiscale neighbors achieves competitive improvements on all datasets. For example, on the ML-100kr dataset, the MSNAN + NGCF, MSNAN + GCN, MSNAN + LightGCN, and MSNAN + GAT methods improve by 5.32%, 3.83%; 3.38%, 3.14%; 4.28%, 4.00%; and 3.91%, 3.10% over the NGCF, GCN, LightGCN, and GAT baselines on MAE and RMSE metrics. In addition, on the BK-10C dataset, the MSNAN + NGCF, MSNAN + GCN, MSNAN + LightGCN, and MSNAN + GAT methods also improve MAE and RMSE values by 1.89%, 1.29%; 2.60%, 1.18%; 1.46%, 1.56%; and 1.80%, 1.39% over the corresponding graph model baselines, respectively. This is because the node embedding of MSNAN combines collaborative semantics of multiscale neighbors, which purifies the important information of similar neighbors in rating prediction. However, for different graph neural network models, the representation quality of nodes is different, and the superposition effect of MSNAN is also different. As is known, LightGCN simplifies the transformation matrix and activation function, which obtains the highest embedding quality of node representation, and can also improve the performance of this method. The performance of the general GCN method is relatively backward. In addition, compared with ACCM and AFM methods, the MSNAN approach can achieve a smaller error in the rating prediction scenario, which indicates that the proposed model can better learn the embedding representation of users and items with the collaborative signal of multiscale neighbors.

7. Discussion and Analysis

7.1. Sparsity Analysis

Sparse recommendation is a challenging problem in recommender systems. The sparsity problem of the user–item interaction matrix makes it difficult to learn the semantic representation of users and items, and reduces the accuracy of rating prediction results. In order to observe the robust advantage of multiscale neighbor signals for sparse recommendation, we conduct the comparative experiments by randomly shielding the rating labels and setting different sparsity proportions for three datasets. In the experiment, we compare the performance of proposed MSNAN model with several graph neural network methods on MAE and RMSE metrics. The experimental results are shown in Figure 2, Figure 3, Figure 4, Figure 5, Figure 6 and Figure 7.
In the above figures, with the increase in data sparsity, the MAE and RMSE values of all methods show an increasing trend, which indicates that the error of rating prediction is gradually increasing. This demonstrates that high data sparsity can affect the quality of embedding representation of users and items and reduces the accuracy of recommendation results. In addition, with the improvement in data sparsity, the MAE and RMSE values of the graph neural network methods fluctuate. This is because the benchmark models of graph neural networks only consider one type of neighbor information of the user–item interaction matrix. Insufficient interaction records lead to the difficulties of higher-order information propagation of single-type neighbors, which limits the stability of node embedding representation.
Compared with graph neural network methods, the proposed MSNAN model consistently yields the best performance for MAE and RMSE on three datasets to adapt the sparsity ratio of different scales. For different graph neural network models, the result of blending the MSNAN model is more generalized than the original model. As shown in Figure 4, Figure 5, Figure 6 and Figure 7, when there is a low drop ratio, LightGCN can obtain better performance compared with other graph neural network methods due to its better node representation quality. This is because LightGCN itself effectively learns the embedding representation of nodes by simplifying the nonlinear structure and reducing complexity. In large-scale e-commerce platforms, the sparsity of user–item interactions can be high. It is difficult to find collaborative neighbors based on the similarity of interaction behaviors. Although the user–item matrix loses a part of interaction records, the proposed model fuses various neighbor information from attribute views and rating views, which fully learns the representations of users and items. According to the attributes of users or items, we can find neighbors of different scales and conduct collaborative filtering recommendation. Moreover, the model takes advantage of high-quality collaborative signals from multiscale neighbors for improving the quality of embedding representation, which is suitable for the practice of large-scale e-commerce and alleviates the performance impact of data sparsity to some extent.

7.2. Impact of Neighbor Embedding

For the multiscale neighbor embedding, the neighbor embedding of different views contributes differently, which affects the result of rating prediction. To observe the roles of different types of neighbor embedding, Figure 8, Figure 9 and Figure 10 give the performance of methods by removing different attribute-view neighbor embedding. In Figure 8, for the ML-100kr dataset, the users have three types of neighbor embedding. We can see that the MAE and RMSE of three att-view neighbor embedding change greatly compared with the other two types, which states that three att-view neighbor embedding plays an important role in predicting the interest rating of the target user. That is, through the neighbor interaction graphs with multiple similar attributes, the multiscale neighbor embedding achieves a stronger ability to express the preference of target user, which also demonstrates that neighbors with fine-grained interests play a greater role in collaborative recommendation. Similarly, for the BK-10C dataset, after removing the four att-view neighbor embedding, the MAE and RMSE values change the most, which indicates that four att-view neighbor embedding makes the greatest contribution on modeling the preference of the user. For the Douban dataset, two att-view neighbor embedding is important. These results indicate that multiscale neighbors on different views have a collaborative effect on the user’s preference decisions.

7.3. Visual Explanation Study

In order to further observe the influence of multiscale neighbor embedding on user decision-making, Figure 11, Figure 12 and Figure 13 report the attention weight of attribute-view neighbor embedding and interaction-view neighbor embedding of 10 users for three datasets. In Figure 11a, the weight of three att-view is the largest, which indicates that similar neighbors with multiple attributes can have a major contribution on capturing collaborative signals. One att-view neighbor embedding can also capture coarse-grained semantic interests for collaborative filtering. This important insight allows us to select neighbors for users and items with different numbers of attributes, and capture rich signals for collaborative filtering recommendation tasks. In addition, in Figure 11, the user’s final embedding representation is most affected by the four att-view embedding and eight sco-view embedding. By observing the dataset of BK-10Core, we find that most users are used to evaluating the items with high score values, which leads to a large contribution of the high-score-view embedding. Meanwhile, the rating neighbor matrix is relatively sparse, which makes the difference between the weight values of various types of high-score-view embeddings small.

8. Conclusions

In this paper, we propose a multiscale neighbor-aware attention network for collaborative filtering recommendation. The proposed strategy fuses the global semantics of various types of neighbors and important local embedding of multiscale neighbors. Multiple attribute-view neighbors and interaction-view neighbors provide collaborative signals to predict the user’s rating of items. Experiments verify the effectiveness of collaborative contributions of multiscale neighbors for learning user and item representation. The key finding is that the combination of multiscale attribute neighbors and interactive neighbors can improve the accuracy of recommendation, and alleviate the poor performance of recommendation in the case of sparse data. The disadvantage is that the computation of multiscale neighbors requires different graph structures to learn the representation of nodes. The platform can construct the graph structure offline and calculate multiscale neighbors, which can reduce the online resource requirements. However, in the e-commerce scenario, the proposed method can realize targeted personalized recommendation according to different attribute neighbors of users. In particular, for some cold-start users who do not have interactive behaviors, the method can select neighbors with similar attributes for the target user according to one’s social attribute set, and then conduct collaborative filtering recommendation. Some products and services can be recommended based on similar attribute preferences of neighbors for target users. In addition, by combining with the behavioral preferences of group users, we can make rating predictions and recommend popular products for target users.
In future work, by investigating the semantic difference of various attribute and interactive behavior views, we can focus on a consistency study of node representations on different behavior views and improve the accuracy and interpretability of the recommendation system. In addition, heterogeneous types of semantic information from different types of user behaviors such as evaluation, clicking, and buying can describe the ordered semantic interests of the user. We will distinguish the types of multiple interaction behaviors to learn heterogeneous semantic representations and model the sequential relations between different behaviors.

Author Contributions

Methodology, J.Z., F.C. and Q.C.; Software, T.J.; Investigation, J.Z. and Y.L.; Writing—original draft, J.Z.; Writing—review & editing, Y.K. and Q.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by the National Natural Science Foundation of China (nos. 62272286, 62072291), the Natural Science Foundation of Shanxi Province (nos. 20210302123468, 202203021221021, 202203021221001).

Data Availability Statement

Data used in this manuscript consist of publicly available standard benchmark datasets.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, Z.; Cui, Z.; Wu, S.; Zhang, X.; Wang, L. Fi-gnn: Modeling feature interactions via graph neural networks for ctr prediction. In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, Beijing, China, 3–10 November 2019; pp. 539–548. [Google Scholar]
  2. Naghiaei, M.; Rahmani, H.; Deldjoo, Y. CPFair: Personalized Consumer and Producer Fairness Re-ranking for Recommender Systems. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, 11–15 July 2022; pp. 770–779. [Google Scholar]
  3. Shi, S.; Zhang, M.; Liu, Y.; Ma, S. Attention-based adaptive model to unify warm and cold starts recommendation. In Proceedings of the 27th ACM International Conference on Information and Knowledge Management, Turin, Italy, 22–26 October 2018; pp. 127–136. [Google Scholar]
  4. Ge, Y.; Tan, J.; Zhu, Y.; Xia, Y.; Luo, J.; Liu, S.; Fu, Z.; Geng, S.; Li, Z.; Zhang, Y. Explainable Fairness for Feature-aware Recommender Systems. In Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval, Madrid, Spain, 11–15 July 2022; pp. 1–11. [Google Scholar]
  5. Xiao, J.; Ye, H.; He, X.; Zhang, H.; Wu, F.; Chua, T. Attentional factorization machines: Learning the weight of feature interactions via attention networks. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, Beijing, China, 21–22 May 2017; pp. 3119–3125. [Google Scholar]
  6. Wang, X.; He, X.; Cao, Y.; Liu, M.; Chua, T. Kgat: Knowledge graph attention network for recommendation. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 950–958. [Google Scholar]
  7. Koren, Y.; Bell, R.; Volinsky, C. Matrix Factorization Techniques for Recommender Systems. Computer 2009, 42, 30–37. [Google Scholar] [CrossRef]
  8. Salakhutdinov, R.; Mnih, A. Probabilistic matrix factorization. In Proceedings of the 20th International Conference on Neural Information Processing Systems, Vancouver, WA, USA, 3–6 December 2007; pp. 849–858. [Google Scholar]
  9. Koren, Y. Factorization meets the neighborhood: A multifaceted collaborative filtering model. In Proceedings of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Las Vegas, NV, USA, 24–27 August 2008; pp. 426–434. [Google Scholar]
  10. Lee, J.; Kim, S.; Lebanon, G.; Singer, Y. Local low-rank matrix approximation. In Proceedings of the 30th International Conference on International Conference on Machine Learning, Atlanta, GA, USA, 16–21 June 2013; pp. 82–90. [Google Scholar]
  11. Ling, G.; Lyu, M.; King, I. Ratings meet reviews, a combined approach to recommend. In Proceedings of the 8th ACM Conference on Recommender Systems, Foster City, CA, USA, 6–10 October 2014; pp. 105–112. [Google Scholar]
  12. Rendle, S.; Freudenthaler, C. Improving pairwise learning for item recommendation from implicit feedback. In Proceedings of the 7th ACM International Conference on Web Search and Data Mining, New York, NY, USA, 24–28 February 2014; pp. 273–282. [Google Scholar]
  13. Ning, X.; Karypis, G. Sparse linear methods with side information for top-n recommendations. In Proceedings of the 6th ACM Conference on Recommender Systems, Dublin, Ireland, 9–13 September 2012; pp. 155–162. [Google Scholar]
  14. Sedhain, S.; Menon, A.; Sanner, S.; Xie, L. Autorec: Autoencoders meet collaborative filtering. In Proceedings of the 24th International Conference on World Wide Web, Florence, Italy, 18–22 May 2015; pp. 111–112. [Google Scholar]
  15. Park, J.; Nam, K. Group recommender system for store product placement. Data Min. Knowl. Disc. 2019, 33, 204–229. [Google Scholar] [CrossRef]
  16. Zheng, J.; Liu, J.; Shi, C.; Zhuang, F.; Li, J.; Wu, B. Dual Similarity Regularization for Recommendation. In Proceedings of the 2016 Pacific-Asia Conference on Knowledge Discovery and Data Mining, Auckland, New Zealand, 19–22 April 2016; pp. 542–554. [Google Scholar]
  17. He, X.; Liao, L.; Zhang, H.; Nie, L.; Hu, X.; Chua, T. Neural collaborative filtering. In Proceedings of the 26th International Conference on World Wide Web, Perth, Australia, 3–7 April 2017; pp. 173–182. [Google Scholar]
  18. Lian, J.; Zhou, X.; Zhang, F.; Chen, Z.; Xie, X. DeepFM: Combining explicit and implicit feature interactions for recommender systems. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, London, UK, 19–23 August 2018; pp. 1754–1763. [Google Scholar]
  19. Cai, Y.; Ke, W.; Cui, E.; Yu, F. A deep recommendation model of cross-grained sentiments of user reviews and ratings. Inf. Process Manag. 2022, 59, 102842. [Google Scholar] [CrossRef]
  20. Cheng, H.; Koc, L.; Harmsen, J.; Shaked, T.; Chandra, T.; Aradhye, H.; Anderson, G.; Corrado, G.; Chai, W.; Ispir, M.; et al. Wide & deep learning for recommender systems. In Proceedings of the 1st Workshop on Deep Learning for Recommender Systems, Boston, MA, USA, 15–19 September 2016; pp. 7–10. [Google Scholar]
  21. He, X.; Chua, T. Neural factorization machines for sparse predictive analytics. In Proceedings of the 40th International ACM SIGIR conference on Research and Development in Information Retrieval, New York, NY, USA, 7–12 February 2017; pp. 355–364. [Google Scholar]
  22. Wang, X.; Wang, R.; Shi, C.; Song, G.; Li, Q. Multi-component graph convolutional collaborative filtering. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 6267–6274. [Google Scholar]
  23. Magron, P.; Fevotte, C. Neural content-aware collaborative filtering for cold-start music recommendation. Data Min. Knowl. Disc. 2022, 36, 1971–2005. [Google Scholar] [CrossRef]
  24. Jin, B.; Gao, C.; He, X.; Jin, D.; Li, Y. Multi-behavior recommendation with graph convolutional networks. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Xi’an, China, 25–30 July 2020; pp. 659–668. [Google Scholar]
  25. Su, Z.; Dou, Z.; Zhu, Y.; Qin, X.; Wen, J. Modeling Intent Graph for Search Result Diversification. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Online, 11–15 July 2021; pp. 736–746. [Google Scholar]
  26. Li, H.; Chen, Z.; Li, C.; Xiao, R.; Deng, H.; Zhang, P.; Liu, Y.; Tang, H. Path-based Deep Network for Candidate Item Matching in Recommenders. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Online, 11–15 July 2021; pp. 1493–1502. [Google Scholar]
  27. Tai, C.; Huang, L.; Huang, C.; Ku, L. User-Centric Path Reasoning towards Explainable Recommendation. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Online, 11–15 July 2021; pp. 879–889. [Google Scholar]
  28. Duan, H.; Zhu, Y.; Liang, X.; Zhu, Z.; Liu, P. Multi-feature fused collaborative attention network for sequential recommendation with semantic-enriched contrastive learning. Inf. Process Manag. 2023, 60, 103416. [Google Scholar] [CrossRef]
  29. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 5998–6008. [Google Scholar]
  30. Wang, X.; Jin, H.; Zhang, A.; He, X.; Xu, T.; Chua, T. Disentangled graph collaborative filtering. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Xi’an, China, 25–30 July 2020; pp. 1001–1010. [Google Scholar]
  31. Chen, J.; Zhang, H.; He, X.; Nie, L.; Liu, W.; Chua, T. Attentive collaborative filtering: Multimedia recommendation with item-and component-level attention. In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval, Tokyo, Japan, 7–11 August 2017; pp. 335–344. [Google Scholar]
  32. Tang, X.; Wang, T.; Yang, H.; Song, H. AKUPM: Attention-enhanced knowledge-aware user preference model for recommendation. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 1891–1899. [Google Scholar]
  33. Wang, X.; He, X.; Wang, M.; Feng, F.; Chua, T. Neural graph collaborative filtering. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, Paris, France, 21–25 July 2019; pp. 165–174. [Google Scholar]
  34. Wang, Z.; Lin, G.; Tan, H.; Chen, Q.; Liu, X. CKAN: Collaborative knowledge-aware attentive network for recommender systems. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Xi’an, China, 25–30 July 2020; pp. 219–228. [Google Scholar]
  35. Niu, G.; Li, Y.; Tang, C.; Geng, R.; Dai, J.; Liu, Q.; Wang, H.; Sun, J.; Huang, F.; Si, L. Relational Learning with Gated and Attentive Neighbor Aggregator for Few-Shot Knowledge Graph Completion. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, Online, 11–15 July 2021; pp. 213–222. [Google Scholar]
  36. Xiao, W.; Zhao, H.; Pan, H.; Song, Y.; Zheng, V.; Yang, Q. Social explorative attention based recommendation for content distribution platforms. Data Min. Knowl. Disc. 2021, 35, 533–567. [Google Scholar] [CrossRef]
  37. Ye, H.; Song, Y.; Li, M.; Cao, F. A new deep graph attention approach with influence and preference relationship reconstruction for rate prediction recommendation. Inf. Process Manag. 2023, 60, 103439. [Google Scholar] [CrossRef]
  38. He, X.; Deng, K.; Wang, X.; Li, Y.; Zhang, Y.; Wang, M. Lightgcn: Simplifying and powering graph convolution network for recommendation. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Xi’an, China, 25–30 July 2020; pp. 639–648. [Google Scholar]
  39. Zheng, Y.; Tang, B.; Ding, W.; Zhou, H. A Neural Autoregressive Approach to Collaborative Filtering. In Proceedings of the 33rd International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; pp. 764–773. [Google Scholar]
  40. Kipf, T.; Welling, M. Semi-supervised classification with graph convolutional networks. Commun. ACM 2007, 50, 36–44. [Google Scholar]
  41. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Liò, P.; Bengio, Y. Graph Attention Networks. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  42. Liu, X.; Wu, J.; Zhou, C.; Pan, S.; Cao, Y.; Wang, B. Task-adaptive neural process for user cold-start recommendation. In Proceedings of the International World Wide Web Conference, Ljubljana, Slovenia, 19–23 April 2021; pp. 1306–1316. [Google Scholar]
Figure 1. The framework of the multiscale neighbor-aware attention network.
Figure 1. The framework of the multiscale neighbor-aware attention network.
Electronics 12 04372 g001
Figure 2. MAE results of different methods on the ML-100kr dataset with different sparsity proportions.
Figure 2. MAE results of different methods on the ML-100kr dataset with different sparsity proportions.
Electronics 12 04372 g002
Figure 3. RMSE results of different methods on the ML-100kr dataset with different sparsity proportions.
Figure 3. RMSE results of different methods on the ML-100kr dataset with different sparsity proportions.
Electronics 12 04372 g003
Figure 4. MAE results of different methods on the BK-10C dataset with different sparsity proportions.
Figure 4. MAE results of different methods on the BK-10C dataset with different sparsity proportions.
Electronics 12 04372 g004
Figure 5. RMSE results of different methods on the BK-10C dataset with different sparsity proportions.
Figure 5. RMSE results of different methods on the BK-10C dataset with different sparsity proportions.
Electronics 12 04372 g005
Figure 6. MAE results of different methods on the Douban dataset with different sparsity proportions.
Figure 6. MAE results of different methods on the Douban dataset with different sparsity proportions.
Electronics 12 04372 g006
Figure 7. RMSE results of different methods on the Douban dataset with different sparsity proportions.
Figure 7. RMSE results of different methods on the Douban dataset with different sparsity proportions.
Electronics 12 04372 g007
Figure 8. Ablation study of multiscale neighbor embedding on the ML-100kr dataset.
Figure 8. Ablation study of multiscale neighbor embedding on the ML-100kr dataset.
Electronics 12 04372 g008
Figure 9. Ablation study of multiscale neighbor embedding on the BK-10C dataset.
Figure 9. Ablation study of multiscale neighbor embedding on the BK-10C dataset.
Electronics 12 04372 g009
Figure 10. Ablation study of multiscale neighbor embedding on the Douban dataset.
Figure 10. Ablation study of multiscale neighbor embedding on the Douban dataset.
Electronics 12 04372 g010
Figure 11. Attention weight of multiscale neighbor embedding for 10 users on the ML-100kr dataset.
Figure 11. Attention weight of multiscale neighbor embedding for 10 users on the ML-100kr dataset.
Electronics 12 04372 g011
Figure 12. Attention weight of multiscale neighbor embedding for 10 users on the BK-10C dataset.
Figure 12. Attention weight of multiscale neighbor embedding for 10 users on the BK-10C dataset.
Electronics 12 04372 g012
Figure 13. Attention weight of multiscale neighbor embedding for 10 users on the Douban dataset.
Figure 13. Attention weight of multiscale neighbor embedding for 10 users on the Douban dataset.
Electronics 12 04372 g013
Table 1. Statistical information of experimental datasets.
Table 1. Statistical information of experimental datasets.
DatasetsUsersItemsInteractionsRatingSparsity
ML-100kr9431682100,0001–594.12%
BK-10C1820203041,4561–1098.87%
Douban30226971195,4931–599.07%
Table 2. MAE and RMSE results of different methods on three datasets.
Table 2. MAE and RMSE results of different methods on three datasets.
MethodML-100krBK-10CDouban
MAE RMSE MAE RMSE MAE RMSE
NCF0.74570.93421.16111.52880.57810.7304
NGCF0.72980.91951.12411.47760.57680.7271
GCN0.72530.91781.13331.47720.57660.7259
LightGCN0.72600.91821.11661.47940.57090.7213
GAT0.72570.91861.12501.48130.57700.7238
AFM0.73190.92381.11701.47860.56430.7136
Wide&Deep0.72040.91521.11511.48070.56540.7141
ACCM0.71450.90271.19831.53730.57890.7301
TANP0.78810.97571.20811.53360.57670.7274
MSNAN + NGCF0.69100.88421.10291.45860.55360.7069
MSNAN + GCN0.70080.88901.10381.45980.56430.7091
MSNAN + LightGCN0.69490.88151.10031.45630.55170.7003
MSNAN + GAT0.69730.89011.10471.46070.56170.7034
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zheng, J.; Jing, T.; Cao, F.; Kang, Y.; Chen, Q.; Li, Y. A Multiscale Neighbor-Aware Attention Network for Collaborative Filtering. Electronics 2023, 12, 4372. https://doi.org/10.3390/electronics12204372

AMA Style

Zheng J, Jing T, Cao F, Kang Y, Chen Q, Li Y. A Multiscale Neighbor-Aware Attention Network for Collaborative Filtering. Electronics. 2023; 12(20):4372. https://doi.org/10.3390/electronics12204372

Chicago/Turabian Style

Zheng, Jianxing, Tengyue Jing, Feng Cao, Yonghong Kang, Qian Chen, and Yanhong Li. 2023. "A Multiscale Neighbor-Aware Attention Network for Collaborative Filtering" Electronics 12, no. 20: 4372. https://doi.org/10.3390/electronics12204372

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop