Next Article in Journal
Analysis of a Pre-Emptive Two-Priority Queuing System with Impatient Customers and Heterogeneous Servers
Next Article in Special Issue
A Semantic Enhancement Framework for Multimodal Sarcasm Detection
Previous Article in Journal
Epidemic Waves in a Stochastic SIRVI Epidemic Model Incorporating the Ornstein–Uhlenbeck Process
Previous Article in Special Issue
Multimodal Interaction and Fused Graph Convolution Network for Sentiment Classification of Online Reviews
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Syntactic Structure-Enhanced Dual Graph Convolutional Network for Aspect-Level Sentiment Classification

School of Electronics and Information Engineering, South China Normal University, Foshan 528225, China
*
Author to whom correspondence should be addressed.
Mathematics 2023, 11(18), 3877; https://doi.org/10.3390/math11183877
Submission received: 5 August 2023 / Revised: 29 August 2023 / Accepted: 10 September 2023 / Published: 11 September 2023

Abstract

:
Aspect-level sentiment classification (ALSC) is a fine-grained sentiment analysis task that aims to predict the sentiment of the given aspect in a sentence. Recent studies mainly focus on using the Graph Convolutional Networks (GCN) to deal with both the semantics and the syntax of a sentence. However, the improvement is limited since the syntax dependency trees are not aspect-oriented and the exploitation of syntax structure information is inadequate. In this paper, we propose a Syntactic Structure-Enhanced Dual Graph Convolutional Network (SSEDGCN) model for an ALSC task. Firstly, to enhance the relation between aspect and its opinion words, we propose an aspect-wise dependency tree by reconstructing the basic syntax dependency tree. Then, we propose a syntax-aware GCN to encode the new tree. For semantics information learning, a semantic-aware GCN is established. In order to exploit syntactic structure information, we design a syntax-guided contrastive learning objective that makes the model aware of syntactic structure and improves the quality of the feature representation of the aspect. The experimental results on three benchmark datasets show that our model significantly outperforms the baseline models and verifies the effectiveness of our model.

1. Introduction

Aspect-level sentiment classification (ALSC) is a fine-grained task in the field of sentiment analysis. The main purpose of ALSC is to identify the user-expressed sentiment polarity at aspect-level [1]. In practical use, ALSC not only focuses on analyzing opinions in a given text but also looks into the aspects and its sentiment, and thus, giving a much clearer understanding. For example, in the sentence “I am never disappointed at the system and software”, the sentiment toward aspects “system” and “software” are both positive (Figure 1).
A key issue for addressing ALSC lies in modeling the relationship between the aspect and its opinion words. Early studies [2,3,4,5,6,7,8] principally focus on integrate attention mechanisms into recurrent neural networks (RNNs) to capture the opinion words that semantically relate to the aspect. The devising of pre-trained model BERT allows to exploit abundant semantic information in ALSC tasks. Song et al. proposed an attention-based encoder and a BERT-SPC model to learn features of the aspect and its context words [9]. By contrast, the syntax-based models also pave a way for resolving ALSC issues for sentences of complex syntactic structures. The application of graph neural networks to syntax dependency tree encoding is most pronounced [10,11,12,13,14,15,16,17]. Both GCN and GAT are capable of learning syntactic-based feature representations by aggregating neighboring node features through node connection on a syntax dependency tree. Zhang and Song [10,11] took GCN to encode the syntax dependency tree and achieve satisfying results, indicating the introduction of sentence syntax benefits the relation modeling. Note that syntax-based models have deficiencies in syntactic-insensitive sentences; thus, the integration of syntax and semantics is developed [18,19]. Zhang et al. [20] defined a syntactic mask matrix and enhance the interaction between semantics and syntax by multiplying the syntactic mask matrix with the semantic-based attention matrix.
Despite the improvement in ALSC, existing GCN-based methods have two limitations. On the one hand, for syntax-based models, the syntax dependency trees generated by the parser are not aspect-oriented. As such, the dependency tree may miss important connections to the aspect. As presented in Figure 1, if the context toward aspect “software” is processed using one-layer GCN, only the information from “system” and “and” can be aggregated. By contrast, no connection is built between the aspect “software” with its opinion words “never disappointed”. Although an aspect-based dependency tree is proposed [16], the sentiment is still mis-identified due to its neglecting the information flow in the basic syntax dependency tree. Taking the words “never disappointed” as an example, if the syntactic connection between “never” and “disappointed” is disregarded, the sentiment polarity can be classified as negative. The main reason is that neither “never” nor “disappointed” expresses a positive sentiment. On the other hand, GCN aggregates node feature merely via the relation in the graph. Such GCN-based models mainly focus on node features and do not fully leverage syntactic structure, which may affect the performance of GCN-based models in ALSC.
Aiming to address the issues mentioned above, we propose Syntactic Structure-Enhanced Dual Graph Convolutional Network for ALSC task. To start with, an aspect-wise dependency tree is developed, which preserves the basic syntax and enhances the connection toward the aspect. The aspect-wise dependency tree is encoded via GCN with the integration of dependency relation aware attention mechanism. Likewise, the sentence semantics is also processed by GCN and self-attention mechanism. To fully consider the syntactic structure, we design a contrastive learning objective with a contrastive coefficient based on the syntactic structure similarity which is characterized by the distribution of anonymous walks. Training with this objective encourages the model to generate closer representations for those aspects with the same labels and similar syntactic structure, thereby further improving the classification performance.
The contributions of our work are summarized as follows:
  • A novel aspect-wise dependency tree is established to overcome the deficiencies of a classical syntax dependency tree. The relation between aspect and its opinion words is enhanced, while the basic syntax is preserved.
  • On the task of ALSC, the SSEDGCN model is proposed, which deals with both the syntax and the semantics. Moreover, a contrastive learning module is exploited to learn the feature representation of the aspect.
  • Experimental results on three benchmark datasets reveal that our model is a competitive alternative comparable with the state of the art.
The paper is mainly divided into five sections. In the introduction, we present an overview of the article and propose our solution to address the limitations of current ALSC methods. Section 2 provides a summary of research related to our work. In Section 3, we offer a detailed description of the proposed model and its individual modules. In Section 4, experiments and result analysis are performed on three public datasets. Finally, concluding remarks are given in Section 5.

2. Related Work

2.1. Aspect-Level Sentiment Analysis

As pointed out in the introduction, ALSC involves with classified the sentiment towards a given aspect according to a predefined set of sentiment polarities. Early research focuses on employing RNN-based method, together with the integration of attention mechanisms or knowledge distillation [21,22], to extract aspect-related information. In recent years, advances in GCN-based algorithms significantly improve the working performance in ALSC tasks. The ALSC methods can be loosely classified into the three categories, i.e., semantic-based models, syntactic-based models, and their integration. For semantic-based models, the most widely-used models are developed based on deep neural networks and attention mechanisms, which are capable of modeling the semantic relation between aspect and its contexts [2,3,4,5,6,7,8]. Another main focus is the exploiting of syntax [10,11,12,13,14,15,16,17], because the syntactic information can be applied to set the connection of aspect to its opinion words. On the task of ALSC, the GCNs are first employed in syntax-based models to encode the syntax dependency tree of the given sentence, and thus, model the syntactic relation between aspect and the contexts [10,11]. Wang et al. reconstructed the syntax dependency tree by pruning the irrelevant edges and encode the revised dependency tree using GAT [16]. Considering the interaction between syntactic and semantic information, multi-channel GCN-based approaches has been proposed to extract syntactic and semantic information [18,19,23]. Zhang et al. [20] generated an attention matrix and a syntactic mask matrix based on word-syntactic distances, which are processed to enhance the interaction between semantics and syntax.

2.2. Random Anonymous Walks

Given a graph G = ( V , E ) , with V = { v 1 , v | V | } and E = { ( v i , v j ) } separately representing the node set and the edge set, the random anonymous walk aims to capture the structural pattern of the graph [24]. Let w = ( v 1 , v 2 , v l ) be a random walk with length l and ( v i , v i + 1 ) E . We shall define the random anonymous walk with respect to w as:
a w ( w ) = ( D I S ( w , v 1 ) , D I S ( w , v 2 ) , D I S ( w , v l ) )
where D I S ( w , v i ) represents the number of distinct nodes in w when v i first appears in w. Specifically, the anonymous walks of length l can be written as ω 1 l , ω 2 l , in line with their lexicographical order. For example, the four-length anonymous walks can be given as ω 1 4 = ( 1 , 2 , 3 , 1 ) , ω 2 4 = ( 1 , 2 , 3 , 2 ) , ω 3 4 = ( 1 , 2 , 1 , 3 ) , etc.
On this occasion, an intuitive description of anonymous walks is also provided. Distinguishing from random walks, the essence of anonymous walks lies in denoting the underlying pattern of random walks, regardless of the specific node being visited. According to Figure 2, all three random walks can be mapped into one anonymous walk referring to certain structure, such as triadic closure.

2.3. Graph Neural Networks

Graph Neural Networks (GNNs) [25,26,27] have emerged as a powerful framework for learning representations of graph data. GNNs aim to learn expressive node representations by leveraging the graph structure and node features. Subsequently, researchers have explored the combination of self-supervised learning and GNNs [28,29,30]. For instance, You [28] proposed a graph-contrastive learning framework that enhances graph representations through data augmentation. Wei [29] utilized contrastive learning and the information bottleneck principle to enhance recommendation systems based on graphs. In the field of Aspect-Level Sentiment Classification (ALSC), GNNs have demonstrated significant potential. Numerous studies [10,11,12,13,14,15,16,17,18,19,20] have employed GNNs to capture the syntactic information of sentences, where nodes represent words and edges indicate dependencies. This enables the generation of representation vectors for nodes based on their neighboring features. With the increasing performance of GNN-based models, it has been observed that they effectively enhance contextual and aspectual dependencies in ABSA tasks through the utilization of GNNs.

2.4. Contrastive Learning

In recent years, advances in contrastive learning algorithms significantly improve the working performance in a variety of tasks. The main purpose of contrastive learning is to pull an anchor and a positive sample closer in the feature space while pushing apart the anchor with negative samples. Motivated by its success in computer vision tasks [31,32,33], research in contrastive learning-based methods for NLP primarily focuses on the learning of representations. Gao et al. [34] proposed a contrastive learning framework to derive the sentence embeddings, aiming at improving the precision of sentence representations through contrastive learning. Wang et al. [35] applied a contrastive learning-enhanced KNN mechanism to Multi-Label Text Classification (MLTC) task, based on which the representation quality of the retrieved neighbors can be upgraded. For the ALSC task, Liang et al. [36] took contrastive learning to extract aspect-invariant and aspect-dependent features to discriminate the sentiment features among the sentiment pattern and polarity perspectives.

3. Methodology

Figure 3 shows the architecture of the SSEDGCN model. The proposed model consists of five major components, i.e., a Sentence Encoder, a Syntax-aware GCN, a Semantic-aware GCN, a Syntax-guided contrastive learning module, and a Sentiment classifier. More details of each component are described as follows. We start with an aspect-oriented dependency tree constructed in our model.

3.1. Aspect-Wise Dependency Tree

In syntax-based ALSC methods, most existing models [11,17,37] tend to derive the sentence syntactic information by processing the dependency relation. Typically, the syntax dependency tree that depicts the sentence dependency relation is directly developed via dependency parsers. Nevertheless, such dependency trees are not aspect-oriented, in which case the aspect word may not connect to its opinion words in line with the sentence syntax dependency. Notably, the ALSC task aims to assign the predominant focus to the aspect instead of the root nodes of the syntax dependency tree. For this reason, an aspect-wise dependency tree is devised based on the classical syntax dependency tree. Specifically, we designate the aspect as the root node and connect all other nodes to it. Concurrently, we assign syntactic dependency relationships to these connecting edges, based on their respective syntactic distances.
Figure 4 shows the basic syntax dependency tree of the sentence and the aspect-wise dependency tree. Comprehensively, we start with obtaining the classical sentence syntax dependency tree via syntactic parser. Then, each unidirectional connection is revised to a bidirectional relation. The aspect-wise dependency tree is established by setting the aspect as the root node of the basic syntax dependency tree. With respect to an aspect composed of multiple words, all these words are considered as one root node. For nodes that have no direct syntactic connection with the aspect in the basic syntax dependency tree, a directional relationship to the aspect is developed. According to Figure 4, the directional relation between the node and the aspect is denoted as [n: con] where n represents the shortest distance between the two nodes in the original syntactic tree. For example, in the sentence “I love the system and the software”, the relation between “software” and “and” is defined as [2: con]. The reason is that these words are not exactly connected to each other in the basic syntax dependency tree, but the word “software” reaches “and” through the path of software-system-and. That is, the distance between “software” and “and” in the basic syntactic dependency tree is 2, so the connection between them is written as [2: con].
Two primary superiorities can be observed from the aspect-wise dependency tree. For one thing, each aspect retains a distinctive dependency tree, avoiding the condition that multiple aspects in one sentence possess the same syntax dependency tree. For another, every single node directionally relates to the aspect, which enhances the connection between the opinion word and the aspect word.

3.2. Sentence Encoder

Let s = { w 1 , w 2 , , w n } be a n-word sentence with aspect a = { w s t a r t , w s t a r t + 1 , , w e n d } , where s t a r t and e n d represent the starting index and the ending index of a, respectively. In light of the recent success of BERT [38] in word embedding, we employ a pre-trained BERT model to obtain the hidden representation of the sentence. The sequence “ [ C L S ] s [ S E P ] a [ S E P ] ” is generated and sent to BERT encoder. Thus, we obtain the contextual feature vector H = { h 1 , h 2 , , h n } :
H = B E R T { [ C L S ] s [ S E P ] a [ S E P ] }
where H R n × d and d represent the hidden layer dimension. Notably, the subvector h a = { h a 1 , h a 2 , h a m } indicates the aspect representation.

3.3. Syntax-Aware GCN with DRA Attention

As previously described [10,11,12], we transform the aspect-wise dependency tree into a syntactic graph G s y n ( A s y n , H ) , where A s y n R n × n refers to the adjacency matrix of the graph and H = { h 1 , h 2 , , h n } is the hidden representation of the sentence. We take H s y n ( 0 ) as the initialized feature representation of nodes and feed it into the syntax-aware GCN.
Normally, A GCN iteratively updates a node representation by averagely aggregating neighboring nodes, which is carried out based on the element in the adjacency matrix. Specifically, the element A i j s y n stands for the relation between node i and node j. Only if A i j s y n = 1 are the two nodes regarded as connected, whereas since the dependencies of different neighboring nodes toward the aspect vary, the specific dependency relation and its effect have to be considered during aggregating. Thereby, dependency relation aware attention (DRA attention) is performed in the process of aspect node updating. In such a manner, the dependency weights are assigned to the aspect-neighboring nodes for information aggregation. We map the dependency relations of the aspect node and its neighboring nodes into vector representations. The dependency relation between nodes is, thus, incorporated into the graph convolution. The attentive weights are taken to control the information flow from neighboring nodes to the aspect. Computation with DRA attention is conducted using the following equations:
H s y n ( 0 ) = { h 1 s y n , 0 , h 2 s y n , 0 , h n s y n , 0 } = H
h i s y n , l = σ ( j = 1 n A i j s y n W s y n l h j s y n , l 1 + b s y n l )
A i j s y n = A i j s y n i ( s t a r t , e n d ) δ i j i ( s t a r t , e n d )
δ i j = exp ( g i j ) j = 1 N i exp ( g i j )
g i j = σ ( r e l u ( r i j W 1 + b 1 ) W 2 + b 2 )
where r i j indicates the dependency relation between words w i and w j , h i s y n , l R d is the embedding of node i in the l-th GCN layer, δ i j is a normalized attention coefficient for the DRA attention mechanism, W s y n l and b s y n l are the trainable parametric matrix and bias in the l-th GCN layer, respectively, and W 1 , W 2 and b 1 , b 2 are the trainable parametric matrices and biases for computing the attention weights.

3.4. Semantic-Aware GCN with Self-Attention

The working performance of ALSC methods is not just affected by the syntax, but also the semantics [20]. In this way, the semantic-aware module is introduced. Similar to the syntax-aware GCN, we define G s e m ( A s e m , H ) as the semantic graph, with A s e m R n × n as the semantic graph adjacency matrix. H = { h 1 , h 2 , , h n } is also taken as the initialized feature representations H s e m ( 0 ) of nodes in the semantic-aware GCN. With respect to semantic information learning, the self-attention mechanism is exploited to build the semantic adjacency matrix A s e m R n × n . The idea behind this approach is that the attention matrix generated by self-attention can represent semantic correlations between words.
H s e m ( 0 ) = { h 1 s e m , 0 , h 2 s e m , 0 , h n s e m , 0 } = H
A s e m = softmax ( ( H s e m ( 0 ) W s e m , k ) ( H s e m ( 0 ) W s e m , q ) T d )
h i s e m , l = σ ( j = 1 n A i j s e m W s e m l h j s e m , l 1 + b s e m l )
where h i s e m , l R d is the embedding of node i in the l-th GCN layer, d is the dimension of the node feature representation, W s e m , k , W s e m , q are trainable parametric matrices for self-attention mechanism, and W s e m l and b s e m l are trainable parametric matrix and bias of the l-th layer in GCN, respectively.

3.5. Sentiment Classifier

For the purpose of information interaction, the BiAffine mechanism [17,18] is exploited to integrate the syntactic and semantic information in advance to sentiment classification. The processing of outputs from syntax-aware GCN and semantic-aware GCN are presented in Equations (10) and (11):
H s e m ( l ) = softmax ( H s e m ( l ) W 3 ( H s y n ( l ) ) T ) H s y n ( l )
H s y n ( l ) = softmax ( H s y n ( l ) W 4 ( H s e m ( l ) ) T ) H s e m ( l )
where W 3 and W 4 are trainable parametric matrices.
The pooling operation f ( · ) is performed to derive the semantic representation h a s e m and the syntactic representation h a s y n of the aspect. Then, h a s e m and h a s y n are concatenated to obtain the final aspect representation r, which is fed into the linear layer and the softmax classifier to identify the sentiment polarity of the aspect.
h a s e m = f ( { h a 1 s e m , h a 2 s e m , , h a m s e m } )
h a s y n = f ( { h a 1 s y n , h a 2 s y n , , h a m s y n } )
r = [ h a s e m , h a s y n ]
y ^ ( a ) = softmax ( W y r + b y )
where W y and b y are the trainable parametric matrix and the bias, respectively, and { h a 1 s e m , h a 2 s e m , h a m s e m } R m × d and { h a 1 s y n , h a 2 s y n , h a m s y n } R m × d stand for aspect sequence derived from H s e m ( L ) R n × d and H s y n ( L ) R n × d .

3.6. Syntax-Guided Contrastive Learning

Inspired by [32,35], the syntax-guided contrastive learning is designed to learn the feature representation of the aspect. Specifically, we design a contrastive learning objective with a contrastive coefficient based on the syntactic structure similarity for positive instance pair. For each input aspect, the remaining aspects with the same sentiment polarity within the same batch are considered as the positive examples, otherwise as negative examples. Then, we use the distribution of anonymous walks to characterize syntactic graph structures since the anonymous walk is a powerful tool that represents graph structural patterns. Subsequently, we employ KL divergence to compute the differences in the anonymous walk distribution of positive samples, such as f G ( i ) and f G ( j ) in Figure 5. The resulting KL divergence is then used as the contrast coefficient β i j in contrastive learning to measure the similarity of syntactic structure between two positive samples.
As reported by [39], the number of anonymous walks with length l in a given graph is fixed. For instance, there are only two kinds of anonymous walks ω 2 with length 2, i.e., ω 1 2 = ( 1 , 1 ) and ω 2 2 = ( 1 , 2 ) . In our model, we take the aspect node as the starting node of the sentence s in the syntactic graph. As Figure 5 show, a set of γ random walks R W with the length of l is sampled, which corresponds to a set of η different anonymous walks ( ω 1 l , ω 2 l , , ω η l ) . The empirical distribution of these γ random walks over the η anonymous walks is thus determined, which is:
f G ( s ) = ( p s ( ω 1 l ) , p s ( ω 2 l ) , , p s ( ω η l ) )
p s ( ω i l ) = w R W I ( a w ( w ) = ω i l ) γ , i = 0 , 1 , η
where p s ( ω i l ) refers to the probability of an anonymous walk ω i l with the starting point of the aspect node, I is the indicator function, and f G ( s ) denotes the distribution of γ random walks over η anonymous walks.
Given an anchor aspect r i , we designate other aspect with the same sentiment P = { j B , y j = y i , j i } as the positive example set of r i , and otherwise as a negative example set. The syntax-guided contrastive loss is written as:
L C o n = 1 B i B 1 | P | j P L ( r i , r j )
L ( r i , r j ) = β i j log exp ( s i m ( r i , r j ) / τ ) k B , k i exp ( s i m ( r i , r k ) / τ )
β i j = C ( i , j ) m P C ( i , m )
C ( i , j ) = 1 K L ( f G ( i ) f G ( j ) )
where B indicates the total number of samples in the mini-batch, τ is the temperature coefficient in contrastive learning, and s i m ( r i , r j ) is taken to calculate the cosine similarity of the vectors r i and r j . For the positive example pair ( r i , r j ) , the more similar the local structure of the aspect nodes is, the smaller the KL divergence can be calculated, and hence, the larger C ( i , j ) is determined. Furthermore, a larger C ( i , j ) results in a larger contrastive coefficient β i j , which enlarges the loss term L ( r i , r j ) of ( r i , r j ) . In this way, the distance between their representation vectors will be optimized to be closer.

3.7. Model Training

The loss function in our model is given by:
L ( θ ) = L C r o s s + λ L C o n
L C r o s s = ( s , a ) D c C log y ^ ( a )
where θ refers to the set of trainable parameters, λ is the contrastive learning loss weight, L C r o s s is the cross entropy loss function for ALSC task, D is the set of all sentence-aspect pairs, and C is the class of sentiment polarity.

4. Experiment

4.1. Datasets

We evaluate our model on three publicly available benchmark datasets, i.e., Restaurant and Laptop from SemEval 2014 Task 4 [40] and Twitter [41]. Each sample contains a sentence, with the aspect (single or multiple words) and its corresponding sentiment polarity (positive, negative or neutral). The statistics of the datasets are shown in Table 1.

4.2. Implementation Details

In this experiment, the Stanford parser (https://stanfordnlp.github.io/CoreNLP/, accessed on 2 November 2022) is employed to build the basic syntax-dependent tree, based on which the aspect-wise dependency tree is re-constructed. The pre-trained bert-base-uncased model is used for sentence encoding. Notably, the dimension of the relation embedding between aspect nodes and the remaining nodes is set to 100 while the hidden layer dimension for both GCN modules is 384. In addition, the layer number for SynGCN and SemGCN is set to 2 to prevent overfitting. The dropout function, with a dropout rate of 0.1, is used in both modules. With respect to anonymous walks, the length l is 6 and the sampling number γ is 100. The weight λ of contrastive loss for contrastive learning scheme is set to 0.4, 0.6, and 0.1 for the three datasets. Besides, the Adam optimizer is adopted with a learning rate of 0.002. The SSEDGCN model is trained for 20 epochs with a batch size of 16. We use accuracy (Acc) and Macro-F1 (F1) to evaluate the model performance, which are the primary evaluation measures used in ALSC models.

4.3. Baseline Methods

The following methods are taken as the baselines in the experiment:
  • CDT+BERT [11]: The first model that employs GCN to learn syntactic information based on dependency trees and generate aspect node representation. The model is trained for 20 epochs with a batch size of 16 and a learning rate of 0.02.
  • BERT-SPC [9]: The sentence-aspect pair is fed into BERT model, whose token output is used for sentiment classification.
  • R-GAT+BERT [16]: An aspect-specific dependency tree is constructed, which is further encoded using the relational graph attention network. The model is trained for 30 epochs with a batch size of 16 and a learning rate of 0.00005.
  • DGEDT+BERT [17]: A dual-transformer structure is devised, which learns the flat representations via the Transformer and the graph-based representations via the corresponding dependency graph in an iterative interaction manner. The model is trained for 50 epochs with a batch size of 32 and a learning rate of 0.001.
  • T-GCN+BERT [37]: A type-aware graph convolutional network is proposed to encode distinguishing dependency types, while the attentive layer ensemble is used to learn the contextual information within GCN layers. The model is trained for 30 epochs with a batch size of 16 and a learning rate of 0.00002.
  • DualGCN+BERT [18]: A dual-channel GCN is developed to tackle both the syntax and the semantics. The orthogonal regularization and differential regularization are also applied. The model is trained for 20 epochs with a batch size of 16 and a learning rate of 0.002.
  • DMGCN+BERT [19]: A dynamic and multi-channel graph convoiution network is proposed to encode the syntactic and semantic information for aspect-based sentiment analysis. The model is trained for 50 epochs with a batch size of 32 and a learning rate of 0.001.
  • TCL+BERT [42]: A triple contrastive learning network, focusing on achieving the alignment of syntactic and semantic features for sentiment classification. The model is trained for 15 epochs with a batch size of 16 and a learning rate of 0.00002.
  • SSEGCN+BERT [20]: Both the semantic information and the syntactic information is comprehensively learned while the word representations are enhanced via GCN. The model is trained for 15 epochs with a batch size of 16 and a learning rate of 0.00002.
  • DR-BERT [43]: A dynamic re-weighting BERT model is built, which tends to learn the dynamic aspect-oriented semantic information.
  • SSK-GAT+BERT [23]: A novel graph attention network model is proposed to incorporate syntactic, semantic, and knowledge-based features. The model is trained for 15 epochs with a batch size of 16 and a learning rate of 0.000022.
  • HD-GCN+BERT [22]: A hierarchical graph convolutional network is proposed to fuse the outputs of multiple GCN layers as the final representation for prediction. The model is trained for 20 epochs with a batch size of 16 and a learning rate of 0.002.

4.4. Comparison of Results

The overall results on three benchmark datasets are recorded in Table 2. We provide accuracy and Macro-F1 score as evaluation metrics. Among all the methods, SSEDGCN obtains the best and most consistent results in sentiment classification accuracy. In contrast with classical syntax dependency tree-based methods (DGEDT and T-GCN), our model consistently dominates these baselines. The main reason is that the relation between aspect and its opinion words is enhanced in the aspect-wise dependency tree, which facilitates the sentiment classification. Comparing with R-GAT+BERT, which also concerns the encoding of aspect-oriented dependency tree, our model shows its superiority in all evaluation settings. Furthermore, the proposed model outperforms the state-of-the-art methods that deal with semantics and syntax (i.e., DualGCN and SSEGCN). In addition, our model also exceeds the contrastive-learning network that tackles the overall alignment (i.e., TCL+BERT), indicating the significance of explicitly considering syntactic structural information. Therefore, SSEDGCN presents a more precise way to extract the aspect-related information and model both the semantic and syntactic relation between aspect and its corresponding opinion words, which validates the effectiveness of our model.

4.5. Ablation Study

We carry out an ablation study to further analyze the impact of different components in SSEDGCN; see Table 3. The results show that the most significant module for our method is SynGCN. The removal of SynGCN results in the largest accuracy for the datasets Restaurant and Laptop, which indicates the effectiveness of syntactic feature extraction in ALSC. Moreover, the ablating of dependency relation aware attention (DRA attention) in SynGCN also degrades the performance, especially for Twitter. A possible explanation is that the colloquial expression in Twitter introduces an amount noise, which confuses the model in relation setting without the DRA attention. In addition, removing SemGCN lead to different degrees of model performance degradation, which suggests that the working performance of ALSC methods is not just affects by the syntactic information, but also the semantic information. As for the “w/o SGCL” model, the accuracy decreases 0.83%, 1.18%, and 0.9% on Restaurant, Laptop, and Twitter, respectively, demonstrating the capability of contrastive learning in obtaining the aspect representation.
To validate the effectiveness of the aspect-wise dependency tree, we conducted comparative experiments based on the two types of dependency trees. The implementation results are shown in Table 4. From Table 4, it can be observed that the model based on the aspect-wise dependency tree outperform the model based on the classical dependency tree on all three datasets. This indicates the superiority of the aspect-wise dependency tree proposed in our work.

4.6. Attention Visualization

To validate whether the DRA attention mechanism effectively helps aspect nodes capture important nodes, we visualized the attention weight distribution. The experimental results are shown in Figure 6, where brighter colors indicate higher weights and vice versa. Additionally, to compare the effectiveness of the DRA attention mechanism, we also compared it with the classical attention mechanism. In this experiment, the classical attention mechanism refers to using the aspect as a query to calculate attention scores based on the interaction between the aspect and other tokens.
Figure 6 shows the attention weight distribution. It can be observed that, after using the DRA attention mechanism, the opinion word “quick” receives a higher attention score. This indicates that the introduction of syntactic dependency relationships helps aspect nodes capture important nodes. Furthermore, it can be seen that the weight distribution of the classical attention mechanism is more uniform, and the attention score for the opinion word “quick” is only slightly higher than other tokens. We believe that this is because the semantic correlation between “web browsing” and “browser” is high, which leads to an increased attention score for “browser” and reduces the focus on “quick”.

4.7. Impact of the GCN Layer Number

To examine the effect of varying the number of GCN layers, we conducted experiments with the number of GCN layers ranging from 1 to 8, and present the results in Figure 7. The results indicate that our model achieves best performance with two GCN layers, which we adopt in our model. It is noteworthy that a single-layer GCN yields suboptimal accuracy and F1, suggesting that a single layer is insufficient to effectively leverage syntactic and semantic information. Furthermore, we observe a decline trend in performance when the number of GCN layers is excessively large, indicating that an indiscriminate increase in the number of GCN layers may hinder the model’s learning ability due to the sharp increase parameters.

4.8. Impact of the Training/Testing Data Ratio

To investigate the impact of the training/testing data ratio the accuracy, we have included a supplementary experiment using the Restaurant dataset. The initial ratio in the Restaurant dataset is 3.22/1, meaning that the training set contains 3.22 times more samples than the testing set. To better understand the influence of the training/testing data ratio on accuracy, we gradually reduced this ratio by 20% increments. The purpose of this approach is to observe whether there are any changes in model accuracy as the ratio decreases. The experimental results are shown in the Table 5.
Based on our experimental results, we observed a consistent decrease in accuracy as the ratio decreased. We believe that this trend can be attributed to the reduced size and diversity of the training set when the ratio decreases. Consequently, the model receives fewer samples during the training process, which may limit its ability to fully capture the complexity and variability of the data. As a result, the performance of the model is compromised.

4.9. Case Study

The impacts of aspect-wise dependency tree and the proposed model for ALSC are investigated on case examples in the Table 6. To this end, we visualize the aspect and the predicted sentiment polarity. Specifically, the CDT model, which is devised based on classical syntax dependency tree, is taken for comparison. In the first sentence, both CDT and our model set the connection between the aspect “food” and its opinion word “enjoyed”. For the second sentence, the distance between aspect “sandwiches” and the sentiment word “good” is as long as four hops, as indicated by the red line in Figure 8a. Therefore, the two-layer GCN model (i.e., CDT) fails to capture this valuable word and misidentifies the sentiment as neutral. By contrast, our model is capable of establishing a direct syntactic relation of the sentiment word toward the aspect, as indicated by the red line in Figure 8b. In the third sample, more importance is attached to the noise word “tiny” instead of the opinion word “fast”, which causes the misclassification. In the proposed aspect-wise dependency tree, both “tiny” and “fast” relate to the aspect word “service”. The exploiting of DRA attention assigns more attention weights to the word “fast”, based on which the sentiment is classified as positive. Lastly, no explicit sentiment word is provided in the last sentence. Under this condition, both syntactic and semantic information has to taken into account. It is clear that our model identifies the positive polarity to the aspect with the enhancing of sentence semantics.

4.10. Analysis of the Contrastive Coefficient

The contrastive coefficients in SGCL are studied to clarify its significance. The working performance on ALSC with and without contrastive coefficient is presented in Table 7. One can easily see that the proposed model overperforms the one using simply contrastive learning without contrastive coefficient to a large extent. Accordingly, the contrastive coefficient plays a pivot role in learning the word syntactic structure of ALSC tasks.
The way of enhancing sentiment feature to the aspect using SGCL is presented as well. We visualize the obtained aspect vectors generated on the Restaurant test set using t-SNE [44] for SSEDGCN (w/o CL&SGCL), SSEDGCN (w/ CL), and SSSEGCN (w/ SGCL), respectively. The sentiment classification results of three categories of methods are shown in Figure 9. The dots in red, green, and blue correspond to the sentiment polarities of positive, negative, and neutral, respectively. Without contrastive learning, a certain proportion of the aspect nodes of different sentiment polarities overlap with each other, as presented in Figure 9a. With the application of basic contrastive learning scheme, the aspect vectors are partially distinguished, as presented in Figure 9b. Furthermore, it can be observed that the separations of representations derived from SGCL are significantly clearer than that generated by basic contrastive learning scheme among different sentiment polarities. The experimental result verifies not just the importance of contrastive coefficient in SGCL, but also the superiority of SGCL in learning aspect representations.

5. Conclusions

In this work, we propose a Syntactic Structure-Enhanced Dual Graph Convolutional Network which focuses on effectively encoding comprehensive syntax information for ALSC. To start with, a novel aspect-wise dependency tree is proposed by reconstructing the basic syntax dependency tree. Then, the sentence syntax and semantics are encoded via two specific GCNs: a self-attention mechanism is adopted by SemGCN and a DRA-attention mechanism is adopted by SynGCN. Moreover, to fully consider the syntactic structural information, the syntax-guided contrastive learning is designed to learn the feature representation of the aspect. The experimental results on three benchmark datasets demonstrate the effectiveness of our proposed method, achieving state-of-the-art performance in terms of accuracy and macro-F1. Moreover, the ablation study and visual analysis have validated the role of the new tree structure and each component in our proposed model.

Author Contributions

Conceptualization, J.C. and Y.X.; methodology, J.C.; formal analysis, J.C. and Z.Q.; writing—original draft preparation, J.C. and Z.Q.; writing—review and editing, J.C. and J.L.; supervision, Q.C. and Y.X.; funding acquisition Q.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Guangdong Basic and Applied Basic Research Foundation under Grant 2023A1515011370, the Characteristic Innovation Projects of Guangdong Colleges and Universities (No. 2018KTSCX049), and the Science and Technology Plan Project of Guangzhou under Grant No. 202102080258.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pontiki, M.; Galanis, D.; Papageorgiou, H.; Androutsopoulos, I.; Manandhar, S.; Al-Smadi, M.; Al-Ayyoub, M.; Zhao, Y.; Qin, B.; De Clercq, O.; et al. Semeval-2016 task 5: Aspect based sentiment analysis. In Proceedings of the International Workshop on Semantic Evaluation, San Diego, CA, USA, 16–17 June 2016; pp. 19–30. [Google Scholar]
  2. Wang, Y.; Huang, M.; Zhu, X.; Zhao, L. Attention-based LSTM for aspect-level sentiment classification. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, Austin, TX, USA, 1–5 November 2016; pp. 606–615. [Google Scholar]
  3. Tang, D.; Qin, B.; Liu, T. Aspect Level Sentiment Classification with Deep Memory Network. In Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, Austin, TX, USA, 1–5 November 2016; pp. 214–224. [Google Scholar]
  4. Ma, D.; Li, S.; Zhang, X.; Wang, H. Interactive attention networks for aspect-level sentiment classification. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, Melbourne, Australia, 19–25 August 2017; pp. 4068–4074. [Google Scholar]
  5. Chen, P.; Sun, Z.; Bing, L.; Yang, W. Recurrent attention network on memory for aspect sentiment analysis. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, Denmark, 7–11 September 2017; pp. 452–461. [Google Scholar]
  6. Fan, F.; Feng, Y.; Zhao, D. Multi-grained attention network for aspect-level sentiment classification. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, 31 October 2018–4 November 2018; pp. 3433–3442. [Google Scholar]
  7. Huang, B.; Ou, Y.; Carley, K.M. Aspect level sentiment classification with attention-over-attention neural networks. In Proceedings of the International Conference on Social Computing, Behavioral-Cultural Modeling and Prediction and Behavior Representation in Modeling and Simulation, Washington DC, USA, 10–13 July 2018; pp. 197–206. [Google Scholar]
  8. Gu, S.; Zhang, L.; Hou, Y.; Song, Y. A position-aware bidirectional attention network for aspect-level sentiment analysis. In Proceedings of the 27th International Conference on Computational Linguistics, Santa Fe, NM, USA, 20–26 August 2018; pp. 774–784. [Google Scholar]
  9. Song, Y.; Wang, J.; Jiang, T.; Liu, Z.; Rao, Y. Attentional encoder network for targeted sentiment classification. arXiv 2019, arXiv:1902.09314. [Google Scholar]
  10. Zhang, C.; Li, Q.; Song, D. Aspect-based sentiment classification with aspect-specific graph convolutional networks. arXiv 2019, arXiv:1909.03477. [Google Scholar]
  11. Sun, K.; Zhang, R.; Mensah, S.; Mao, Y.; Liu, X. Aspect-level sentiment analysis via convolution over dependency tree. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; pp. 5679–5688. [Google Scholar]
  12. Huang, B.; Carley, K.M. Syntax-Aware Aspect Level Sentiment Classification with Graph Attention Networks. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, China, 3–7 November 2019; pp. 5469–5477. [Google Scholar]
  13. Zhang, M.; Qian, T. Convolution over hierarchical syntactic and lexical graphs for aspect level sentiment analysis. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Punta Cana, Dominican Republic, 8–12 November 2020; pp. 3540–3549. [Google Scholar]
  14. Chen, C.; Teng, Z.; Zhang, Y. Inducing target-specific latent structures for aspect sentiment classification. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Punta Cana, Dominican Republic, 8–12 November 2020; pp. 5596–5607. [Google Scholar]
  15. Liang, B.; Yin, R.; Gui, L.; Du, J.; Xu, R. Jointly learning aspect-focused and inter-aspect relations with graph convolutional networks for aspect sentiment analysis. In Proceedings of the 28th International Conference on Computational Linguistics, Barcelona, Spain, Online, 8–13 December 2020; pp. 150–161. [Google Scholar]
  16. Wang, K.; Shen, W.; Yang, Y.; Quan, X.; Wang, R. Relational Graph Attention Network for Aspect-based Sentiment Analysis. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 3229–3238. [Google Scholar]
  17. Tang, H.; Ji, D.; Li, C.; Zhou, Q. Dependency graph enhanced dual-transformer structure for aspect-based sentiment classification. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; pp. 6578–6588. [Google Scholar]
  18. Li, R.; Chen, H.; Feng, F.; Ma, Z.; Wang, X.; Hovy, E. Dual graph convolutional networks for aspect-based sentiment analysis. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), Bangkok, Thailand, 1–6 August 2021; pp. 6319–6329. [Google Scholar]
  19. Pang, S.; Xue, Y.; Yan, Z.; Huang, W.; Feng, J. Dynamic and multi-channel graph convolutional networks for aspect-based sentiment analysis. In Proceedings of the Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, Online, 1–6 August 2021; pp. 2627–2636. [Google Scholar]
  20. Zhang, Z.; Zhou, Z.; Wang, Y. SSEGCN: Syntactic and semantic enhanced graph convolutional network for aspect-based sentiment analysis. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Seattle, WA, USA, 10–15 July 2022; pp. 4916–4925. [Google Scholar]
  21. Gou, J.; Yu, B.; Maybank, S.J.; Tao, D. Knowledge distillation: A survey. Int. J. Comput. Vis. 2021, 129, 1789–1819. [Google Scholar] [CrossRef]
  22. Zhou, T.; Shen, Y.; Chen, K.; Cao, Q. Hierarchical dual graph convolutional network for aspect-based sentiment analysis. Knowl.-Based Syst. 2023, 276, 110740. [Google Scholar] [CrossRef]
  23. Zhang, S.; Gong, H.; She, L. An aspect sentiment classification model for graph attention networks incorporating syntactic, semantic, and knowledge. Knowl.-Based Syst. 2023, 275, 110662. [Google Scholar] [CrossRef]
  24. Micali, S.; Zhu, Z.A. Reconstructing markov processes from independent and anonymous experiments. Discret. Appl. Math. 2016, 200, 108–122. [Google Scholar] [CrossRef]
  25. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. In Proceedings of the International Conference on Learning Representations, Palais des Congrès Neptune, Toulon, France, 24–26 April 2017. [Google Scholar]
  26. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  27. Xu, K.; Hu, W.; Leskovec, J.; Jegelka, S. How Powerful are Graph Neural Networks? In Proceedings of the International Conference on Learning Representations, New Orleans, Louisiana, MI, USA, 6–9 May 2019. [Google Scholar]
  28. You, Y.; Chen, T.; Sui, Y.; Chen, T.; Wang, Z.; Shen, Y. Graph contrastive learning with augmentations. Adv. Neural Inf. Process. Syst. 2020, 33, 5812–5823. [Google Scholar]
  29. Wei, C.; Liang, J.; Liu, D.; Wang, F. Contrastive Graph Structure Learning via Information Bottleneck for Recommendation. Adv. Neural Inf. Process. Syst. 2022, 35, 20407–20420. [Google Scholar]
  30. Xia, L.; Huang, C.; Huang, C.; Lin, K.; Yu, T.; Kao, B. Automated Self-Supervised Learning for Recommendation. In Proceedings of the ACM Web Conference, New York, NY, USA, 30 April–4 May 2023; pp. 992–1002. [Google Scholar]
  31. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, PMLR, Vienna, Austria, 12–18 July 2020; pp. 1597–1607. [Google Scholar]
  32. Khosla, P.; Teterwak, P.; Wang, C.; Sarna, A.; Tian, Y.; Isola, P.; Maschinot, A.; Liu, C.; Krishnan, D. Supervised contrastive learning. Adv. Neural Inf. Process. Syst. 2020, 33, 18661–18673. [Google Scholar]
  33. Chen, T.; Kornblith, S.; Swersky, K.; Norouzi, M.; Hinton, G.E. Big self-supervised models are strong semi-supervised learners. Adv. Neural Inf. Process. Syst. 2020, 33, 22243–22255. [Google Scholar]
  34. Gao, T.; Yao, X.; Chen, D. SimCSE: Simple Contrastive Learning of Sentence Embeddings. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, EMNLP 2021. Association for Computational Linguistics (ACL), Punta Cana, Dominican Republic, Online, 7–11 November 2021; pp. 6894–6910. [Google Scholar]
  35. Wang, R.; Dai, X. Contrastive Learning-Enhanced Nearest Neighbor Mechanism for Multi-Label Text Classification. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), Dublin, Ireland, 22–27 May 2022; pp. 672–679. [Google Scholar]
  36. Liang, B.; Luo, W.; Li, X.; Gui, L.; Yang, M.; Yu, X.; Xu, R. Enhancing aspect-based sentiment analysis with supervised contrastive learning. In Proceedings of the 30th ACM International Conference on Information & Knowledge Management, Queensland, Australia, 1–5 November 2021; pp. 3242–3247. [Google Scholar]
  37. Tian, Y.; Chen, G.; Song, Y. Aspect-based sentiment analysis with type-aware graph convolutional networks and layer ensemble. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Online, 6–11 June 2021; pp. 2910–2922. [Google Scholar]
  38. Devlin, J.; Chang, M.; Lee, K.; Toutanova, K. Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
  39. Ivanov, S.; Burnaev, E. Anonymous walk embeddings. In Proceedings of the International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 2186–2195. [Google Scholar]
  40. Pontiki, M.; Galanis, D.; Pavlopoulos, J.; Papageorgiou, H.; Androutsopoulos, I.; Manandhar, S. SemEval-2014 Task 4: Aspect based sentiment analysis. In Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval 2014), Dublin, Ireland, 23–24 August 2014. [Google Scholar]
  41. Dong, L.; Wei, F.; Tan, C.; Tang, D.; Zhou, M.; Xu, K. Adaptive recursive neural network for target-dependent twitter sentiment classification. In Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics (volume 2: Short Papers), Baltimore, MD, USA, 22–27 June 2014; pp. 49–54. [Google Scholar]
  42. Xiong, H.; Yan, Z.; Zhao, H.; Huang, Z.; Xue, Y. Triplet Contrastive Learning for Aspect Level Sentiment Classification. Mathematics 2022, 10, 4099. [Google Scholar] [CrossRef]
  43. Zhang, K.; Zhang, K.; Zhang, M.; Zhao, H.; Liu, Q.; Wu, W.; Chen, E. Incorporating Dynamic Semantics into Pre-Trained Language Model for Aspect-based Sentiment Analysis. In Proceedings of the Findings of the Association for Computational Linguistics: ACL 2022, Dublin, Ireland, 22–27 May 2022; pp. 3599–3610. [Google Scholar]
  44. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. An example sentence with its dependency tree where aspects are highlighted in green and opinion words are highlighted in blue.
Figure 1. An example sentence with its dependency tree where aspects are highlighted in green and opinion words are highlighted in blue.
Mathematics 11 03877 g001
Figure 2. An example of anonymous walks.
Figure 2. An example of anonymous walks.
Mathematics 11 03877 g002
Figure 3. The overall architecture of SSEDGCN.
Figure 3. The overall architecture of SSEDGCN.
Mathematics 11 03877 g003
Figure 4. Aspect-wise dependency tree.
Figure 4. Aspect-wise dependency tree.
Mathematics 11 03877 g004
Figure 5. The local syntactic structure distribution.
Figure 5. The local syntactic structure distribution.
Mathematics 11 03877 g005
Figure 6. Attention visualization for DRA Attention (DRA-Att) and Classic Attention (Classic-Att) for the aspect word ‘web browsing’.
Figure 6. Attention visualization for DRA Attention (DRA-Att) and Classic Attention (Classic-Att) for the aspect word ‘web browsing’.
Mathematics 11 03877 g006
Figure 7. Effect of the number of GCN layers (a) Accuracy performance based on different numbers of GCN layers. (b) F1-score performance based on different numbers of GCN layers.
Figure 7. Effect of the number of GCN layers (a) Accuracy performance based on different numbers of GCN layers. (b) F1-score performance based on different numbers of GCN layers.
Mathematics 11 03877 g007
Figure 8. Visualization of dependency trees in the case study. (a) Classical dependency tree of the second sentence. (b) Aspect-wise dependency tree of the second sentence.
Figure 8. Visualization of dependency trees in the case study. (a) Classical dependency tree of the second sentence. (b) Aspect-wise dependency tree of the second sentence.
Mathematics 11 03877 g008
Figure 9. Visualization of sentiment classification results. SSEDGCN (w/o CL& SGCL) refers to the model without contrast learning and syntactic guidance, SSEDGCN (w/CL) refers to the use of contrastive learning, and SSEDGCN is (w/SGCL) is the model with SGCL.
Figure 9. Visualization of sentiment classification results. SSEDGCN (w/o CL& SGCL) refers to the model without contrast learning and syntactic guidance, SSEDGCN (w/CL) refers to the use of contrastive learning, and SSEDGCN is (w/SGCL) is the model with SGCL.
Mathematics 11 03877 g009
Table 1. Statistics of the experimental datasets.
Table 1. Statistics of the experimental datasets.
DatasetPositiveNegativeNeutral
TrainTestTrainTestTrainTest
Laptop976337851128455167
Restaurant2164727807196637196
Twitter150717215281693016336
Table 2. Experimental results comparison on three publicly available datasets.
Table 2. Experimental results comparison on three publicly available datasets.
ModelsRestaurantLaptopTwiitter
AccuracyF1AccuracyF1AccuracyF1
CDT+BERT [11]86.2480.6680.0676.1777.1075.90
BERT-SPC [9]84.4676.9878.9975.0373.5572.14
R-GAT+BERT [16]86.6081.3578.2174.0776.1574.88
DGEDT+BERT [17]86.3081.3579.8075.6077.9075.40
T-GCN+BERT [37]86.1679.9580.8877.9676.4575.25
DualCGN+BERT [18]87.1381.1681.8078.1077.4076.02
DMCGN+BERT [19]87.6682.7980.2278.1078.0677.06
TCL+BERT [42]87.4082.1281.8078.9677.5576.57
SSEGCN+BERT [20]87.3181.0981.0177.9677.4076.02
DR-BERT [43]87.7282.3181.4578.1677.2476.10
SSK-GAT+BERT [23]87.4181.6580.2575.8575.7274.44
HD-GCN+BERT [22]87.1381.4081.8078.8877.3476.12
SSEDGCN(Ours)87.8582.0882.9180.1278.1477.08
Table 3. Ablation study results.
Table 3. Ablation study results.
ModelsRestaurantLaptopTwitter
AccuracyF1AccuracyF1AccuracyF1
SSEDGCN87.8582.0882.9180.1278.1477.08
w/o SGCL87.0281.4681.7378.6177.2476.12
w/o SemGCN86.5680.5081.2477.2377.1076.34
w/o DRA attention86.3379.9281.6377.7776.6975.56
w/o SynGCN85.8879.2580.9477.4576.5475.42
Table 4. Results of SSEDGCN based on two different dependency trees.
Table 4. Results of SSEDGCN based on two different dependency trees.
Dependency TreeModelsRestaurantLaptopTwitter
AccuracyF1AccuracyF1AccuracyF1
Aspect-wiseSSEDGCN87.8582.0882.9180.1278.1477.08
ClassicSSEDGCN86.0679.7881.4878.1477.5976.13
Table 5. Impact of the training/testing data ratio on accuracy.
Table 5. Impact of the training/testing data ratio on accuracy.
Train:TestAccuracy (%)
3.22:187.85
2.58:184.72
1.93:184.53
1.29:173.02
0.64:164.96
Table 6. Case study. ALSC results of CDT and SSEDGCN on testing examples, along with their predictions. The marker ✔ and ✗ indicate the correct classification and incorrect classification, respectively.
Table 6. Case study. ALSC results of CDT and SSEDGCN on testing examples, along with their predictions. The marker ✔ and ✗ indicate the correct classification and incorrect classification, respectively.
#Testing InstanceCDTOurs
1I definitely enjoyed the food as wellpos ✔pos ✔
2Most of the sandwiches are made with soy mayonaise which is actually pretty goodneu ✗pos ✔
3tiny restaurant with very fast serviceneg ✗pos ✔
4and the food, well the food will keep you coming backneg ✗pos ✔
Table 7. Experimental results with and without the contrastive coefficient.
Table 7. Experimental results with and without the contrastive coefficient.
ModelsRestaurantLaptopTwitter
AccuracyF1AccuracyF1AccuracyF1
SSEDGCN (w/ ω )87.8582.0882.9280.1278.1477.08
SSEDGCN (w/o ω )87.1381.5482.0778.7677.3476.89
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, J.; Qiu, Z.; Liu, J.; Xue, Y.; Cai, Q. Syntactic Structure-Enhanced Dual Graph Convolutional Network for Aspect-Level Sentiment Classification. Mathematics 2023, 11, 3877. https://doi.org/10.3390/math11183877

AMA Style

Chen J, Qiu Z, Liu J, Xue Y, Cai Q. Syntactic Structure-Enhanced Dual Graph Convolutional Network for Aspect-Level Sentiment Classification. Mathematics. 2023; 11(18):3877. https://doi.org/10.3390/math11183877

Chicago/Turabian Style

Chen, Jiehai, Zhixun Qiu, Junxi Liu, Yun Xue, and Qianhua Cai. 2023. "Syntactic Structure-Enhanced Dual Graph Convolutional Network for Aspect-Level Sentiment Classification" Mathematics 11, no. 18: 3877. https://doi.org/10.3390/math11183877

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop