Aspect-Level Sentiment Analysis of Bi-Graph Convolutional Networks Based on Enhanced Syntactic Structural Information

Abstract

Aspect-oriented sentiment analysis is a meticulous sentiment analysis task that aims to analyse the sentiment polarity of specific aspects. Most of the current research builds graph convolutional networks based on dependent syntactic trees, which improves the classification performance of the models to some extent. However, the technical limitations of dependent syntactic trees can introduce considerable noise into the model. Meanwhile, it is difficult for a single graph convolutional network to aggregate both semantic and syntactic structural information of nodes, which affects the final sentence classification. To cope with the above problems, this paper proposes a bi-channel graph convolutional network model. The model introduces a phrase structure tree and transforms it into a hierarchical phrase matrix. The adjacency matrix of the dependent syntactic tree and the hierarchical phrase matrix are combined as the initial matrix of the graph convolutional network to enhance the syntactic information. The semantic information feature representations of the sentences are obtained by the graph convolutional network with a multi-head attention mechanism and fused to achieve complementary learning of dual-channel features. Experimental results show that the model performs well and improves the accuracy of sentiment classification on three public benchmark datasets, namely Rest14, Lap14 and Twitter.

Share and Cite:

Hu, J. and Li, Y. (2025) Aspect-Level Sentiment Analysis of Bi-Graph Convolutional Networks Based on Enhanced Syntactic Structural Information. Journal of Computer and Communications, 13, 72-89. doi: 10.4236/jcc.2025.131006.

1. Introduction

Aspect-based Sentiment analysis (ABSA) is a fine-grained field of emotion analysis. Its main task is to identify and analyze the emotional polarity expressed in a given text, such as positive, negative or neutral, for a specific aspect or entity. This kind of analysis method can let us understand the feelings expressed in the text more deeply, especially when dealing with complex text containing multiple aspects or entities, such as product reviews, restaurant reviews, etc. Taking “the waiter was attentive, the food is unpleasant” as an example, the sentence contains two aspects of “waiter” and “food”, and the corresponding emotional polarity is positive and negative respectively. The goal of the ABSA task is to establish the relationship between the aspect item and the opinion item, so as to judge that the emotional polarity of the two aspects is positive and negative respectively.

Aspect level emotion analysis methods can be divided into two types: machine learning based methods and deep learning based methods [1]. The method based on machine learning relies on rich and correct feature representation and an emotional dictionary that needs to be updated at any time, which is more expensive and difficult. With the rise of deep learning, attention mechanisms have become widely used in aspect-level emotion analysis tasks [2]-[4]. Although the attention mechanism is widely used in aspect level tasks and has achieved some results, it cannot obtain the syntactic relationship between specific aspect words and contexts. Many scholars have applied Graph Convolutional Networks (GCN) [5] to aspect level emotion analysis tasks to solve this problem. GCN can effectively process the dependency tree or syntactic structure of a sentence, where nodes represent words and edges represent the grammatical relationship between words. By applying convolution operations on such a graph structure, GCN can capture the dependencies between words and help the model better understand the semantic content of sentences. Although GCN can obtain syntactic information in sentences, it is insufficient to obtain information, and it also brings a lot of noise. External emotional knowledge is also needed to enhance the syntactic features of sentences. In existing research, emotional features are usually enhanced by establishing an emotional knowledge map, introducing emotional dictionary knowledge and introducing common sense knowledge into external emotional knowledge, and many scholars have proved that emotional knowledge plays a positive role in emotional classification. And has a positive impact on emotional classification. How to effectively use deep semantic information and refined syntactic structure information to effectively fuse them to enhance graph neural networks and extract emotional features more fully has become a problem worth exploring.

Therefore, this paper proposes a two-channel graph convolutional network model that introduces emotional knowledge. The main contributions are as follows:

(1) Strengthen the relevance of syntactic information, associate syntactic dependency matrix and phrase matrix, and combine graph convolution neural network to extract more accurate feature information.

(2) The syntactic structure information and semantic information are dynamically updated through the dual channel graph convolution network. In order to fully integrate the syntactic information and semantic information, the importance of different representations is adaptively learned using the trainable parameter matrix.

(3) The performance of the model is evaluated on three public data sets. The experimental results show that the accuracy and macro F1 value of the model are improved for different data sets.

2. Related Work

In recent years, more and more deep learning methods have been applied to aspect level emotion analysis tasks, which can be roughly divided into two categories in existing work. The first type is to obtain the semantic information of sentences through the attention mechanism. Tang et al. [6] proposed a long-term and short-term memory network model based on goal dependency. Long Short Term Memory (LSTM) is used to obtain the feature information of target words and contexts, and effectively identify the corresponding emotional polarity. ATAT-LSTM [7] is an LSTM model based on attention. It uses the context and aspect word information in sentences to splice and model through the attention mechanism. Ma et al. [2] proposed that the Interactive Attention Network (IAN) can effectively model the relationship between the target word composed of multiple words and the context, focusing on the parts with higher relevance to the target word. Xu et al. proposed the Multi Attention Network (MAN), which uses the global attention module and the local attention module to capture the global semantic information and the semantic information between the context of the aspect words respectively. The combination of the two improves the accuracy of the model. However, such attention mechanism based methods only focus on the characteristics of aspect words and context in sentences, and cannot capture the syntactic dependency between context words in sentences.

The second type is to obtain syntactic structure information through dependency resolution. The syntactic structure information can obtain words that are dependent on aspect words, which greatly shortens the distance between aspect words and opinion words. Modeling through GCN can make up for the deficiency of attention mechanisms in aspect-level emotion analysis tasks. CDT [8] uses Bi LSTM to learn the feature representation of sentences, and GCN convolution based on syntactic dependency tree to learn the feature representation of sentences. TD-GAT [9] directly conveys emotional features from the context of the syntactic structure of aspect words, making the opinion words closer to aspect words, and resolving potential syntactic ambiguity, so that the model can obtain more accurate syntactic information. Most of the above GCN models based on dependency syntax trees ignore the emotional information between aspect words and contexts. In order to add emotional knowledge to aspect level emotion analysis tasks, Zhou et al. [10] proposed a graph convolution network model based on syntactic dependency and emotional knowledge modeling. This model is composed of two graph convolution neural networks, which construct syntactic dependency graphs and emotional knowledge graphs respectively and apply the multi head attention mechanism. Liang et al. [11] proposed a graph convolution network model for emotional knowledge enhancement, which complements a single syntactic dependency by introducing emotional knowledge in SenticNet, while constructing the adjacency matrix for emotional enhancement to extract the syntactic structure features of strong dependency.

Based on the above observations, this paper proposes to extract more accurate emotional information through the dual channel graph convolution neural network (BiC-GCN) to help find aspect words, opinion words and their emotional polarity.

3. Binary Channels Graph Convolution Model (BiC-GCN)

As shown in Figure 1, the BiC-GCN model proposed in this paper consists of six parts: embedding layer, Bi LSTM layer, feature extraction layer, graph convolution layer, aspect word masking layer, interaction layer, and classification layer.

Figure 1. BiC-GCN model architecture.

3.1. Embedded Layer

Assume that the sentence is consists of n words, Where denotes the ith word and denotes the ith aspect word, set of aspect words . Use pre pre-trained GloVe word vector [12] map each word in the input statement to a vector with low dimensional real value to obtain the embedding matrix , is the size of vocabulary, denotes the dimension of each word vector. Therefore, the sentence has corresponding word embedding .

3.2. Bi-LSTM Layer

LSTM can learn the information of sentences from front to back during training, but cannot encode the information from back to front. Therefore, Bi-LSTM is used to learn the hidden information in the context. Bi LSTM is a combination of forward and backward LSTM. The forward LSTM of sentence S is expressed as: , the backward LSTM of sentence S is expressed as , the hidden state vector finally generated by Bi LSTM is spliced with the vectors corresponding to and , where , using H as the initial input for the model.

3.3. Feature Extraction Layer

Build an enhanced emotional dependency matrix for feature extraction. In aspect level sentiment analysis tasks, phrase structure tree and dependency syntax tree are two syntactic structures used to capture the collocation relationships between words in a sentence. Essentially, it is to represent each word in a sentence as a node in a tree, with dependency relationships represented as directed edges. This article uses the Stanford NLP tool for syntactic analysis and constructs a dependency syntactic tree. The dependency syntactic tree can be interpreted as a graph with nodes, where nodes represent words in the sentence and edges represent syntactic dependency paths between words in the graph. The dependency tree G of any sentence can be expressed as the adjacency matrix D of , If node is connected to node j through a single dependent path in tree G, Then , otherwise , as shown in Formula (1):

(1)

In the ASGCN model, we believe that the dependency between words is mutual, so we can build an undirected graph, where .

The phrase structure tree is a grammatical structure that helps to understand the text structure and capture the phrase level relationship, and can reflect the hierarchical relationship between sentence phrases [6]. The construction of the phrase structure tree usually involves phrase structure analysis. The nodes of the phrase structure tree represent phrases, and the phrases at each level can contain all the words of the sentence. As shown in the Figure 2:

Figure 2. The phrase structure tree.

The method of this paper is to give a sentence and use the natural language processing tool Stanford Consistency Parser developed by Stanford University to generate the phrase structure tree. Different from the dependency syntax tree, the phrase structure tree shows the organizational structure between words in the sentence. This paper uses each node in the phrase structure tree to segment the sentence. Each node in the phrase structure tree can be understood as a phrase, E.g. sentence “The voice quality of iPhone is great but its battery sucks”, the L-layer structure will be analyzed in the structural analysis, the topmost layer is broken down into [“The voice quality of iPhone is great”, “but”, “its battery sucks”], according to the phrase structure tree, the phrase structure at the next level can be expressed as [“The voice quality of iPhone”, “is great”, “but”, “its battery sucks”], Up to the bottom, so you can get the phrase list of each layer , the phrase list of each layer can contain all the words of the sentence. Because the phrase structure has multiple layers, the phrase set of L layer can be obtained , and the set is transformed into a multi-level phrase matrix by formula (2):

(2)

where is the vector representation of phrases at each level, is the number of levels of phrase set, and p is the number of phrases at that level. If only syntactic dependency trees are used, the association between great and socks will lead to noise problems, which is not conducive to capturing the emotional context of each aspect, thus leading to the misjudgment of emotional polarity. Therefore, phrase structure is added to the traditional syntactic dependency matrix to make it more closely related to the context. Because the subtree in the phrase structure governs all words appearing in the syntactic dependency subtree [7]. So, the dependency tree has the same words in its corresponding phrase subtree. From the perspective of these subtrees, the association between the transformed syntactic dependency matrix and the phrase matrix at each level can be established. Therefore, this paper combines the syntactic dependency matrix and the phrase matrix, as shown in Formula (3):

(3)

The information of the two is integrated into a common representation, including the internal context information and syntax information. By deleting the edges that may cause noise, the remaining dependent edge positions are added, and the edge merging is completed by matching the dependency type or phrase structure type. As shown in Figure 3:

Figure 3. Dependency matrix and phrase matrix fusion.

refers to the phrase matrix at each level, while D refers to the syntactic dependency matrix generated by the syntactic dependency tree.

The adjacency matrix obtained through the fusion of phrase matrix and syntactic dependency matrix contains both phrase and sentence information. The phrase level grammatical information and sentence level grammatical information are fused to obtain the set of adjacency matrix of , which is used as the input of GCN. Since there are multiple adjacency matrices in the set, multiple outputs G are obtained through formula (4), and the final GCN output is obtained by averaging the output hidden vectors of each layer, where represents the feature output from the first node to the ith node. The syntactic feature information is enhanced after the fused adjacency matrix passes through the GCN layer.

(4)

where, refers to multiple GCN outputs after fusion, refers to the adjacency matrix of a fused layer, i and j represent the relationship between words, refers to the weight matrix, refers to the deviation vector, refers to the activation function, and h refers to the initial hidden vector input.

3.4. SemGCN

Inspired by the work of Huang [12] et al. and He [13] et al., semantic graph convolution is introduced to distinguish the importance of different contexts. Its process is shown in Figure 2. The semantic map convolution is mainly divided into two parts, one is the initialization part, and the other is the update part.

The initialization part is used as the initialization input, and self attention operation is carried out on the initialized input to construct weight matrices based on . Summing weight matrices and applying top-k function to retain important context words, we can get the initial adjacency matrix of semantic map convolution by this method, and extract the preliminary semantic information by graph convolution operation.

(5)

(6)

In formula (5), is the hidden state generated by the coding layer Bi-LSTM, and is the ith adjacency matrix, is a trainable weight parameter, is the dimension of attention, is the number of attention heads.

For the update part, in the initialization semantic map , the update formula is as follows:

(7)

(8)

(9)

(10)

(11)

The difference between the update part and the initialization part lies in the input. The input of the update part is obtained from the convolution training of the previous layer graph. The softmax function operation of the weight matrix can obtain the probability matrix that is most closely related to the maximum probability in the current semantic map. Finally, the top-k selection of multi-layer graph convolution is used to obtain deeper semantic information, and the final output is and .

3.5. Covering Layer of Aspect Words

In order to highlight the important features of aspect words, mask the output of layer graph convolution before information fusion, and reserve the vector of aspect words, that is, the vector corresponding to aspect words is set to 1, the vector of non-aspect words is set to 0, and the output after setting is and .

(12)

(13)

(14)

3.6. Interaction Layer

The syntactic structure information and semantic information of a sentence are inseparable. If a certain kind of information is calculated separately, the effect of the model will be poor. Therefore, the interaction layer is designed so that the model can extract the feature information of the sentence from different perspectives. Specifically, the text feature information obtained from the coding layer Bi-LSTM interacts with the enhanced dependency feature after masking and the aspect word feature after masking the aspect feature.

Similarity (15)

(16)

(17)

The interactive and are fused and learned by using the trainable parameter matrix, so that the appropriate weight coefficient can be found in the continuous iterative training, and the final representation of the text vector under this model can be obtained.

(18)

3.7. Classification Layer

Take the final output of the model as the input of the full connection layer, and determine the final emotional polarity through the softmax function.

(19)

is the final output of the fusion layer, and and are the weights and bias terms of the full connection layer. Gradient descent algorithm is used to train the model during training to achieve the final classification purpose.

3.8. Loss Function

(20)

In formula (20), S represents the number of training samples, C represents the number of categories, and represents the real label of the training set, represents the prediction label of the model, represents all parameters that can be trained, and represents the coefficient of regularization term.

4. Experiment and Analysis

4.1. Data Set

This paper conducted experiments on three publicly available fine-grained emotion analysis based datasets, including restaurant comment dataset Rest14 [2], laptop comment dataset Lap14, and Twitter tweet comment dataset [3]. Each dataset is composed of real comments, and each comment has its corresponding emotional polarity, including positive, neutral and negative. This paper adopts Sun et al.’s pretreatment method to process data sets, and the statistical results of each data set after processing are shown in Table 1.

Table 1. Statistics for the three experimental datasets.

data set

grouping

positive

neutral

negative

Rest14

Train

2164

637

807

Test

727

196

196

Laptop

Train

976

455

851

Test

337

167

128

Twitter

Train

1507

3016

1528

Test

172

336

169

4.2. Hyperparameter Setting

The BiC-GCN model was tested with the parameters used, Glove [7] word vector dimensions were used with 300 dimensions, the number of graph convolution layers was tested to be optimal at 2, and the optimiser used Adam, which performs well in many deep learning tasks as it is able to adaptively adjust the learning rate for each parameter. The specific hyperparameter settings are shown in Table 2:

Table 2. Model parameter settings.

Parameter

parameter value

Parameter

parameter value

Glove [19] Word vector dimension

300

optimizer

Adam

Learning rate

0.001

Batch size

16

Training Iterations

100

dropout

0.3

GCN Number of layers

2

L2 Regularization coefficient

0.00001

The model uses accuracy and F1 value as indicators to evaluate whether the model is effective.

4.3. Comparative Experiment

In order to verify the effectiveness of the model, it will be compared with a series of benchmark models on three data sets. The following is a brief introduction of each model:

LSTM [14]: LSTM processes sequential data through a unique gate mechanism that effectively captures and preserves long-term dependencies and contextual information in text. Its output is a hidden state vector that contains a comprehensive feature representation of the text, incorporating important information about specific aspect words and their context.

ASGCN: A model that combines Graph Convolutional Networks (GCN) and Attention Mechanisms has been developed to model the syntactic information of a sentence by constructing a graph structure based on dependency trees and learning the dependencies between words. ASGCN has been shown to be capable of more accurately identifying and understanding the sentiment tendency of a sentence, thus enabling effective sentiment classification.

DGCN [15]: The integration of Bi-LSTM networks, syntactic dependency trees and graph convolutional networks is intended to facilitate the extraction of contextual information and features from text at multiple levels for the purpose of efficient sentiment analysis and related NLP tasks.

TD-GAT [12]: Graph Attention Networks are utilised for the purpose of aggregating aspect-related feature information. By means of this process, the model is capable of focusing on key information that is closely related to a particular aspect, while simultaneously ignoring irrelevant contextual information. This enables more accurate understanding and analysis of emotional tendencies or aspect-specific information in a text.

BiGCN [16]: The model has been designed as a two-layer interactive graph convolutional network. It is capable of interacting and fusing information between two graphs (grammatical and lexical) to extract a deeper and more comprehensive representation of features.

MIGCN [17]: A multi-interaction attention layer is utilised to capture semantic and syntactic information in the text. The integration of information from disparate layers enables the model to comprehend the underlying structure and relationships in the text more effectively.

MSD-GCN [18]: This paper introduces a novel fusion of the cross-attention mechanism with graph convolutional neural networks. The cross-attention mechanism enables the model to process text by focusing on key parts that are relevant to a particular task or goal while ignoring other irrelevant information. This approach enables the model to comprehend text in a multi-granular manner, thereby capturing rich syntactic information.

SGAN [17]: A syntactic graph attention network is proposed with the objective of incorporating syntactic dependency knowledge into graph attention networks to facilitate more efficient processing of text data. When processing text, the network is able to identify and focus on keywords or phrases based on the syntactic structure of the sentence and associate them with specific aspect terms.

R-GAT [12]: The information concerning dependency labels in the syntactic dependency tree is introduced as features in order to construct the relationship graph. Consequently, R-GAT is capable of capturing complex dependencies between words in a text and modelling them effectively.

DGEDT [1]: A combination of Transformer and Graph Convolutional Networks (GCN) has been designed to learn semantic features and graph representations of sentences. The model processes textual and graphical information through different modules and combines them through an interaction mechanism. The Transformer module is used to learn semantic features of sentences. The Graph Convolutional Network (GCN) is employed to learn graph representation. The employment of bi-affine transformations facilitates the interaction of features, thereby ensuring the capture of richer textual information.

As demonstrated in Table 3, the Glove word embedding-based model demonstrates significantly superior outcomes in comparison to the LSTM baseline model on the same dataset when employing BiC-GCN. Furthermore, a comparison of the latest GCN model reveals that all models exhibit varying degrees of enhancement. Among them, BiGCN and MIGCN, which employ semantic syntax interaction, exhibit an enhancement in model accuracy by 3.46% and 2.21%, respectively, on average, when compared to ASGCN. However, these models are limited to shallow semantic syntax interaction, which results in a deficiency in context and aspectual word information, consequently leading to a decline in model accuracy. In contrast, the GCN-based models (MSD-GCN, R-GAT, DGEDT) exhibited an enhancement in model accuracy by an average of 1.35%, 0.64%, and 1.12%, respectively. R-GAT emerged as the most effective model among the baseline models. It incorporated the label information of the dependencies in the syntactic dependency tree as a feature into the model to construct the relational graphical attention network. However, it did not utilise syntactic parsing to obtain the complete dependency tree. The complete dependency tree obtained leads to ignoring part of the syntactic information, resulting in limited model enhancement.

Table 3. Experimental results.

Models

Res14

Laptop

Twitter

Acc

F1

Acc

F1

Acc

F1

LSTM

78.13

67.47

69.28

63.09

69.56

67.70

ASGCN

80.77

72.02

75.55

71.05

72.15

70.40

TD-GAT

81.32

71.72

75.63

70.64

72.70

70.45

BiGCN

81.97

73.48

74.59

71.84

74.16

73.35

MIGCN

82.32

74.31

76.59

72.44

73.31

72.12

MSD-GCN

82.67

75.44

77.59

74.18

74.52

73.68

R-GAT

83.55

75.99

78.02

74.00

75.36

74.15

DualGCN

87.13

81.16

81.80

78.10

77.40

76.02

BiC-GCN

87.16

81.24

81.70

78.45

78.27

77.29

4.4. Ablation Experiment

In order to demonstrate the significance of each module in DS-GCN, this paper conducts an experimental study on the model. The results of these experiments are presented in Table 4. The baseline model is represented by DS-GCN, with the semantic branch and syntactic branch modules removed, respectively.

The notations “w/o dsp” and “w/o sem” in Table 4 denote the removal of the semantic branching module and the removal of the syntactic branching module respectively. The notations “w/o dsp-ren” denote the removal of dependency information, with the retention of phrase structure information.

Table 4. Experimental results of ablation study.

Models

Res14

Laptop

Twitter

Acc

F1

Acc

F1

Acc

F1

w/o dsp

85.71

78.32

79.57

75.25

75.47

74.66

w/o sem

86.16

79.94

80.62

77.10

76.45

75.42

w/o dsp-ren

86.45

80.28

79.59

76.93

76.52

76.07

BiC-GCN

87.16

81.24

81.70

78.45

78.27

77.29

A series of ablation experiments were conducted to ascertain the impact of employing distinct modules for modelling. Utilising solely the semantic module resulted in a diminished capacity to acquire information and exhibited suboptimal generalisation capabilities, consequently leading to a decline in accuracy of 1.61%, 1.49%, and 1.30% for Res14, Laptop, and Twitter, respectively. This finding signifies that the structural information inherent within a sentence constitutes an indispensable component of the model. In contrast, when the syntactic module is used alone, there is a decline in accuracy of 1.83%, 1.80%, and 1.73% for Res14, Laptop, and Twitter, respectively. This suggests that incorporating semantic information enhances the classification accuracy of the model. However, when the model removes the sentiment node and retains only sentence dependency and semantic information, the accuracy of the model is observed to decrease on all three datasets. This finding indicates that the sentiment node plays a positive role in enhancing the syntactic information. Furthermore, the w/o dsp-ren reflects that the syntactic and semantic information can be complementary to each other, which further underscores the importance of integrating semantic information with syntactic information. In conclusion, it can be stated that the sentence representations obtained from different branches of the model all have an important impact on the model of this paper.

4.5. hyperparametric Analysis

4.5.1. GCN Layer Analysis

In order to verify the effect of the number of GCN layers on the model, the number of GCN layers was set to 1 - 5, respectively. Following experimentation on three datasets, changes in accuracy and F1 value were analysed for different layers. The experimental results are shown in Figure 4 and Figure 5:

As demonstrated in Figure 1, the model achieves optimal performance when the number of GCN layers is set at 2. However, as the number of GCN layers increases, the model’s overall performance exhibits a decline. This phenomenon can be attributed to the fact that, with an augmented number of GCN layers, the model’s parameters also rise, consequently hindering its capacity to learn effectively, leading to convergence difficulties or even the disappearance of the gradient. This, in turn, results in a decline in model accuracy.

Figure 4. Effect of number of GCN layers on model accuracy.

Figure 5. Effect of number of GCN layers on F1 values.

4.5.2. Top_k Value Analysis

In the semantic analysis module, the top_k function is utilised to identify and retain significant contextual words, whilst also conducting a comprehensive analysis of the impact of these words on the model’s final outcomes. The experimental results are illustrated in Figure 6 and Figure 7:

Figure 6. Effect of the value of top_k on accuracy.

Figure 7. Effect of the value of top_k on the value of F1.

It is evident that the number of keywords selected by the model with the best results, top_k, varies for different datasets, primarily due to the varying number of aspect words across different datasets of differing data lengths. The model’s overall performance is optimised when 1 - 4 high-scoring words are selected. However, as the number of selected keywords increases, the introduction of irrelevant words becomes more prevalent, thereby introducing noise that interferes with the model’s ability to accurately discern sentiment polarity.

4.5.3. Case Studies

A case study was conducted, with several examples categorised using different models. The results are shown in Table 5. P, N and O are used to denote positive, negative and neutral, respectively, and Label represents the true affective polarity of the aspect word.

Table 5. Case study.

example sentence

Label

R-GAT

DualGCN

our

(a) The menu is limited but the dishes are excellent.

(N, P)

(N, P)

(N, P)

(N, P)

(b) The menu is limited but the dishes are excellent.

(P, N)

(P, P)

(P, N)

(P, N)

(c) I hand great interest in this restaurant due to its atmosphere, but the service was disappointing.

(P, N)

(N, N)

(O, N)

(P, N)

Despite the close proximity of both “speed” and “battery life” to the opinion word “better”, an analysis of the data reveals a slight disparity between the two. Specifically, “battery life” exhibits a marginally greater distance from the opinion word “speed” compared to “battery life”. Despite the proximity of both “speed” and “battery life” to the opinion word “better”, the analysis demonstrates that “battery life” is marginally more distant from the opinion word “speed”. This finding underscores the limitations of the R-GAT model in accurately extracting dependency information, consequently leading to prediction errors.

In the third sentence, the opinion expression corresponding to the “atmosphere” aspect is “great interest”, which is actually the same as the opinion expression “appointing”. “Appointing” is closer to the expression of opinion. Consequently, the R-GAT and DualGCN models demonstrate a tendency towards bias in their predictions. However, the model proposed in this paper is able to more accurately capture the emotional cues in real opinion expressions by introducing emotional knowledge, which improves the accuracy of prediction. The enhanced ability to manage intricate emotional interrelationships ensures a more precise evaluation of products or services.

5. Conclusions

In this paper, we propose a two-channel graph convolution neural network model based on increasing sentiment knowledge for aspect-level sentiment analysis tasks. Initially, the syntactic dependency graph, augmented with sentiment knowledge, undergoes a graph convolution operation. Concurrently, the semantic features of the entire text are obtained through the utilisation of a multi-head attention mechanism. Thereafter, an adaptive learning process occurs, whereby the semantic and syntactic features are integrated. The efficacy of the model is validated through experiments on three datasets, which demonstrate that the model outperforms the benchmark model in terms of accuracy and F1 values. Furthermore, the individual modules of the model are demonstrated to contribute to improved classification results through a series of ablation experiments.

The proposed BiC-GCN model in this paper has achieved positive research results in fine-grained sentiment analysis tasks, but there are still some limitations and directions worth further exploring: firstly, the experimental validation of this paper has been carried out in Res14, Laptop, and Twitter, and the dataset has limited domain diversity. In the future, we can consider expanding the domain coverage of the dataset to further validate the generalisation and robustness of the proposed method. Secondly, issues such as emoji usage and character expression in the task of fine-grained sentiment analysis of comment text increase the difficulty of the task, and the accuracy needs to be improved. However, these limitations also point to the direction of future research in this study. Future research can further improve the accuracy of fine-grained sentiment analysis by increasing the dataset domain coverage and enhancing emoji and symbol sentiment representation.

Conflicts of Interest

The authors declare no conflicts of interest regarding the publication of this paper.

References

[1] Phan, H.T., Nguyen, N.T. and Hwang, D. (2023) Aspect-Level Sentiment Analysis: A Survey of Graph Convolutional Network Methods. Information Fusion, 91, 149-172.
https://doi.org/10.1016/j.inffus.2022.10.004
[2] Ma, D., Li, S., Zhang, X. and Wang, H. (2017) Interactive Attention Networks for Aspect-Level Sentiment Classification. Proceedings of the Twenty-Sixth International Joint Conference on Artificial Intelligence, Melbourne, 19-25 August 2017, 4068-4074.
https://doi.org/10.24963/ijcai.2017/568
[3] Chen, P., Sun, Z., Bing, L. and Yang, W. (2017) Recurrent Attention Network on Memory for Aspect Sentiment Analysis. Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Copenhagen, 7-11 September 2017, 452-461.
https://doi.org/10.18653/v1/d17-1047
[4] Fan, F., Feng, Y. and Zhao, D. (2018) Multi-Grained Attention Network for Aspect-Level Sentiment Classification. Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, 31 October-4 November 2018, 3433-3442.
https://doi.org/10.18653/v1/d18-1380
[5] Kipf, T.N. and Welling, M. (2017) Semi-Supervised Classification with Graph Convolutional Networks. The 5th International Conference on Learning Representations, Toulon, 24-26 April 2017, 1-14.
[6] Tang, D., Qin, B., Feng, X., et al. (2016) Effective LSTMs for Target-Dependent Sentiment Classification. Proceedings of the 26th International Conference on Computational Linguistics: Technical Papers, Osaka, 11-17 December 2016, 3298-3307.
[7] Wang, Y., Huang, M., Zhu, X. and Zhao, L. (2016) Attention-Based LSTM for Aspect-Level Sentiment Classification. Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing, Austin, 1-5 November 2016, 606-615.
https://doi.org/10.18653/v1/d16-1058
[8] Sun, K., Zhang, R., Mensah, S., Mao, Y. and Liu, X. (2019) Aspect-Level Sentiment Analysis via Convolution over Dependency Tree. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, 3-7 November 2019, 5679-5688.
https://doi.org/10.18653/v1/d19-1569
[9] He, R., Lee, W., Ng, H., et al. (2018) Effective Attention Modeling for Aspect-Level Sentiment Classification. Proceedings of the 27th International Conference on Computational Linguistics, Stroudsburg, 20-26 August 2018, 1121-1131.
[10] Zhou, J., Huang, J.X., Hu, Q.V. and He, L. (2020) SK-GCN: Modeling Syntax and Knowledge via Graph Convolutional Network for Aspect-Level Sentiment Classification. Knowledge-Based Systems, 205, Article 106292.
https://doi.org/10.1016/j.knosys.2020.106292
[11] Liang, B., Su, H., Gui, L., Cambria, E. and Xu, R. (2022) Aspect-Based Sentiment Analysis via Affective Knowledge Enhanced Graph Convolutional Networks. Knowledge-Based Systems, 235, Article 107643.
https://doi.org/10.1016/j.knosys.2021.107643
[12] Huang, B. and Carley, K. (2019) Syntax-Aware Aspect Level Sentiment Classification with Graph Attention Networks. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, 3-7 November 2019, 5469-5477.
https://doi.org/10.18653/v1/d19-1549
[13] Zhang, C., Li, Q. and Song, D. (2019) Aspect-Based Sentiment Classification with Aspect-Specific Graph Convolutional Networks. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Hong Kong, 3-7 November 2019, 4568-4578.
https://doi.org/10.18653/v1/d19-1464
[14] Zou, L., Xia, L., Ding, Z., Song, J., Liu, W. and Yin, D. (2019) Reinforcement Learning to Optimize Long-Term User Engagement in Recommender Systems. Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, 4-8 August 2019, 2810-2818.
https://doi.org/10.1145/3292500.3330668
[15] Wang, K., Shen, W., Yang, Y., Quan, X. and Wang, R. (2020) Relational Graph Attention Network for Aspect-Based Sentiment Analysis. Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5-10 July 2020, 3229-3238.
https://doi.org/10.18653/v1/2020.acl-main.295
[16] Zhang, M. and Qian, T. (2020) Convolution over Hierarchical Syntactic and Lexical Graphs for Aspect Level Sentiment Analysis. Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Online, 5-10 July 2020, 3540-3549.
https://doi.org/10.18653/v1/2020.emnlp-main.286
[17] Wang, R., Tao, Z., Zhao, R., et al. (2022) Multi-Interaction Graph Convolutional Networks for Aspectual Sentiment Analysis. Journal of Electronics and Information, 44, 1111-1118.
[18] Zhao, Z., Liu, Y., Gao, J., Wu, H., Yue, Z. and Li, J. (2022) Multi-Grained Syntactic Dependency-Aware Graph Convolution for Aspect-Based Sentiment Analysis. 2022 International Joint Conference on Neural Networks (IJCNN), Padua, 18-23 July 2022, 1-8.
https://doi.org/10.1109/ijcnn55064.2022.9892472
[19] Yuan, L., Wang, J., Yu, L. and Zhang, X. (2024) Syntactic Graph Attention Network for Aspect-Level Sentiment Analysis. IEEE Transactions on Artificial Intelligence, 5, 140-153.
https://doi.org/10.1109/tai.2022.3227535

Copyright © 2025 by authors and Scientific Research Publishing Inc.

Creative Commons License

This work and the related PDF file are licensed under a Creative Commons Attribution 4.0 International License.