This paper presents a method for aspect based sentiment classification tasks, named convolutional multi-head self-attention memory network(CMA-Mem Net). This is an improved model based on memory networks, and makes it...This paper presents a method for aspect based sentiment classification tasks, named convolutional multi-head self-attention memory network(CMA-Mem Net). This is an improved model based on memory networks, and makes it possible to extract more rich and complex semantic information from sequences and aspects. In order to fix the memory network’s inability to capture context-related information on a word-level,we propose utilizing convolution to capture n-gram grammatical information. We use multi-head self-attention to make up for the problem where the memory network ignores the semantic information of the sequence itself. Meanwhile, unlike most recurrent neural network(RNN) long short term memory(LSTM), gated recurrent unit(GRU) models, we retain the parallelism of the network. We experiment on the open datasets Sem Eval-2014 Task 4 and Sem Eval-2016 Task 6. Compared with some popular baseline methods, our model performs excellently.展开更多
基金supported by the National Key Research and Development Program of China(2018YFC0830700)。
文摘This paper presents a method for aspect based sentiment classification tasks, named convolutional multi-head self-attention memory network(CMA-Mem Net). This is an improved model based on memory networks, and makes it possible to extract more rich and complex semantic information from sequences and aspects. In order to fix the memory network’s inability to capture context-related information on a word-level,we propose utilizing convolution to capture n-gram grammatical information. We use multi-head self-attention to make up for the problem where the memory network ignores the semantic information of the sequence itself. Meanwhile, unlike most recurrent neural network(RNN) long short term memory(LSTM), gated recurrent unit(GRU) models, we retain the parallelism of the network. We experiment on the open datasets Sem Eval-2014 Task 4 and Sem Eval-2016 Task 6. Compared with some popular baseline methods, our model performs excellently.