Self-attentive Biaffine Dependency Parsing
Self-attentive Biaffine Dependency Parsing
Ying Li, Zhenghua Li, Min Zhang, Rui Wang, Sheng Li, Luo Si
Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence
Main track. Pages 5067-5073.
https://doi.org/10.24963/ijcai.2019/704
The current state-of-the-art dependency parsing approaches employ BiLSTMs to encode input sentences.Motivated by the success of the transformer-based machine translation, this work for the first time applies the self-attention mechanism to dependency parsing as the replacement of the BiLSTM-based encoders, leading to competitive performance on both English and Chinese benchmark data. Based on the detailed error analysis, we then combine the power of both BiLSTM and self-attention via model ensembles, demonstrating their complementary capability of capturing contextual information. Finally, we explore the recently proposed contextualized word representations as extra input features, and further improve the parsing performance.
Keywords:
Natural Language Processing: Natural Language Processing
Natural Language Processing: Tagging, chunking, and parsing