pith. machine review for the scientific record. sign in

arxiv: 1902.10667 · v2 · submitted 2019-02-27 · 💻 cs.CL · cs.AI

Recognition: unknown

Bridging the Gap: Attending to Discontinuity in Identification of Multiword Expressions

Authors on Pith no claims yet
classification 💻 cs.CL cs.AI
keywords mwesdiscontinuityexpressionsinformationmodelmultiwordself-attentionarchitecture
0
0 comments X
read the original abstract

We introduce a new method to tag Multiword Expressions (MWEs) using a linguistically interpretable language-independent deep learning architecture. We specifically target discontinuity, an under-explored aspect that poses a significant challenge to computational treatment of MWEs. Two neural architectures are explored: Graph Convolutional Network (GCN) and multi-head self-attention. GCN leverages dependency parse information, and self-attention attends to long-range relations. We finally propose a combined model that integrates complementary information from both through a gating mechanism. The experiments on a standard multilingual dataset for verbal MWEs show that our model outperforms the baselines not only in the case of discontinuous MWEs but also in overall F-score.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.