Contribute to Semantic Parsing

Broad-Coverage Semantic Parsing as Transduction

Sheng Zhang, Xutai Ma, Kevin Duh, Benjamin Van Durme. EMNLP 2019

[ArXiV] [PDF]      
amr transformer glove cnn

We unify different broad-coverage semantic parsing tasks into a transduction parsing paradigm, and propose an attention-based neural transducer that incrementally builds meaning representation via a sequence of semantic relations. By leveraging multiple attention mechanisms, the neural transducer can be effectively trained without relying on a pre-trained aligner. Experiments separately conducted on three broad-coverage semantic parsing tasks – AMR, SDP and UCCA – demonstrate that our attention-based neural transducer improves the state of the art on both AMR and UCCA, and is competitive with the state of the art on SDP.