Comparing Knowledge-Intensive and Data-Intensive Models for English Resource Semantic Parsing

Authors

  • Junjie Cao Peking University
  • Zi Lin Peking University
  • Weiwei Sun University of Cambridge
  • Xiaojun Wan Wangxuan Institute of Computer Technology, Peking University

Abstract

In this work, we present a phenomenon-oriented comparative analysis of the two dominant approaches in English Resource Semantic (ERS) parsing: classic, knowledge-intensive and neural, data-intensive models. To reflect state-of-the-art neural NLP technologies, a factorization-based parser is introduced that can produce Elementary Dependency Structures much more accurately than previous data-driven parsers. We conduct a suite of tests for different linguistic phenomena to analyze the grammatical competence of different parsers, where we show that, despite comparable performance overall, knowledge- and data-intensive models produce different types of errors, in a way that can be explained by their theoretical properties. This analysis is beneficial to in-depth evaluation of several representative parsing techniques and leads to new directions for parser development.

Author Biography

  • Weiwei Sun, University of Cambridge
    Weiwei Sun is a Senior Lecturer at the Department of Computer Science and Technology of University of Cambridge. She completed her PhD in Department of Computational Linguistics from Saarland University under the supervision of Prof. Hans Uszkoreit. Before that, she studied at Peking University, where she obtained Bachelor of Art (Computational and Applied Linguistics), Bachelor of Science (Computer Science) and Master of Science (Computer Science). Her research lies at the intersection of computational linguistics and natural language processing.  The main topic is symbolic and statistical parsing, with a special focus on parsing into semantic graphs of various flavors.  She is also interested in experimental linguistics.

Published

2024-11-21