Abstract

The end-to-end performance of natural language processing systems for compound tasks, such as question answering and textual entailment, is often hampered by use of a greedy 1-best pipeline architecture, which causes errors to propagate and compound at each stage. We present a novel architecture, which models these pipelines as Bayesian networks, with each low level task corresponding to a variable in the network, and then we perform approximate inference to find the best labeling. Our approach is extremely simple to apply but gains the benefits of sampling the entire distribution over labels at each stage in the pipeline. We apply our method to two tasks -- semantic role labeling and recognizing textual entailment -- and achieve useful performance gains from the superior pipeline architecture.


Original document

The different versions of the original document can be found in:

http://www.robotics.stanford.edu/~ang/papers/emnlp06-bayesianpipeline.pdf,
http://ai.stanford.edu/~ang/papers/emnlp06-bayesianpipeline.pdf,
http://www-nlp.stanford.edu/~manning/papers/pipeline.pdf,
https://www.aclweb.org/anthology/W06-1673.pdf,
https://www.aclweb.org/anthology/W06-1673,
https://dblp.uni-trier.de/db/conf/emnlp/emnlp2006.html#FinkelMN06,
https://dl.acm.org/citation.cfm?id=1610075.1610162,
https://academic.microsoft.com/#/detail/2083195487
Back to Top

Document information

Published on 01/01/2010

Volume 2010, 2010
DOI: 10.3115/1610075.1610162
Licence: CC BY-NC-SA license

Document Score

0

Views 2
Recommendations 0

Share this document

Keywords

claim authorship

Are you one of the authors of this document?