TAILIEUCHUNG - Báo cáo khoa học: "Neutralizing Linguistically Problematic Annotations in Unsupervised Dependency Parsing Evaluation"

Dependency parsing is a central NLP task. In this paper we show that the common evaluation for unsupervised dependency parsing is highly sensitive to problematic annotations. We show that for three leading unsupervised parsers (Klein and Manning, 2004; Cohen and Smith, 2009; Spitkovsky et al., 2010a), a small set of parameters can be found whose modification yields a significant improvement in standard evaluation measures. These parameters correspond to local cases where no linguistic consensus exists as to the proper gold annotation. . | Neutralizing Linguistically Problematic Annotations in Unsupervised Dependency Parsing Evaluation Roy Schwartz1 Omri Abend1 Roi Reichart2 Ari Rappoport1 institute of Computer Science Hebrew University of Jerusalem roys02 omria011arir @ 2Computer Science and Artificial Intelligence Laboratory Massachusetts Institute of Technology roiri@ Abstract Dependency parsing is a central NLP task. In this paper we show that the common evaluation for unsupervised dependency parsing is highly sensitive to problematic annotations. We show that for three leading unsupervised parsers Klein and Manning 2004 Cohen and Smith 2009 Spitkovsky et al. 2010a a small set of parameters can be found whose modification yields a significant improvement in standard evaluation measures. These parameters correspond to local cases where no linguistic consensus exists as to the proper gold annotation. Therefore the standard evaluation does not provide a true indication of algorithm quality. We present a new measure Neutral Edge Direction NED and show that it greatly reduces this undesired phenomenon. 1 Introduction Unsupervised induction of dependency parsers is a major NLP task that attracts a substantial amount of research Klein and Manning 2004 Cohen et al. 2008 Headden et al. 2009 Spitkovsky et al. 2010a Gillenwater et al. 2010 Berg-Kirkpatrick et al. 2010 Blunsom and Cohn 2010 inter alia . Parser quality is usually evaluated by comparing its output to a gold standard whose annotations are linguistically motivated. However there are cases in which there is no linguistic consensus as to what the correct annotation is Kubler et al. 2009 . Examples include which verb is the head in a verb group structure . can or eat in can eat and which Omri Abend is grateful to the Azrieli Foundation for the award of an Azrieli Fellowship. 663 noun is the head in a sequence of proper nouns . John or Doe in John Doe . We refer to such annotations as linguistically problematic. For

TỪ KHÓA LIÊN QUAN
TAILIEUCHUNG - Chia sẻ tài liệu không giới hạn
Địa chỉ : 444 Hoang Hoa Tham, Hanoi, Viet Nam
Website : tailieuchung.com
Email : tailieuchung20@gmail.com
Tailieuchung.com là thư viện tài liệu trực tuyến, nơi chia sẽ trao đổi hàng triệu tài liệu như luận văn đồ án, sách, giáo trình, đề thi.
Chúng tôi không chịu trách nhiệm liên quan đến các vấn đề bản quyền nội dung tài liệu được thành viên tự nguyện đăng tải lên, nếu phát hiện thấy tài liệu xấu hoặc tài liệu có bản quyền xin hãy email cho chúng tôi.
Đã phát hiện trình chặn quảng cáo AdBlock
Trang web này phụ thuộc vào doanh thu từ số lần hiển thị quảng cáo để tồn tại. Vui lòng tắt trình chặn quảng cáo của bạn hoặc tạm dừng tính năng chặn quảng cáo cho trang web này.