TAILIEUCHUNG - Báo cáo khoa học: "Incremental Parsing with the Perceptron Algorithm"

This paper describes an incremental parsing approach where parameters are estimated using a variant of the perceptron algorithm. A beam-search algorithm is used during both training and decoding phases of the method. The perceptron approach was implemented with the same feature set as that of an existing generative model (Roark, 2001a), and experimental results show that it gives competitive performance to the generative model on parsing the Penn treebank. We demonstrate that training a perceptron model to combine with the generative model during search provides a percent F-measure improvement over the generative model alone, to percent. . | Incremental Parsing with the Perceptron Algorithm Michael Collins MIT CSAIL mcollins@ Brian Roark AT T Labs - Research roark@ Abstract This paper describes an incremental parsing approach where parameters are estimated using a variant of the perceptron algorithm. A beam-search algorithm is used during both training and decoding phases of the method. The perceptron approach was implemented with the same feature set as that of an existing generative model Roark 2001a and experimental results show that it gives competitive performance to the generative model on parsing the Penn treebank. We demonstrate that training a perceptron model to combine with the generative model during search provides a percent F-measure improvement over the generative model alone to percent. 1 Introduction In statistical approaches to NLP problems such as tagging or parsing it seems clear that the representation used as input to a learning algorithm is central to the accuracy of an approach. In an ideal world the designer of a parser or tagger would be free to choose any features which might be useful in discriminating good from bad structures without concerns about how the features interact with the problems of training parameter estimation or decoding search for the most plausible candidate under the model . To this end a number of recently proposed methods allow a model to incorporate arbitrary global features of candidate analyses or parses. Examples of such techniques are Markov Random Fields Rat-naparkhi et al. 1994 Abney 1997 Della Pietra et al. 1997 Johnson et al. 1999 and boosting or perceptron approaches to reranking Freund et al. 1998 Collins 2000 Collins and Duffy 2002 . A drawback of these approaches is that in the general case they can require exhaustive enumeration of the set of candidates for each input sentence in both the training and decoding phases1. For example Johnson et al. 1999 and Riezler et al. 2002 use all parses generated by .

TAILIEUCHUNG - Chia sẻ tài liệu không giới hạn
Địa chỉ : 444 Hoang Hoa Tham, Hanoi, Viet Nam
Website : tailieuchung.com
Email : tailieuchung20@gmail.com
Tailieuchung.com là thư viện tài liệu trực tuyến, nơi chia sẽ trao đổi hàng triệu tài liệu như luận văn đồ án, sách, giáo trình, đề thi.
Chúng tôi không chịu trách nhiệm liên quan đến các vấn đề bản quyền nội dung tài liệu được thành viên tự nguyện đăng tải lên, nếu phát hiện thấy tài liệu xấu hoặc tài liệu có bản quyền xin hãy email cho chúng tôi.
Đã phát hiện trình chặn quảng cáo AdBlock
Trang web này phụ thuộc vào doanh thu từ số lần hiển thị quảng cáo để tồn tại. Vui lòng tắt trình chặn quảng cáo của bạn hoặc tạm dừng tính năng chặn quảng cáo cho trang web này.