TAILIEUCHUNG - Báo cáo khoa học: "Self-Training for Biomedical Parsing"

Parser self-training is the technique of taking an existing parser, parsing extra data and then creating a second parser by treating the extra data as further training data. Here we apply this technique to parser adaptation. In particular, we self-train the standard Charniak/Johnson Penn-Treebank parser using unlabeled biomedical abstracts. This achieves an f -score of on a standard test set of biomedical abstracts from the Genia corpus. This is a 20% error reduction over the best previous result on biomedical data ( on the same test set). . | Self-Training for Biomedical Parsing David McClosky and Eugene Charniak Brown Laboratory for Linguistic Information Processing BLLIP Brown University Providence RI 02912 dmcc ec @ Abstract Parser self-training is the technique of taking an existing parser parsing extra data and then creating a second parser by treating the extra data as further training data. Here we apply this technique to parser adaptation. In particular we self-train the standard Char-niak Johnson Penn-Treebank parser using unlabeled biomedical abstracts. This achieves an -score of on a standard test set of biomedical abstracts from the Genia corpus. This is a 20 error reduction over the best previous result on biomedical data on the same test set . 1 Introduction Parser self-training is the technique of taking an existing parser parsing extra data and then creating a second parser by treating the extra data as further training data. While for many years it was thought not to help state-of-the art parsers more recent work has shown otherwise. In this paper we apply this technique to parser adaptation. In particular we self-train the standard Charniak Johnson Penn-Treebank C J parser using unannotated biomedical data. As is well known biomedical data is hard on parsers because it is so far from more standard English. To our knowledge this is the first application of self-training where the gap between the training and self-training data is so large. In section two we look at previous work. In particular we note that there is in fact very little data on self-training when the corpora for self-training is so different from the original labeled data. Section three describes our main experiment on standard test data Clegg and Shepherd 2005 . Section four looks at some preliminary results we obtained on development data that show in slightly more detail how selftraining improved the parser. We conclude in section five. 2 Previous Work While self-training has worked in several .

TAILIEUCHUNG - Chia sẻ tài liệu không giới hạn
Địa chỉ : 444 Hoang Hoa Tham, Hanoi, Viet Nam
Website : tailieuchung.com
Email : tailieuchung20@gmail.com
Tailieuchung.com là thư viện tài liệu trực tuyến, nơi chia sẽ trao đổi hàng triệu tài liệu như luận văn đồ án, sách, giáo trình, đề thi.
Chúng tôi không chịu trách nhiệm liên quan đến các vấn đề bản quyền nội dung tài liệu được thành viên tự nguyện đăng tải lên, nếu phát hiện thấy tài liệu xấu hoặc tài liệu có bản quyền xin hãy email cho chúng tôi.
Đã phát hiện trình chặn quảng cáo AdBlock
Trang web này phụ thuộc vào doanh thu từ số lần hiển thị quảng cáo để tồn tại. Vui lòng tắt trình chặn quảng cáo của bạn hoặc tạm dừng tính năng chặn quảng cáo cho trang web này.