TAILIEUCHUNG - Báo cáo khoa học: "Unsupervised Learning of Dependency Structure for Language Modeling"

This paper presents a dependency language model (DLM) that captures linguistic constraints via a dependency structure, ., a set of probabilistic dependencies that express the relations between headwords of each phrase in a sentence by an acyclic, planar, undirected graph. Our contributions are three-fold. First, we incorporate the dependency structure into an n-gram language model to capture long distance word dependency. Second, we present an unsupervised learning method that discovers the dependency structure of a sentence using a bootstrapping procedure. . | Unsupervised Learning of Dependency Structure for Language Modeling Jianfeng Gao Microsoft Research Asia 49 Zhichun Road Haidian District Beijing 100080 China jfgao@ Abstract This paper presents a dependency language model DLM that captures linguistic constraints via a dependency structure . a set of probabilistic dependencies that express the relations between headwords of each phrase in a sentence by an acyclic planar undirected graph. Our contributions are three-fold. First we incorporate the dependency structure into an n-gram language model to capture long distance word dependency. Second we present an unsupervised learning method that discovers the dependency structure of a sentence using a bootstrapping procedure. Finally we evaluate the proposed models on a realistic application Japanese Kana-Kanji conversion . Experiments show that the best DLM achieves an error rate reduction over the word trigram model. 1 Introduction In recent years many efforts have been made to utilize linguistic structure in language modeling which for practical reasons is still dominated by trigram-based language models. There are two major obstacles to successfully incorporating linguistic structure into a language model 1 capturing longer distance word dependencies leads to higher-order n-gram models where the number of parameters is usually too large to estimate 2 capturing deeper linguistic relations in a language model requires a large annotated training corpus and a decoder that assigns linguistic structure which are not always available. This paper presents a new dependency language model DLM that captures long distance linguistic constraints between words via a dependency Hisami Suzuki Microsoft Research One Microsoft Way Redmond WA 98052 USA hisamis@ structure . a set of probabilistic dependencies that capture linguistic relations between headwords of each phrase in a sentence. To deal with the first obstacle mentioned above we .

TỪ KHÓA LIÊN QUAN
TAILIEUCHUNG - Chia sẻ tài liệu không giới hạn
Địa chỉ : 444 Hoang Hoa Tham, Hanoi, Viet Nam
Website : tailieuchung.com
Email : tailieuchung20@gmail.com
Tailieuchung.com là thư viện tài liệu trực tuyến, nơi chia sẽ trao đổi hàng triệu tài liệu như luận văn đồ án, sách, giáo trình, đề thi.
Chúng tôi không chịu trách nhiệm liên quan đến các vấn đề bản quyền nội dung tài liệu được thành viên tự nguyện đăng tải lên, nếu phát hiện thấy tài liệu xấu hoặc tài liệu có bản quyền xin hãy email cho chúng tôi.
Đã phát hiện trình chặn quảng cáo AdBlock
Trang web này phụ thuộc vào doanh thu từ số lần hiển thị quảng cáo để tồn tại. Vui lòng tắt trình chặn quảng cáo của bạn hoặc tạm dừng tính năng chặn quảng cáo cho trang web này.