TAILIEUCHUNG - Báo cáo khoa học: "A Pylonic Decision-Tree Language Model with Optimal Question Selection"

This paper discusses a decision-tree approach to the problem of assigning probabilities to words following a given text. In contrast with previous decision-tree language model attempts, an algorithm for selecting nearly optimal questions is considered. The model is to be tested on a standard task, The Wall Street Journal, allowing a fair comparison with the well-known trigram model. | A Pylonic Decision-Tree Language Model with Optimal Question Selection Adrian Corduneanu University of Toronto 73 Saint George St 299 Toronto Ontario M5S 2E5 Canada g7adrian@ Abstract This paper discusses a decision-tree approach to the problem of assigning probabilities to words following a given text. In contrast with previous decision-tree language model attempts an algorithm for selecting nearly optimal questions is considered. The model is to be tested on a standard task The Wall Street Journal allowing a fair comparison with the well-known trigram model. 1 Introduction In many applications such as automatic speech recognition machine translation spelling correction etc. a statistical language model LM is needed to assign probabilities to sentences. This probability assignment may be used . to choose one of many transcriptions hypothesized by the recognizer or to make decisions about capitalization. Without any loss of generality we consider models that operate left-to-right on the sentences assigning a probability to the next word given its word history. Specifically we consider statistical LM s which compute probabilities of the type P wn I W1 W2 . Wn-1 where Wi denotes the i-th word in the text. Even for a small vocabulary the space of word histories is so large that any attempt to estimate the conditional probabilities for each distinct history from raw frequencies is infeasible. To make the problem manageable one partitions the word histories into some classes C wi W2 . wn_i and identifies the word probabilities with p wn I C iiq W2 . Wn-1 . Such probabilities are easier to estimate as each class gets significantly more counts from a training corpus. With this setup building a language model becomes a classification problem group the word histories into a small number of classes while preserving their predictive power. Currently popular 7V-gram models classify the word histories by their last N 1 words. N varies from 2 to 4 and the .

TAILIEUCHUNG - Chia sẻ tài liệu không giới hạn
Địa chỉ : 444 Hoang Hoa Tham, Hanoi, Viet Nam
Website : tailieuchung.com
Email : tailieuchung20@gmail.com
Tailieuchung.com là thư viện tài liệu trực tuyến, nơi chia sẽ trao đổi hàng triệu tài liệu như luận văn đồ án, sách, giáo trình, đề thi.
Chúng tôi không chịu trách nhiệm liên quan đến các vấn đề bản quyền nội dung tài liệu được thành viên tự nguyện đăng tải lên, nếu phát hiện thấy tài liệu xấu hoặc tài liệu có bản quyền xin hãy email cho chúng tôi.
Đã phát hiện trình chặn quảng cáo AdBlock
Trang web này phụ thuộc vào doanh thu từ số lần hiển thị quảng cáo để tồn tại. Vui lòng tắt trình chặn quảng cáo của bạn hoặc tạm dừng tính năng chặn quảng cáo cho trang web này.