Abstract
Tokenization, or word boundary detection, is a critical first step for most NLP applications. This is often given little attention in English and other languages which use explicit spaces between written words, but standard orthographies for many languages lack explicit markers. Tokenization systems for such languages are usually engineered on an individual basis, with little re-use. The human ability to decode any written language, however, suggests that a general algorithm exists.This thesis presents simple morphologically-based and statistical methods for identifying word boundaries in multiple languages. Statistical methods tend to over-predict, while lexical and morphological methods fail when encountering unknown words. I demonstrate that a generic hybrid approach to tokenization using both morphological and statistical information generalizes well across multiple languages and improves performance over morphological or statistical methods alone, and show that it can be used for efficient tokenization of English, Korean, and Arabic.
Degree
MA
College and Department
Humanities; Linguistics and English Language
Rights
http://lib.byu.edu/about/copyright/
BYU ScholarsArchive Citation
Kearsley, Logan R., "A Hybrid Approach to Cross-Linguistic Tokenization: Morphology with Statistics" (2016). Theses and Dissertations. 5984.
https://scholarsarchive.byu.edu/etd/5984
Date Submitted
2016-06-01
Document Type
Thesis
Handle
http://hdl.lib.byu.edu/1877/etd8642
Keywords
tokenization, lexing, morphological analysis
Language
english