Fine-Grained Linguistic Soft Constraints on Statistical Natural Language Processing Models
MARYLAND UNIV COLLEGE PARK
Pagination or Media Count:
This dissertation focuses on effective combination of data-driven natural language processing NLP approaches with linguistic knowledge sources that are based on manual text annotation or word grouping according to semantic commonalities. I gainfully apply fine-grained linguistic soft constraints - of syntactic or semantic nature - on statistical NLP models, evaluated in end-to-end state-of-the-art statistical machine translation SMT systems. The introduction of semantic soft constraints involves intrinsic evaluation on word-pair similarity ranking tasks, extension from words to phrases, application in a novel distributional paraphrase generation technique and an introduction of a generalized framework of which these soft semantic and syntactic constraints can be viewed as instances, and in which they can be potentially combined. Fine granularity is key in the successful combination of these soft constraints in many cases. I show how to softly constrain SMT models by adding fine-grained weighted features, each preferring translation of only a specific syntactic constituent. Previous attempts using coarse-grained features yielded negative results. I also show how to softly constrain corpus-based semantic models of words distributional profiles to effectively create word-sense-aware models, by using semantic word grouping information found in a manually compiled thesaurus. Previous attempts using hard constraints and resulting in aggregated, coarse-grained models, yielded lower gains. A novel paraphrase generation technique incorporating these soft semantic constraints is then also evaluated in a SMT system. This paraphrasing technique is based on the Distributional Hypothesis. The main advantage of this novel technique over current pivoting techniques for paraphrasing is the independence from parallel texts, which are a limited resource.
- Statistics and Probability