Latent variable models of distributional lexical semantics

Computer Sciences === In order to respond to increasing demand for natural language interfaces---and provide meaningful insight into user query intent---fast, scalable lexical semantic models with flexible representations are needed. Human concept organization is a rich phenomenon that has yet to be...

Full description

Bibliographic Details
Main Author: Reisinger, Joseph Simon
Format: Others
Published: 2014
Subjects:
Online Access:http://hdl.handle.net/2152/26889
id ndltd-UTEXAS-oai-repositories.lib.utexas.edu-2152-26889
record_format oai_dc
spelling ndltd-UTEXAS-oai-repositories.lib.utexas.edu-2152-268892015-09-20T17:27:40ZLatent variable models of distributional lexical semanticsReisinger, Joseph SimonLexical semanticsNatural language processingMachine learningComputer SciencesIn order to respond to increasing demand for natural language interfaces---and provide meaningful insight into user query intent---fast, scalable lexical semantic models with flexible representations are needed. Human concept organization is a rich phenomenon that has yet to be accounted for by a single coherent psychological framework: Concept generalization is captured by a mixture of prototype and exemplar models, and local taxonomic information is available through multiple overlapping organizational systems. Previous work in computational linguistics on extracting lexical semantic information from unannotated corpora does not provide adequate representational flexibility and hence fails to capture the full extent of human conceptual knowledge. In this thesis I outline a family of probabilistic models capable of capturing important aspects of the rich organizational structure found in human language that can predict contextual variation, selectional preference and feature-saliency norms to a much higher degree of accuracy than previous approaches. These models account for cross-cutting structure of concept organization---i.e. selective attention, or the notion that humans make use of different categorization systems for different kinds of generalization tasks---and can be applied to Web-scale corpora. Using these models, natural language systems will be able to infer a more comprehensive semantic relations, which in turn may yield improved systems for question answering, text classification, machine translation, and information retrieval.text2014-10-24T18:38:51Z2012-052014-10-24May 20122014-10-24T18:38:51ZThesisapplication/pdfhttp://hdl.handle.net/2152/26889
collection NDLTD
format Others
sources NDLTD
topic Lexical semantics
Natural language processing
Machine learning
spellingShingle Lexical semantics
Natural language processing
Machine learning
Reisinger, Joseph Simon
Latent variable models of distributional lexical semantics
description Computer Sciences === In order to respond to increasing demand for natural language interfaces---and provide meaningful insight into user query intent---fast, scalable lexical semantic models with flexible representations are needed. Human concept organization is a rich phenomenon that has yet to be accounted for by a single coherent psychological framework: Concept generalization is captured by a mixture of prototype and exemplar models, and local taxonomic information is available through multiple overlapping organizational systems. Previous work in computational linguistics on extracting lexical semantic information from unannotated corpora does not provide adequate representational flexibility and hence fails to capture the full extent of human conceptual knowledge. In this thesis I outline a family of probabilistic models capable of capturing important aspects of the rich organizational structure found in human language that can predict contextual variation, selectional preference and feature-saliency norms to a much higher degree of accuracy than previous approaches. These models account for cross-cutting structure of concept organization---i.e. selective attention, or the notion that humans make use of different categorization systems for different kinds of generalization tasks---and can be applied to Web-scale corpora. Using these models, natural language systems will be able to infer a more comprehensive semantic relations, which in turn may yield improved systems for question answering, text classification, machine translation, and information retrieval. === text
author Reisinger, Joseph Simon
author_facet Reisinger, Joseph Simon
author_sort Reisinger, Joseph Simon
title Latent variable models of distributional lexical semantics
title_short Latent variable models of distributional lexical semantics
title_full Latent variable models of distributional lexical semantics
title_fullStr Latent variable models of distributional lexical semantics
title_full_unstemmed Latent variable models of distributional lexical semantics
title_sort latent variable models of distributional lexical semantics
publishDate 2014
url http://hdl.handle.net/2152/26889
work_keys_str_mv AT reisingerjosephsimon latentvariablemodelsofdistributionallexicalsemantics
_version_ 1716824110020952064