Neural Language Models with Explicit Coreference Decision

Coreference is an important and frequent concept in any form of discourse, and Coreference Resolution (CR) a widely used task in Natural Language Understanding (NLU). In this thesis, we implement and explore two recent models that include the concept of coreference in Recurrent Neural Network (RNN)-...

Full description

Bibliographic Details
Main Author: Kunz, Jenny
Format: Others
Language:English
Published: Uppsala universitet, Institutionen för lingvistik och filologi 2019
Subjects:
LM
RNN
Online Access:http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-371827
id ndltd-UPSALLA1-oai-DiVA.org-uu-371827
record_format oai_dc
spelling ndltd-UPSALLA1-oai-DiVA.org-uu-3718272019-01-11T05:52:08ZNeural Language Models with Explicit Coreference DecisionengKunz, JennyUppsala universitet, Institutionen för lingvistik och filologi2019CoreferenceReferenceEntityLanguage ModelsLMNeural NetworksRNNAttentionDeep LearningLanguage Technology (Computational Linguistics)Språkteknologi (språkvetenskaplig databehandling)Coreference is an important and frequent concept in any form of discourse, and Coreference Resolution (CR) a widely used task in Natural Language Understanding (NLU). In this thesis, we implement and explore two recent models that include the concept of coreference in Recurrent Neural Network (RNN)-based Language Models (LM). Entity and reference decisions are modeled explicitly in these models using attention mechanisms. Both models learn to save the previously observed entities in a set and to decide if the next token created by the LM is a mention of one of the entities in the set, an entity that has not been observed yet, or not an entity. After a theoretical analysis where we compare the two LMs to each other and to a state of the art Coreference Resolution system, we perform an extensive quantitative and qualitative analysis. For this purpose, we train the two models and a classical RNN-LM as the baseline model on the OntoNotes 5.0 corpus with coreference annotation. While we do not reach the baseline in the perplexity metric, we show that the models’ relative performance on entity tokens has the potential to improve when including the explicit entity modeling. We show that the most challenging point in the systems is the decision if the next token is an entity token, while the decision which entity the next token refers to performs comparatively well. Our analysis in the context of a text generation task shows that a wide-spread error source for the mention creation process is the confusion of tokens that refer to related but different entities in the real world, presumably a result of the context-based word representations in the models. Our re-implementation of the DeepMind model by Yang et al. 2016 performs notably better than the re-implementation of the EntityNLM model by Ji et al. 2017 with a perplexity of 107 compared to a perplexity of 131. Student thesisinfo:eu-repo/semantics/bachelorThesistexthttp://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-371827application/pdfinfo:eu-repo/semantics/openAccess
collection NDLTD
language English
format Others
sources NDLTD
topic Coreference
Reference
Entity
Language Models
LM
Neural Networks
RNN
Attention
Deep Learning
Language Technology (Computational Linguistics)
Språkteknologi (språkvetenskaplig databehandling)
spellingShingle Coreference
Reference
Entity
Language Models
LM
Neural Networks
RNN
Attention
Deep Learning
Language Technology (Computational Linguistics)
Språkteknologi (språkvetenskaplig databehandling)
Kunz, Jenny
Neural Language Models with Explicit Coreference Decision
description Coreference is an important and frequent concept in any form of discourse, and Coreference Resolution (CR) a widely used task in Natural Language Understanding (NLU). In this thesis, we implement and explore two recent models that include the concept of coreference in Recurrent Neural Network (RNN)-based Language Models (LM). Entity and reference decisions are modeled explicitly in these models using attention mechanisms. Both models learn to save the previously observed entities in a set and to decide if the next token created by the LM is a mention of one of the entities in the set, an entity that has not been observed yet, or not an entity. After a theoretical analysis where we compare the two LMs to each other and to a state of the art Coreference Resolution system, we perform an extensive quantitative and qualitative analysis. For this purpose, we train the two models and a classical RNN-LM as the baseline model on the OntoNotes 5.0 corpus with coreference annotation. While we do not reach the baseline in the perplexity metric, we show that the models’ relative performance on entity tokens has the potential to improve when including the explicit entity modeling. We show that the most challenging point in the systems is the decision if the next token is an entity token, while the decision which entity the next token refers to performs comparatively well. Our analysis in the context of a text generation task shows that a wide-spread error source for the mention creation process is the confusion of tokens that refer to related but different entities in the real world, presumably a result of the context-based word representations in the models. Our re-implementation of the DeepMind model by Yang et al. 2016 performs notably better than the re-implementation of the EntityNLM model by Ji et al. 2017 with a perplexity of 107 compared to a perplexity of 131.
author Kunz, Jenny
author_facet Kunz, Jenny
author_sort Kunz, Jenny
title Neural Language Models with Explicit Coreference Decision
title_short Neural Language Models with Explicit Coreference Decision
title_full Neural Language Models with Explicit Coreference Decision
title_fullStr Neural Language Models with Explicit Coreference Decision
title_full_unstemmed Neural Language Models with Explicit Coreference Decision
title_sort neural language models with explicit coreference decision
publisher Uppsala universitet, Institutionen för lingvistik och filologi
publishDate 2019
url http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-371827
work_keys_str_mv AT kunzjenny neurallanguagemodelswithexplicitcoreferencedecision
_version_ 1718813189683019776