Exploring Transformer-Based Contextual Knowledge Graph Embeddings : How the Design of the Attention Mask and the Input Structure Affect Learning in Transformer Models

The availability and use of knowledge graphs have become commonplace as a compact storage of information and for lookup of facts. However, the discrete representation makes the knowledge graph unavailable for tasks that need a continuous representation, such as predicting relationships between entit...

Full description

Bibliographic Details
Main Author: Holmström, Oskar
Format: Others
Language:English
Published: Linköpings universitet, Artificiell intelligens och integrerade datorsystem 2021
Subjects:
Online Access:http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-175400