Show simple item record

dc.contributor.author Morales, Cristian
dc.date.accessioned 2021-02-04T14:07:34Z
dc.date.available 2021-02-04T14:07:34Z
dc.date.issued 2020-09
dc.identifier.uri http://hdl.handle.net/10230/46339
dc.description Treball fi de màster de: Master in Intelligent Interactive Systems
dc.description Tutors: Leo Wanner, Bernd Bohnet
dc.description.abstract Graph-structured data is ubiquitous in the field of Natural Language Processing. For instance, directed acyclic graphs are used in semantic and syntactic dependency representations. Thus, several applications in NLP use graph-structured data, such as sequence labeling, neural machine translation and relation extraction. Most approaches first linearize graphs and then apply off-the-shelf algorithms, which leave out important information of node connectivity. It is clear that the state-of-the art falls short in offering graph-to-graph transduction models. The motivation of this thesis is to expand the limited literature in tree-to-tree learning and provide an instrument capable of treebank transformations, that could enrich the corpora available in NLP. The starting point is previous work on Gated Graph Neural Networks [1], which we modified to output sequences per node as opposed to sequences per graph. We also modified the general architecture by using two GGNNs, one responsible for predicting heads and the other one for predicting edge types. For testing, we used the Stanford dependency treebank and the Matsumoto dependency treebank. These treebanks are substantially different especially in the granularity of their dependency tagsets. The proposed model achieved over 95% Labeled Attachment Score (LAS) when converting from one treebank to the other. As compared to the baseline, which ignores graph data, it achieved an average improvement of 16.42% in LAS, which highlights the value of incorporating graph-structured data. We also showed that feeding the network with each node’s position within the sentence yielded a 2.32% LAS improvement. Thus, including sequential data proved to be beneficial. We concluded that GGNNs are capable of tree-to-tree transduction and that this research is a step forward in bringing attention to graph-to-graph transduction in NLP.
dc.format.mimetype application/pdf
dc.language.iso eng
dc.rights Reconeixement-NoComercial-SenseObraDerivada 3.0 Espanya (CC BY-NC-ND 3.0 ES)
dc.rights.uri https://creativecommons.org/licenses/by-nc-nd/3.0/es
dc.title Neural Tree-to-Tree Transduction
dc.type info:eu-repo/semantics/masterThesis
dc.subject.keyword Tree transduction
dc.subject.keyword Graphs
dc.subject.keyword Graph-to-graph learning
dc.subject.keyword Natural Language Processing
dc.rights.accessRights info:eu-repo/semantics/openAccess

Thumbnail

This item appears in the following Collection(s)

Show simple item record

Search DSpace


Advanced Search

Browse

My Account

Statistics

In collaboration with Compliant to Partaking