Challenges in including extra-linguistic context in pre-trained language models

dc.contributor.authorSorodoc, Ionut-Teodor
dc.contributor.authorAina, Laura
dc.contributor.authorBoleda, Gemma
dc.date.accessioned2023-03-15T07:20:51Z
dc.date.available2023-03-15T07:20:51Z
dc.date.issued2022
dc.descriptionComunicació presentada a Third Workshop on Insights from Negative Results in NLP (Insights 2022), celebrat el 26 de maig de 2022 a Dublín, Irlanda.
dc.description.abstractTo successfully account for language, computational models need to take into account both the linguistic context (the content of the utterances) and the extra-linguistic context (for instance, the participants in a dialogue). We focus on a referential task that asks models to link entity mentions in a TV show to the corresponding characters, and design an architecture that attempts to account for both kinds of context. In particular, our architecture combines a previously proposed specialized module (an “entity library”) for character representation with transfer learning from a pre-trained language model. We find that, although the model does improve linguistic contextualization, it fails to successfully integrate extra-linguistic information about the participants in the dialogue. Our work shows that it is very challenging to incorporate extra-linguistic information into pretrained language models.
dc.description.sponsorshipThis project has received funding from the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (grant agreement No 715154). We are also grateful to the NVIDIA Corporation for the donation of GPUs used for this research.
dc.format.mimetypeapplication/pdf
dc.identifier.citationSorodoc IT, Aina L, Boleda G. Challenges in including extra-linguistic context in pre-trained language models. In: Tafreshi S, Sedoc J, Rogers A, Drozd A, Rumshisky A, Akula A, editors. The Third Workshop on Insights from Negative Results in NLP (Insights 2022): proceedings of the Workshop; 2022 May 26; Dublin, Ireland. [Stroudsburg]: Association for Computational Linguistics; 2022. p. 134-8. DOI: 10.18653/v1/2022.insights-1.18
dc.identifier.doihttp://dx.doi.org/10.18653/v1/2022.insights-1.18
dc.identifier.urihttp://hdl.handle.net/10230/56228
dc.language.isoeng
dc.publisherACL (Association for Computational Linguistics)
dc.relation.ispartofTafreshi S, Sedoc J, Rogers A, Drozd A, Rumshisky A, Akula A, editors. The Third Workshop on Insights from Negative Results in NLP (Insights 2022): proceedings of the Workshop; 2022 May 26; Dublin, Ireland. [Stroudsburg]: Association for Computational Linguistics; 2022. p. 134-8.
dc.relation.projectIDinfo:eu-repo/grantAgreement/EC/H2020/715154
dc.rights© ACL, Creative Commons Attribution 4.0 License
dc.rights.accessRightsinfo:eu-repo/semantics/openAccess
dc.rights.urihttp://creativecommons.org/licenses/by/4.0/
dc.subject.otherModels lingüístics
dc.titleChallenges in including extra-linguistic context in pre-trained language models
dc.typeinfo:eu-repo/semantics/conferenceObject
dc.type.versioninfo:eu-repo/semantics/publishedVersion

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Sorodoc_Thi_Chal.pdf
Size:
385.3 KB
Format:
Adobe Portable Document Format

License

Rights