When the whole is less than the sum of its parts: how composition affects PMI values in distributional semantic vectors

dc.contributor.authorPaperno, Denis
dc.contributor.authorBaroni, Marco
dc.date.accessioned2020-12-02T09:07:47Z
dc.date.available2020-12-02T09:07:47Z
dc.date.issued2016
dc.description.abstractDistributional semantic models, deriving vector-based word representations from patterns of word usage in corpora, have many useful applications (Turney and Pantel 2010). Recently, there has been interest in compositional distributional models, which derive vectors for phrases from representations of their constituent words (Mitchell and Lapata 2010). Often, the values of distributional vectors are pointwise mutual information (PMI) scores obtained from raw co-occurrence counts. In this article we study the relation between the PMI dimensions of a phrase vector and its components in order to gain insights into which operations an adequate composition model should perform. We show mathematically that the difference between the PMI dimension of a phrase vector and the sum of PMIs in the corresponding dimensions of the phrase's parts is an independently interpretable value, namely, a quantification of the impact of the context associated with the relevant dimension on the phrase's internal cohesion, as also measured by PMI. We then explore this quantity empirically, through an analysis of adjective–noun composition.en
dc.description.sponsorshipWe would like to thank the Computational Linguistics editor and reviewers: Yoav Goldberg, Omer Levy, Katya Tentori, Germán Kruszewski, Nghia Pham, and the other members of the Composes team for useful feedback. Our work is funded by ERC 2011 Starting Independent Research Grant n. 283554 (COMPOSES).
dc.format.mimetypeapplication/pdf
dc.identifier.citationPaperno D, Baroni M. When the whole is less than the sum of its parts: how composition affects PMI values in distributional semantic vectors. Computational Linguistics. 2016 Jun;42(2):345-50. DOI: 10.1162/COLI_a_00250
dc.identifier.doihttp://dx.doi.org/10.1162/COLI_a_00250
dc.identifier.issn0891-2017
dc.identifier.urihttp://hdl.handle.net/10230/45934
dc.language.isoeng
dc.publisherMIT Press
dc.relation.ispartofComputational Linguistics. 2016 Jun;42(2):345-50
dc.relation.projectIDinfo:eu-repo/grantAgreement/EC/FP7/283554
dc.rights© MIT Press (Publisher version at http://mitpress.mit.edu) All articles are published under a CC BY-NC-ND 4.0 license. (https://creativecommons.org/licenses/by-nc-nd/4.0/)
dc.rights.accessRightsinfo:eu-repo/semantics/openAccess
dc.rights.urihttps://creativecommons.org/licenses/by-nc-nd/4.0/
dc.titleWhen the whole is less than the sum of its parts: how composition affects PMI values in distributional semantic vectorsen
dc.typeinfo:eu-repo/semantics/article
dc.type.versioninfo:eu-repo/semantics/publishedVersion

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
paperno_comlin_when.pdf
Size:
159.44 KB
Format:
Adobe Portable Document Format

License

Rights