Deanthropomorphising NLP: can a language model be conscious?

Mostra el registre complet Registre parcial de l'ítem

  • dc.contributor.author Shardlow, Matthew
  • dc.contributor.author Przybyła, Piotr
  • dc.date.accessioned 2025-06-27T07:15:44Z
  • dc.date.available 2025-06-27T07:15:44Z
  • dc.date.issued 2024
  • dc.description.abstract This work is intended as a voice in the discussion over previous claims that a pretrained large language model (LLM) based on the Transformer model architecture can be sentient. Such claims have been made concerning the LaMDA model and also concerning the current wave of LLM-powered chatbots, such as ChatGPT. This claim, if confirmed, would have serious ramifications in the Natural Language Processing (NLP) community due to wide-spread use of similar models. However, here we take the position that such a large language model cannot be conscious, and that LaMDA in particular exhibits no advances over other similar models that would qualify it. We justify this by analysing the Transformer architecture through Integrated Information Theory of consciousness. We see the claims of sentience as part of a wider tendency to use anthropomorphic language in NLP reporting. Regardless of the veracity of the claims, we consider this an opportune moment to take stock of progress in language modelling and consider the ethical implications of the task. In order to make this work helpful for readers outside the NLP community, we also present the necessary background in language modelling.
  • dc.description.sponsorship This work is part of the ERINIA project, which has received funding from the European Union’s Horizon Europe research and innovation programme under grant agreement No 101060930. Views and opinions expressed are however those of the author(s) only and do not necessarily reflect those of the European Union. Neither the European Union nor the granting authority can be held responsible for them. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
  • dc.format.mimetype application/pdf
  • dc.identifier.citation Shardlow M, Przybyła P. Deanthropomorphising NLP: can a language model be conscious? PLoS One. 2024;19(12):e0307521. DOI: 10.1371/journal.pone.0307521
  • dc.identifier.doi http://dx.doi.org/10.1371/journal.pone.0307521
  • dc.identifier.issn 1932-6203
  • dc.identifier.uri http://hdl.handle.net/10230/70776
  • dc.language.iso eng
  • dc.publisher Public Library of Science (PLoS)
  • dc.relation.ispartof PLoS One. 2024;19(12):e0307521
  • dc.relation.projectID info:eu-repo/grantAgreement/EC/HE/101060930
  • dc.rights © 2024 Shardlow, Przybyła. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
  • dc.rights.accessRights info:eu-repo/semantics/openAccess
  • dc.rights.uri http://creativecommons.org/licenses/by/4.0/
  • dc.subject.other Tractament del llenguatge natural (Informàtica)
  • dc.subject.other Lingüística computacional
  • dc.title Deanthropomorphising NLP: can a language model be conscious?
  • dc.type info:eu-repo/semantics/article
  • dc.type.version info:eu-repo/semantics/publishedVersion