Usage
  • 105 views
  • 125 downloads

How Brain-Like is an LSTM’s Representation of Nonsensical Language Stimuli?

  • Author / Creator
    Hashemzadeh, Maryam
  • The representations generated by many models of language (word embeddings, recurrent neural networks and transformers) correlate to brain activity recorded while people listen. However, these decoding results are usually based on the brain’s reaction to syntactically and semantically sound language stimuli. In this study, we asked: how does an LSTM (long short term memory) language model, trained (by and large) on semantically and syntactically intact language, represent a language sample with degraded semantic or syntactic information? Does the LSTM representation still resemble the brain’s reaction? We found that, even for some kinds of nonsensical language, there is a statistically significant relationship between the brain’s activity and the representations of an LSTM. More exceptional, a character-based LSTM’s representation of pseudoword sentences is significantly correlated to EEG collected while people listened to those sentences - even though the pseudowords were not in the LSTM training data. This indicates that, at least in some instances, LSTMs and the human brain handle nonsensical data similarly.

  • Subjects / Keywords
  • Graduation date
    Fall 2021
  • Type of Item
    Thesis
  • Degree
    Master of Science
  • DOI
    https://doi.org/10.7939/r3-wz3m-sm87
  • License
    This thesis is made available by the University of Alberta Libraries with permission of the copyright owner solely for non-commercial purposes. This thesis, or any portion thereof, may not otherwise be copied or reproduced without the written consent of the copyright owner, except to the extent permitted by Canadian copyright law.