Role of punctuation in semantic mapping between brain and transformer models

Zenon Lamprou, Frank Pollick, Yashar Moshfeghi*

*Corresponding author for this work

Research output: Chapter in Book/Report/Conference proceedingConference contribution book

2 Citations (Scopus)
5 Downloads (Pure)

Abstract

Modern neural networks specialised in natural language processing (NLP) are not implemented with any explicit rules regarding language. It has been hypothesised that they might learn something generic about language. Because of this property much research has been conducted on interpreting their inner representations. A novel approach has utilised an experimental procedure that uses human brain recordings to investigate if a mapping from brain to neural network representations can be learned. Since this novel approach has been introduced, more advanced models in NLP have been introduced. In this research we are using this novel approach to test four new NLP models to try and find the most brain aligned model. Moreover, in our effort to unravel important information on how the brain processes text semantically, we modify the text in the hope of getting a better mapping out of the models. We remove punctuation using four different scenarios to determine the effect of punctuation on semantic understanding by the human brain. Our results show that the RoBERTa model is most brain aligned. RoBERTa achieves a higher accuracy score on our evaluation than BERT. Our results also show for BERT that when punctuation was removed a higher accuracy was achieved and that as the context length increased the accuracy did not decrease as much as the original results that include punctuation.

Original languageEnglish
Title of host publicationMachine Learning, Optimization, and Data Science - 8th International Conference, LOD 2022, Revised Selected Papers
EditorsGiuseppe Nicosia, Giovanni Giuffrida, Varun Ojha, Emanuele La Malfa, Gabriele La Malfa, Panos Pardalos, Giuseppe Di Fatta, Renato Umeton
Place of PublicationCham, Switzerland
PublisherSpringer Science and Business Media Deutschland GmbH
Pages458-472
Number of pages15
ISBN (Electronic)9783031258916
ISBN (Print)9783031258909
DOIs
Publication statusPublished - 10 Mar 2023
Event8th International Conference on Machine Learning, Optimization, and Data Science, LOD 2022, held in conjunction with the 2nd Advanced Course and Symposium on Artificial Intelligence and Neuroscience, ACAIN 2022 - Certosa di Pontignano, Italy
Duration: 18 Sept 202222 Sept 2022

Publication series

NameLecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Volume13811 LNCS
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349

Conference

Conference8th International Conference on Machine Learning, Optimization, and Data Science, LOD 2022, held in conjunction with the 2nd Advanced Course and Symposium on Artificial Intelligence and Neuroscience, ACAIN 2022
Country/TerritoryItaly
CityCertosa di Pontignano
Period18/09/2222/09/22

Keywords

  • fMRI
  • transformers
  • explainable AI
  • punctuation symbols

Fingerprint

Dive into the research topics of 'Role of punctuation in semantic mapping between brain and transformer models'. Together they form a unique fingerprint.

Cite this