Show simple item record

dc.contributor.advisorTalukdar, Partha Pratim
dc.contributor.advisorMitchell, Tom
dc.contributor.authorJat, Sharmistha
dc.date.accessioned2022-11-03T06:34:27Z
dc.date.available2022-11-03T06:34:27Z
dc.date.submitted2022
dc.identifier.urihttps://etd.iisc.ac.in/handle/2005/5892
dc.description.abstractDeep Neural Networks (DNN) inspired by the human brain have redefined the state-of-the-art performance in AI during the past decade. Much of the research is still trying to understand and explain the function of these networks. In this thesis, we leverage knowledge from the neuroscience literature to evaluate the representations learned in state-of-the-art language models. We use sentences with simple syntax and semantics (e.g., “The bone was eaten by the dog.”), and train multiple neural networks to predict the part of speech, next word. We present other sentences of this same simple form, word-by-word to humans in a magnetoencephalography (MEG) scanner for silent reading and comprehension. We then train a linear regression model to predict observed brain recording from the hidden layers of the trained neural networks and popular pre-trained networks like BERT and ELMo. We find that the middle layers of these networks are the most predictive of the recorded brain activity. But, a more fine-grained evaluation shows that various types of stimuli (determiner, adjective, noun, verb) are represented more dominantly in different layers of the language model. Further, we test the semantic composition capabilities of these networks with respect to the human brain. Semantic composition is defined as the rule-based combination of the parts that constitutes the meaning of the whole. We collect new data and develop a new framework to perform this evaluation incrementally as each word in the sentence is processed in the brain and DNN. As a result, we are able to analyze the effect of the composition function in representing the same word as more of the sentence context becomes available. Our experiments show that DNN models are effective in encoding the sentence being read and are able to predict the word which occurred earlier in the sentence, indicating good composition. We find that in these tests, the right frontal and right temporal brain regions are predicted with best accuracy. Previous research has suggested that these brain regions are responsible for executive and memory function. As an additional contribution, we propose a new dynamic time warping based distance metric to evaluate alignment between the predicted brain activity versus the observed brain activity. The new metric helps tackle the variability observed in a single subject’s recorded brain activity.en_US
dc.description.sponsorshipMinistry of Human Resource and development India (MHRD), Pratiksha Trust, and CMU BrainHuben_US
dc.language.isoen_USen_US
dc.rightsI grant Indian Institute of Science the right to archive and to make available my thesis or dissertation in whole or in part in all forms of media, now hereafter known. I retain all proprietary rights, such as patent rights. I also retain the right to use in future works (such as articles or books) all or part of this thesis or dissertationen_US
dc.subjectComputational Neuroscienceen_US
dc.subjectMEGen_US
dc.subjectNatural Language Processing,en_US
dc.subjectMachine Learningen_US
dc.subjectDeep Neural Networksen_US
dc.subjectNeuroscienceen_US
dc.subject.classificationResearch Subject Categories::INTERDISCIPLINARY RESEARCH AREASen_US
dc.titleRelating Representations in Deep Learning and the Brainen_US
dc.typeThesisen_US
dc.degree.namePhDen_US
dc.degree.levelDoctoralen_US
dc.degree.grantorIndian Institute of Scienceen_US
dc.degree.disciplineEngineeringen_US


Files in this item

This item appears in the following Collection(s)

Show simple item record