dc.contributor.advisor | Talukdar, Partha Pratim | |
dc.contributor.advisor | Mitchell, Tom | |
dc.contributor.author | Jat, Sharmistha | |
dc.date.accessioned | 2022-11-03T06:34:27Z | |
dc.date.available | 2022-11-03T06:34:27Z | |
dc.date.submitted | 2022 | |
dc.identifier.uri | https://etd.iisc.ac.in/handle/2005/5892 | |
dc.description.abstract | Deep Neural Networks (DNN) inspired by the human brain have redefined the state-of-the-art performance in AI during the past decade. Much of the research is still trying to understand and explain the function of these networks. In this thesis, we leverage knowledge from the neuroscience literature to evaluate the representations learned in state-of-the-art language models. We use sentences with simple syntax and semantics (e.g., “The bone was eaten by the dog.”), and train multiple neural networks to predict the part of speech, next word. We present other sentences of this same simple form, word-by-word to humans in a magnetoencephalography (MEG) scanner for silent reading and comprehension. We then train a linear regression model to predict observed brain recording from the hidden layers of the trained neural networks and popular pre-trained networks like BERT and ELMo.
We find that the middle layers of these networks are the most predictive of the recorded brain activity. But, a more fine-grained evaluation shows that various types of stimuli (determiner, adjective, noun, verb) are represented more dominantly in different layers of the language model. Further, we test the semantic composition capabilities of these networks with respect to the human brain. Semantic composition is defined as the rule-based combination of the parts that constitutes the meaning of the whole. We collect new data and develop a new framework to perform this evaluation incrementally as each word in the sentence is processed in the brain and DNN. As a result, we are able to analyze the effect of the composition function in representing the same word as more of the sentence context becomes available. Our experiments show that DNN models are effective in encoding the sentence being read and are able to predict the word which occurred earlier in the sentence, indicating good composition. We find that in these tests, the right frontal and right temporal brain regions are predicted with best accuracy. Previous research has suggested that these brain regions are responsible for executive and memory function.
As an additional contribution, we propose a new dynamic time warping based distance metric to evaluate alignment between the predicted brain activity versus the observed brain activity. The new metric helps tackle the variability observed in a single subject’s recorded brain activity. | en_US |
dc.description.sponsorship | Ministry of Human Resource and development India (MHRD), Pratiksha Trust, and CMU BrainHub | en_US |
dc.language.iso | en_US | en_US |
dc.rights | I grant Indian Institute of Science the right to archive and to make available my thesis or dissertation in whole or in part in all forms of media, now hereafter known. I retain all proprietary rights, such as patent rights. I also retain the right to use in future works (such as articles or books) all or part
of this thesis or dissertation | en_US |
dc.subject | Computational Neuroscience | en_US |
dc.subject | MEG | en_US |
dc.subject | Natural Language Processing, | en_US |
dc.subject | Machine Learning | en_US |
dc.subject | Deep Neural Networks | en_US |
dc.subject | Neuroscience | en_US |
dc.subject.classification | Research Subject Categories::INTERDISCIPLINARY RESEARCH AREAS | en_US |
dc.title | Relating Representations in Deep Learning and the Brain | en_US |
dc.type | Thesis | en_US |
dc.degree.name | PhD | en_US |
dc.degree.level | Doctoral | en_US |
dc.degree.grantor | Indian Institute of Science | en_US |
dc.degree.discipline | Engineering | en_US |