Show simple item record

dc.contributor.advisorRamakrishnan, K R
dc.contributor.advisorSastry, P S
dc.contributor.authorKirthi, Suresh K
dc.date.accessioned2021-10-20T09:22:26Z
dc.date.available2021-10-20T09:22:26Z
dc.date.submitted2019
dc.identifier.urihttps://etd.iisc.ac.in/handle/2005/5444
dc.description.abstractThe convolutional neural networks (CNNs) have become the most successful models for many pattern recognition problems in the areas of computer vision, speech, text and others. One concern about CNNs has always been their need for large amount of training data, large computational re- sources and long training time. In this regard the transfer learning is a technique that can address this concern of inefficient CNN training through reuse of pretrained networks (CNNs). In this thesis we discuss transfer learning in CNNs where the transfer is from multiple source CNNs and done at subnetwork levels. The subnetwork multisource transfer is attempted for the fi rst time and hence we begin by showing the effectiveness of such a transfer. We consider subnetworks at various granularities for the transfer. These granularities begin at a whole network-level then pro-ceed to layer-level and further fi lter-level. In order to realize this kind of transfer we create a set called bank of weight fi lters (BWF) which is a repository of the pretrained subnetworks that are used as candidates for transfer. Through extensive simulations we show that subnetwork level transfer, implemented through random selection from a BWF, is elective and is also efficient in terms of training time. We also present experimental results to show that subnetwork level transfer learning is efficient in terms of the amount of training data needed. It is seen that fi lter-level transfer learning is as effective as the whole-network-level transfer which is the conventional transfer learning used with CNNs. We then show the usefulness of the fi lter-level multisource transfer for the cases of transfer from natural to non-natural (hand drawn sketches) image datasets and transfer across different CNN architectures (having different number of layers, fi lter dimensions etc.). We also discuss transfer from CNNs trained on high-resolution images to the CNNs needed for the low-resolution im- ages and vice-versa. In the multisource transfer of prelearnt weights discussed above, the transferred weights have to be fi netuned to achieve the same accuracy as that of a CNN trained from scratch. It is certainly more bene cfiial and efficient if the fi netuning of transferred weights can be completely avoided. For this, we conceptualize we conceptualize what we call a fi lter-tree which represents the complete feature generation entity that is learnt by a CNN and propose that the a filter-tree represents a subnetwork that can be used for transfer without finetuning. Similar to BWF we create a repository of pre-learnt fllter-trees called bank of filter-trees (BFT) to realize the transfer using fi lter-trees. Through experiments we show that transfer using BFT (where the transferred weights are held fixed and are not fi netunes) has performance that is on par with training from scratch, which is the best achievable performance. The selection of the subnetworks from BWF or BFT so far for all experiments was done uniformly randomly. For the sake of completion we introduce a method that can result in informed choice of fi lters from a BFT. We propose a learnable auxilliary layer called choice layer whose learnt weights give an idea of the importance/utility of different the subnetwork (fi lter-trees here) in the BFT for the target task. We show that when the random choice from BFT does not achieve the best possible accuracy, the choice layer based method can achieve it.en_US
dc.language.isoen_USen_US
dc.relation.ispartofseries;G29281
dc.rightsI grant Indian Institute of Science the right to archive and to make available my thesis or dissertation in whole or in part in all forms of media, now hereafter known. I retain all proprietary rights, such as patent rights. I also retain the right to use in future works (such as articles or books) all or part of this thesis or dissertationen_US
dc.subjectFilter-treesen_US
dc.subjectconvolutional neural networksen_US
dc.subjectbank of weight filtersen_US
dc.subjectbank of filter treesen_US
dc.subject.classificationResearch Subject Categories::TECHNOLOGY::Electrical engineering, electronics and photonics::Electrical engineeringen_US
dc.titleMultisource Subnetwork Level Transfer in Deep CNNs Using Bank of Weight Filtersen_US
dc.typeThesisen_US
dc.degree.namePhDen_US
dc.degree.levelDoctoralen_US
dc.degree.grantorIndian Institute of Scienceen_US
dc.degree.disciplineEngineeringen_US


Files in this item

This item appears in the following Collection(s)

Show simple item record