Show simple item record

dc.contributor.advisorKatewa, Vaibhav
dc.contributor.authorRahul, N R
dc.date.accessioned2025-02-08T05:19:39Z
dc.date.available2025-02-08T05:19:39Z
dc.date.submitted2024
dc.identifier.urihttps://etd.iisc.ac.in/handle/2005/6803
dc.description.abstractWe consider a sequential multi-task problem, where each task is modeled as a stochastic multi-armed bandit with K arms. We study the problem of transfer learning in this setting and propose algorithms based on UCB to transfer reward samples from previous tasks to improve the total regret across all tasks. We consider two different notions of similarity among tasks, (i) universal similarity and (ii) adjacent similarity. In universal similarity, all tasks encountered in the sequence are similar. On the other hand, in adjacent similarity, tasks close to one another in the sequence are more similar than the ones that are farther apart. We provide transfer algorithms and their regret upper bounds for both the similarity notions and then highlight the benefit of transfer. Our regret bounds show that the performance improves as the sequential tasks become closer to each other. Finally, we provide empirical results for our algorithms, which show performance improvement over the standard UCB algorithm without transfer.en_US
dc.language.isoen_USen_US
dc.relation.ispartofseries;ET00813
dc.rightsI grant Indian Institute of Science the right to archive and to make available my thesis or dissertation in whole or in part in all forms of media, now hereafter known. I retain all proprietary rights, such as patent rights. I also retain the right to use in future works (such as articles or books) all or part of this thesis or dissertationen_US
dc.subjectmulti-armed banditen_US
dc.subjectalgorithmsen_US
dc.subjectuniversal similarityen_US
dc.subjectadjacent similarityen_US
dc.subjectUCB algorithmen_US
dc.subject.classificationResearch Subject Categories::TECHNOLOGY::Electrical engineering, electronics and photonics::Electronicsen_US
dc.titleSequential Transfer in Multi-Armed Bandits using Reward Samplesen_US
dc.typeThesisen_US
dc.degree.nameMTech (Res)en_US
dc.degree.levelMastersen_US
dc.degree.grantorIndian Institute of Scienceen_US
dc.degree.disciplineEngineeringen_US


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record