Show simple item record

dc.contributor.advisorNarahari, Y
dc.contributor.advisorKhan, Arindam
dc.contributor.authorPatil, Vishakha
dc.date.accessioned2024-05-22T04:26:51Z
dc.date.available2024-05-22T04:26:51Z
dc.date.submitted2024
dc.identifier.urihttps://etd.iisc.ac.in/handle/2005/6517
dc.description.abstractOnline decision-making under uncertainty is a fundamental aspect of numerous real-world problems across various domains, including online resource allocation, crowd-sourcing, and online advertising. Multi-Armed Bandits (MABs) and Markov Decision Processes (MDPs) are two popular modeling frameworks for capturing decision-making under uncertainty. The inherent nature of applications modeled by frameworks like MABs and MDPs often requires additional considerations and adaptations to effectively address real-world challenges. In this thesis, our primary emphasis is on two specific factors: integrating fairness considerations into the model and leveraging causal relations among different variables in the model to make better decisions. The thesis comprises three contributions: First, we commence with an exploration of fairness within temporally extended decision-making scenarios, specifically those modeled as MDPs. Our novel fairness notion aims to guarantee that each state's long-term visitation frequency surpasses a predefined fraction - a natural extension of quota-based fairness from MAB literature. We propose an algorithm with a dual guarantee: simultaneously satisfying fairness and maximizing the total reward. Second, we shift our focus to a variant of the MAB model that accounts for the dynamic nature of the environment. This model, where arm rewards increase with each pull, is a versatile abstraction for real-world scenarios, particularly in education and employment domains where opportunity allocation impacts community capabilities. We present an algorithm that maximizes the total reward while ensuring that the arms, which may correspond to communities, attain their fullest potential. Third, we study the problem of learning good interventions in causal graphs by modeling it as an MAB problem. This problem called the Causal Multi-Armed Bandit (Causal MAB) problem, captures dependencies between arms through a causal graph. We study the problem of identifying the best intervention in Causal MAB and provide algorithms for three variants of the Causal MAB problem.en_US
dc.description.sponsorshipGoogle PhD Fellowship, CII-SERB PM Fellowship for Doctoral Researchen_US
dc.language.isoen_USen_US
dc.relation.ispartofseries;ET00526
dc.rightsI grant Indian Institute of Science the right to archive and to make available my thesis or dissertation in whole or in part in all forms of media, now hereafter known. I retain all proprietary rights, such as patent rights. I also retain the right to use in future works (such as articles or books) all or part of this thesis or dissertationen_US
dc.subjectMachine Learningen_US
dc.subjectOnline Learningen_US
dc.subjectCausal Inferenceen_US
dc.subjectFairnessen_US
dc.subjectMulti-Armed Banditsen_US
dc.subjectMarkov Decision Processesen_US
dc.subject.classificationResearch Subject Categories::TECHNOLOGY::Information technology::Computer scienceen_US
dc.titleExploring Fairness and Causality in Online Decision-Makingen_US
dc.typeThesisen_US
dc.degree.namePhDen_US
dc.degree.levelDoctoralen_US
dc.degree.grantorIndian Institute of Scienceen_US
dc.degree.disciplineEngineeringen_US


Files in this item

This item appears in the following Collection(s)

Show simple item record