Exploring Fairness and Causality in Online Decision-Making

By: Contributor(s): Material type: BookBookPublication details: Bangalore : Indian Institute of Science, 2024.Description: xiii, 150p . : col. ill. e-Thesis 8.166MbDissertation: PhD;2024;Computer Science and AutomationSubject(s): DDC classification:
  • 006.31 PAT
Online resources: Dissertation note: PhD;2024;Computer Science and Automation Summary: Online decision-making under uncertainty is a fundamental aspect of numerous real-world problems across various domains, including online resource allocation, crowd-sourcing, and online advertising. Multi-Armed Bandits (MABs) and Markov Decision Processes (MDPs) are two popular modeling frameworks for capturing decision-making under uncertainty. The inherent nature of applications modeled by frameworks like MABs and MDPs often requires additional considerations and adaptations to effectively address real-world challenges. In this thesis, our primary emphasis is on two specific factors: integrating fairness considerations into the model and leveraging causal relations among different variables in the model to make better decisions. The thesis comprises three contributions: First, we commence with an exploration of fairness within temporally extended decision-making scenarios, specifically those modeled as MDPs. Our novel fairness notion aims to guarantee that each state's long-term visitation frequency surpasses a predefined fraction - a natural extension of quota-based fairness from MAB literature. We propose an algorithm with a dual guarantee: simultaneously satisfying fairness and maximizing the total reward. Second, we shift our focus to a variant of the MAB model that accounts for the dynamic nature of the environment. This model, where arm rewards increase with each pull, is a versatile abstraction for real-world scenarios, particularly in education and employment domains where opportunity allocation impacts community capabilities. We present an algorithm that maximizes the total reward while ensuring that the arms, which may correspond to communities, attain their fullest potential. Third, we study the problem of learning good interventions in causal graphs by modeling it as an MAB problem. This problem called the Causal Multi-Armed Bandit (Causal MAB) problem, captures dependencies between arms through a causal graph. We study the problem of identifying the best intervention in Causal MAB and provide algorithms for three variants of the Causal MAB problem.
Tags from this library: No tags from this library for this title. Log in to add tags.
Star ratings
    Average rating: 0.0 (0 votes)
Holdings
Item type Current library Call number URL Status Date due Barcode
Thesis Thesis JRD Tata Memorial Library 006.31 PAT (Browse shelf(Opens below)) Link to resource Available ET00526

Includes bibliographical references

PhD;2024;Computer Science and Automation

Online decision-making under uncertainty is a fundamental aspect of numerous real-world problems across various domains, including online resource allocation, crowd-sourcing, and online advertising. Multi-Armed Bandits (MABs) and Markov Decision Processes (MDPs) are two popular modeling frameworks for capturing decision-making under uncertainty. The inherent nature of applications modeled by frameworks like MABs and MDPs often requires additional considerations and adaptations to effectively address real-world challenges. In this thesis, our primary emphasis is on two specific factors: integrating fairness considerations into the model and leveraging causal relations among different variables in the model to make better decisions. The thesis comprises three contributions: First, we commence with an exploration of fairness within temporally extended decision-making scenarios, specifically those modeled as MDPs. Our novel fairness notion aims to guarantee that each state's long-term visitation frequency surpasses a predefined fraction - a natural extension of quota-based fairness from MAB literature. We propose an algorithm with a dual guarantee: simultaneously satisfying fairness and maximizing the total reward. Second, we shift our focus to a variant of the MAB model that accounts for the dynamic nature of the environment. This model, where arm rewards increase with each pull, is a versatile abstraction for real-world scenarios, particularly in education and employment domains where opportunity allocation impacts community capabilities. We present an algorithm that maximizes the total reward while ensuring that the arms, which may correspond to communities, attain their fullest potential. Third, we study the problem of learning good interventions in causal graphs by modeling it as an MAB problem. This problem called the Causal Multi-Armed Bandit (Causal MAB) problem, captures dependencies between arms through a causal graph. We study the problem of identifying the best intervention in Causal MAB and provide algorithms for three variants of the Causal MAB problem.

There are no comments on this title.

to post a comment.

                                                                                                                                                                                                    Facebook    Twitter

                             Copyright © 2023. J.R.D. Tata Memorial Library, Indian Institute of Science, Bengaluru - 560012

                             Contact   Phone: +91 80 2293 2832

Powered by Koha