In a series of three talks, we present an overview of work in bandit optimization, including stochastic and adversarial bandits, and will focus on variants of multi-armed bandits where fairness is an additional constraint. In Fair-MAB, for example, in addition to the objective of maximizing the sum of expected rewards, the algorithm also needs to ensure that at any time, each arm is pulled at least a pre-specified fraction of times. We investigate the interplay between learning and fairness in such settings, and also investigate the cost of fairness.
Talk 1: Thursday 29/8, 2 - 3:30 pm in A-238
Talk 2: Friday 30/8, 4 - 5:30 pm in A-201
Talk 3: Monday 2/9, 2 - 3:30 pm in A-238
Short Bio:
Vishakha Patil works in online learning and optimization, game theory, and mechanism design. She completed her PhD from IISc, supported by a Google Fellowship and a CII-SERB PMRF. She has also been a Heidelberg Laureate Forum Young Researcher, and her MTech thesis received an Honourable Mention for the Best Thesis Award at CSA, IISc.