Achieving User-Side Fairness in Contextual Bandits

  • Wen Huang
  • , Kevin Labille
  • , Xintao Wu
  • , Dongwon Lee
  • , Neil Heffernan

Research output: Contribution to journalArticlepeer-review

9 Scopus citations

Abstract

Personalized recommendation based on multi-arm bandit (MAB) algorithms has shown to lead to high utility and efficiency as it can dynamically adapt the recommendation strategy based on feedback. However, unfairness could incur in personalized recommendation. In this paper, we study how to achieve user-side fairness in personalized recommendation. We formulate our fair personalized recommendation as a modified contextual bandit and focus on achieving fairness on the individual whom is being recommended an item as opposed to achieving fairness on the items that are being recommended. We introduce and define a metric that captures the fairness in terms of rewards received for both the privileged and protected groups. We develop a fair contextual bandit algorithm, Fair-LinUCB, that improves upon the traditional LinUCB algorithm to achieve group-level fairness of users. Our algorithm detects and monitors unfairness while it learns to recommend personalized videos to students to achieve high efficiency. We provide a theoretical regret analysis and show that our algorithm has a slightly higher regret bound than LinUCB. We conduct numerous experimental evaluations to compare the performances of our fair contextual bandit to that of LinUCB and show that our approach achieves group-level fairness while maintaining a high utility.

Original languageEnglish (US)
Pages (from-to)81-94
Number of pages14
JournalHuman-Centric Intelligent Systems
Volume2
Issue number3
DOIs
StatePublished - Dec 2022

All Science Journal Classification (ASJC) codes

  • Information Systems
  • Artificial Intelligence
  • Computer Science (miscellaneous)

Fingerprint

Dive into the research topics of 'Achieving User-Side Fairness in Contextual Bandits'. Together they form a unique fingerprint.

Cite this