Young Data Science Researcher Seminar Zurich

×

Modal title

Modal content

Pleasesubscribe hereif you would you like to be notified about these presentations via e-mail. Moreover you cansubscribeto the iCal/ics Calender.

Autumn Semester 2023

Date / Time Speaker Title Location
3 October 2023
15:30-17:00
Linjun Zhang
Rutgers, University
Mikhail Yurochkin
IBM Research and MIT-IBM Watson AI Lab
Event Details

Young Data Science Researcher Seminar Zurich

Title Joint webinar of the IMS New Researchers Group, Young Data Science Researcher Seminar Zürich, and the YoungStatS Project: Algorithmic Fairness
Speaker, Affiliation Linjun Zhang,Rutgers, University
Mikhail Yurochkin,IBM Research and MIT-IBM Watson AI Lab
Date, Time 3 October 2023, 15:30-17:00
Location Zoom Call
Abstract 1) Fair conformal prediction, Linjun Zhang, Rutgers University
Multi-calibration is a powerful and evolving concept originating in the field of algorithmic fairness. For a predictor $f$ that estimates the outcome y given covariates $x$, and for a function class $C$, multi-calibration requires that the predictor $f(x)$ and outcome y are indistinguishable under the class of auditors in $C$. Fairness is captured by incorporating demographic subgroups into the class of functions $C$. Recent work has shown that, by enriching the class $C$ to incorporate appropriate propensity re-weighting functions, multi-calibration also yields target-independent learning, wherein a model trained on a source domain performs well on unseen, future target domains(approximately) captured by the re-weightings. The multi-calibration notion is extended, and the power of an enriched class of mappings is explored. HappyMap, a generalization of multi-calibration, is proposed, which yields a wide range of new applications, including a new fairness notion for uncertainty quantification (conformal prediction), a novel technique for conformal prediction under covariate shift, and a different approach to analyzing missing data, while also yielding a unified understanding of several existing seemingly disparate algorithmic fairness notions and target-independent learning approaches. A single HappyMap meta-algorithm is given that captures all these results, together with a sufficiency condition for its success.

2) Operationalizing Individual Fairness, Mikhail Yurochkin, IBM Research and MIT-IBM Watson AI Lab
Societal applications of ML proved to be challenging due to algorithms replicating or even exacerbating biases in the training data. In response, there is a growing body of research on algorithmic fairness that attempts to address these issues, primarily via group definitions of fairness. In this talk, I will illustrate several shortcomings of group fairness and present an algorithmic fairness pipeline based on individual fairness (IF). IF is often recognized as the more intuitive notion of fairness: we want ML models to treat similar individuals similarly. Despite the benefits, challenges in formalizing the notion of similarity and enforcing equitable treatment prevented the adoption of IF. I will present our work addressing these barriers via algorithms for learning the similarity metric from data and methods for auditing and training fair models utilizing the intriguing connection between individual fairness and adversarial robustness. Finally, I will demonstrate applications of IF with Large Language Models.

Discussant:Razieh Nabi, Emory University
Assets Registrationfile_download
Joint webinar of the IMS New Researchers Group, Young Data Science Researcher Seminar Zürich, and the YoungStatS Project: Algorithmic Fairnessread_more
Zoom Call

Note: thehighlighted eventmarks the next occurring event.

JavaScript has been disabled in your browser