Schedule

December 14, 2024 - West Meeting Room 111-112

Pre-registration form: https://forms.gle/YBCwn7L8N5AxExMG7

09:00 AM - 09:05 AM PST

Opening remarks

09:05 AM - 09:40 AM 

Hoda Heidari


Reflections on Fairness Measurement: From Predictive to Generative AI


Abstract: In this talk, I will provide an overview of the algorithmic fairness literature, which has historically focused on predictive AI models, designed to automate or assist high-stakes decisions. I will contrast that line of work with the recently growing set of benchmarks, metrics, and measures for bias and unfairness through Generative AI. I will conclude with some reflections on conceptualizing and measuring GenAI unfairness, drawing on my past work. 

Invited  Talk

09:40 AM - 09:45 AM PST  Short break

09:45 AM - 10:20 AM

Kush Varshney

Title: Harm Detectors and Guardian Models for LLMs: Implementations, Uses, and Limitations

Abstract: Large language models (LLMs) are susceptible to a variety of harms, from non-faithful output to biased and toxic generations. Due to several limiting factors surrounding LLMs (training cost, API access, data availability, etc.), it may not always be feasible to impose direct safety constraints on a deployed model. Therefore, an efficient and reliable alternative is required. To this end, we present our ongoing efforts to create and deploy harm detectors and guardian models: compact classification models that provide labels for various harms. In addition to the models themselves, we discuss a wide range of uses for these detectors and guardian models - from acting as guardrails to enabling effective AI governance. We also deep dive into inherent sociotechnical challenges in their development.

Invited  Talk

 10:20 a.m. - 10:25 a.m. Short break

10:25 a.m. - 10:30 a.m.

Alex Tamkin

Title: Evaluating and Mitigating Discrimination in Language Model Decisions 

Contributed  Talk

10:30 a.m. - 10:35 a.m

To Eun Kim

Title: Towards Fair RAG: On the Impact of Fair Ranking in Retrieval-Augmented Generation 

Contributed  Talk

10:35 a.m. - 11:20 a.m.

Poster Session 1

11:25 a.m. - 12:00 p.m.

Seth Lazar

Title: Evaluating the Ethical Competence of LLMs

Abstract: Existing approaches to evaluating LLM ethical competence place too much emphasis on the verdicts—of permissibility and impermissibility—that they render. But ethical competence doesn’t consist in one’s judgments conforming to those of a cohort of crowdworkers. It consists in being able to identify morally relevant features, prioritise among them, associate them with reasons and weave them into a justified conclusion. We identify the limitations of existing evals for ethical competence, provide an account of moral reasoning that can ground better alternatives, and discuss the practical—and philosophical—implications if LLMs ultimately do prove to be adept moral reasoners.

Invited  Talk

 12:00 p.m. - 12:10 p.m. Short break

12:10 p.m. - 1:00 p.m.

Roundtables

Discussions


Leads: Candace Ross (Meta AI), Tom Hartvigsen (University of Virginia)

Lead: Angelina Wang (Stanford)

1:00 p.m. - 2:00 p.m. Lunch break

2:00 p.m. - 2:35 p.m.

Sanmi Koyejo and Angelina Wang

Title: Fairness through Difference Awareness: Measuring Desired Group Discrimination in LLMs

Abstract: Algorithmic fairness has conventionally adopted a perspective of racial color-blindness (i.e., difference unaware treatment). We contend that in a range of important settings, group difference awareness matters. First, we present a taxonomy of such settings, such as in the legal system where it can be legally permissible to discriminate (e.g., Native Americans sometimes have privileged legal status, men enter the compulsory draft in America while women do not). Second, we present a benchmark suite that spans eight different settings for a total of 16k questions that enables us to assess for difference awareness. Third, we show that difference awareness is a distinct dimension of fairness and that existing bias mitigation strategies may backfire on this dimension.

Invited  Talk

2:35 p.m. - 2:40 p.m. Short break

2:40 p.m. - 2:45 p.m.

Prakhar Ganesh

Title: Different Horses for Different Courses: Comparing Bias Mitigation Algorithms in ML 

Contributed  Talk

2:45 p.m. - 2:50 p.m.

Natalie Mackraz

Title: Evaluating Gender Bias Transfer between Pre-trained and Prompt Adapted Language Models

Contributed  Talk

2:50 p.m. - 2:55 p.m.

Benjamin Laufer · Manish Raghavan · Solon Barocas

Title: The Search for Less Discriminatory Algorithms: Limits and Opportunities

Contributed  Talk

2:55 p.m. - 3:55 p.m.

Poster Session 2

3:55 p.m. - 4:15 p.m. Break

 4:15 p.m. - 4:55 p.m.

Panel: Rethinking fairness in the era of large language models

Hoda Heidari (CMU) · Sanmi Koyejo (Stanford University) · Jessica Schrouff (Google DeepMind) · Seth Lazar (Australian National University)

Discussions

4:55 p.m. - 5:00 p.m. Short break

5:00 p.m. - 5:10 p.m.

Closing remarks

5:10 p.m. - 5:30 p.m.

Poster Session 3