MPI-INF Logo
Campus Event Calendar

Event Entry

What and Who

Reliable Measurement for Machine Learning at Scale

A. Feder Cooper
Cornell University
CIS@MPG Colloquium

A. Feder Cooper is a scalable machine-learning (ML) researcher, working on reliable measurement and evaluation of ML. Cooper’s research develops nuanced quality metrics for ML behaviors, and makes sure that we can effectively measure these metrics at scale and in practice. Cooper’s contributions span distributed training, hyperparameter optimization, uncertainty estimation, model selection, and generative AI. To make sure that our evaluation metrics can meaningfully measure our goals for ML, Cooper also leads research in tech policy and law, and spends a lot of time working to effectively communicate the capabilities and limits of AI/ML to the broader public. Cooper is a CS Ph.D. candidate at Cornell University, an Affiliate at the Berkman Klein Center for Internet & Society at Harvard University, co-founder of The Center for Generative AI, Law, and Policy Research (The GenLaw Center), and a student researcher at Google Research. Cooper has received many spotlight and oral awards at top conferences, including NeurIPS, AAAI, and AIES, and was named a "Rising Star in EECS" by MIT.
AG 1, AG 2, AG 3, INET, AG 4, AG 5, D6, SWS, RG1, MMCI  
AG Audience
English

Date, Time and Location

Thursday, 8 February 2024
10:00
60 Minutes
MPI-SP
-
Bochum

Abstract

We need reliable measurement in order to develop rigorous knowledge about ML models and the systems in which they are embedded. But reliable measurement is a really hard problem, touching on issues of reproducibility, scalability, uncertainty quantification, epistemology, and more. In this talk, I will discuss the criteria needed to take reliability seriously — criteria for designing meaningful metrics, and for methodologies that ensure that we can dependably and efficiently measure these metrics at scale and in practice. I will give two examples of my research that put these criteria into practice: (1) large-scale evaluation of training-data memorization in large language models, and (2) evaluating latent arbitrariness in algorithmic fairness binary classification contexts. Throughout this discussion, I will emphasize how important it is to make metrics understandable for other stakeholders in order to facilitate public governance. For this reason, my work aims to design metrics that are legally cognizable — a goal that frames the both my ML and legal scholarship. I will draw on important connections that I have uncovered between ML and law: connections between (1) the generative-AI supply chain and US copyright law, and (2) ML arbitrariness and arbitrariness in legal rules. This talk reflects joint work with collaborators at The GenLaw Center, Cornell CS, Cornell Law School, Google DeepMind, and Microsoft Research.

Contact

Annika Meiser
+49 681 9303 9105
--email hidden

Virtual Meeting Details

Zoom
668 9092 0474
passcode not visible
logged in users only

Lena Schneider, 01/23/2024 15:54
Annika Meiser, 01/22/2024 10:38
Annika Meiser, 01/19/2024 12:06 -- Created document.