Hi! I'm Aspen.

I'm a PhD student at CSAIL in the Madry Lab and the Center for Deployable Machine Learning. I study how machine learning systems are built and deployed.

AI Supply Chains as Systems

I study how chaining or otherwise composing models & data through "AI supply chains" introduces new challenges to AI development + how we can build tools to help.

Interpretability through Representations

I use internal model representations (like weights or activations) as semantic interfaces for understanding and steering model behavior.

Human‑Centered Tooling

I work on interactive tools, visual interfaces, and policy-relevant artifacts that help people make sense of complex data and computational systems.

research

* * * ai supply chains as systems * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

selected model updates

when should model updates propegate?

A new algorithmic challenge for developers has surfaced: deciding when to adopt model updates. Currently, the only reliable approach is to retrain on new upstream models and then test extensively. We introduce a method for assessing when to propagate newly released model versions to downstream applications that does not require access to upstream data or retraining a priori.
w/ Isabella Struckman, Hedi Dress, & Aleksander Madry
selected AI supply chains overview

ai supply chains

AI systems are increasingly built and deployed through AI supply chains (AISCs): complex networks of AI components "glued together" by researchers and developers. These supply chains challenge many of the basic expectations we have about ML development and deployment. We provide a formal model of AI supply chains as directed graphs to study how structure and complexity affect fairness and explainability.
w/ Sarah H. Cen, Andrew Illyas, Isabella Struckman, Aleksander Madry, & Luis Videgaray
selected Markets and redress

ai supply chains, markets, & redress

AI supply chains are market-structured systems composed of organizations with defined roles, incentives, and contracts. Our work examines how outsourcing, integration, and power dynamics shape avenues of redress for all parties when AI failures occur.
w/ Isabella Struckman, Kevin Klyman, & Susan S. Silbey
selected On AI deployment

on ai deployment

Our series On AI Deployment discusses the economic and regulatory implications of AI supply chains.

* * * ml interpretability * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

selected Emergent world representations

emergent world representations

Do complex LMs memorize surface statistics, or develop internal representations of underlying processes? Using a synthetic board game (Othello), we uncover nonlinear internal representations of board state and show, via interventions, that these are causal. We also create latent saliency maps explaining outputs.
w/Kenneth Li, David Bau, Fernanda Viegas, Hanspeter Pfister, & Martin Wattenberg

* * * responsible ai* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

selected Designing data for ML

designing data for ml

The ML pipeline includes data collection and iteration. What data should you collect, how should you collect it, and how do you evaluate what a model has learned prior to deployment?
w/Fred Hohman, Luca Zappella, Xavier Suau Cuadros, & Dominik Moritz
selected cosine distances between model responses

how models evolve over the 2024 election cycle

Large Scale, Longitudinal Study of Large Language Models During the 2024 US Election Season.
w/Sarah H. Cen, Andrew Ilyas, Hedi Driss, Charlotte Park, Aspen Hopkins, Chara Podimata, Aleksander Madry
selected Sampling with LLMs

sampling with llms

People have begun using large language models (LLMs) to induce sample distributions for synthetic data, but there are no guarantees about the resulting distribution. We evaluate LLMs as distribution samplers across modalities and find they struggle to produce a reasonable distribution.
w/ Alex Renda & Michael Carbin
selected Open foundation models

open foundation models

What are the benefits of open models? What are the risks? Led by Sayash Kapoor and Rishi Bommasani, this work collects the thoughts of 25 authors to start answering these questions.
w/ Sayash Kapoor, Rishi Bommasani*, Kevin Klyman, Shayne Longpre, Ashwin Ramaswami, Peter Cihon, Kevin Bankston, Stella Biderman, Miranda Bogen, Rumman Chowdhury, Alex Engler, Peter Henderson, Yacine Jernite, Seth Lazar, Stefano Maffulli, Alondra Nelson, Joelle Pineau, Aviya Skowron, Dawn Song, Victor Storchan, Daniel Zhang, Daniel E. Ho, Percy Liang & Arvind Narayanan
selected ML practices outside big tech

ml practices outside big tech

Support for the democratization of ML is growing rapidly, but responsible ML development outside Big Tech is poorly understood. As more organizations turn to ML, what challenges do they face in creating fair and ethical ML? We explore these challenges and outline research directions in an AIES spotlight paper.
w/ Serena Booth
selected Uncertainty in ML systems

uncertainty in ml systems

ML systems result from complex sociotechnical processes, each introducing distinct forms of uncertainty. Communicating this uncertainty is critical for appropriate trust, but cumulative encodings often obscure its complexity. We explore how and what uncertainty to present to different stakeholders.
w/Harini Suresh*
selected Socializing data

socializing data

Labeled datasets are often treated as authoritative ground truth. How is that ground truth determined, and how can we build historical contexts for these systems? This project focuses on collaborative sensemaking and label provenance.

* * * visualizations * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *

selected Misleading visualizations

misleading visualizations

Misinformation comes in many forms—including charts and graphs. We built a spell-check equivalent for visualizations to encourage best practices, improve data literacy, and foster critique in public domains.
w/Michael Correll & Arvind Satyanarayan
selected Visualizations for the public

visualizations for the public

Air quality, like many environmental and health considerations, is important to communicate to the public. How do you effectively communicate to lay readers, especially under uncertainty and statistical model outputs?
w/Pascal Goffin & Miriah Meyer
selected Visualizing hyperspectral data

visualizing hyperspectral data

CoreInspector allows researchers to intuitively search and analyze patterns in core sections, interactively create complex multi-channel/multi-mineral maps, and curate and annotate found features.

wanna chat?

Aspen

{

  • medium
  • substack
  • linkedin