I am a fifth year JD-PhD student in Computer Science at Stanford University (advised by Chris Ré). I'm a part of the Hazy Research Lab, Stanford Center for Research on Foundation Models, and RegLab. I graduated with a MS in Machine Learning from Carnegie Mellon University ('19) and a BS (with Honors) in Computer Science from Stanford University ('18). I am grateful to be supported by the Stanford Interdisciplinary Graduate Fellowship (SIGF) and the HAI Graduate Fellowship.
My research lies at the intersection of artificial intelligence/machine learning (AI/ML) and law. Most of my work can be organized into four buckets:
-
Using machine learning to study the law and legal institutions. ML provides a useful set of techniques to perform large scale analysis of different legal corpora---cases, statutes, regulations, and more. I apply these techniques in the context of different substantive legal questions (often with a focus on civil procedure). For instance, my prior work has tested theories of private enforcement at the state level, by building a database of state private rights of action (U. Pa. L. Rev. 2024). In ongoing work, I am extending these techniques to build an open and accessible database of state laws, annotated with a broad spectrum of relevant features.
-
Studying questions around AI governance. I'm interested in how technical nuances of AI affect governance.
My recent work has examined trends in how courts have approached liability for medical AI (NEJM 2024), and the technical and institutional trade-offs of commonly proposed regulatory interventions
(GW L. Rev. 2024). In an ongoing project, I am examining how informational and structural properties of AI applications influence regulatory design across different application/use-contexts.
-
Measuring the legal usefulness/capabilities of large language models (LLMs). I work on building benchmarks to assess how well LLMs can perform different legal tasks. Measurement on legal tasks serves as a useful proxy for assessing reasoning capabilities generally, and also clarifies opportunities for real-world use. Examples of this work include
LegalBench (NeurIPS 2023),
CaseHOLD (ICAIL 2021),
RAG benchmarks (CS&Law 2025), and
LoCo (ICML 2024).
-
Improving LLM performance when labeled data is scarce. Producing labeled data for tasks is often expensive or impractical. I've studied how information contained within other prexisting sources may be extracted and leveraged to improve model performance for specific tasks. Examples of auxiliary sources of information include pretrained embedding models (e.g., BERT), and knowledge bases (e.g., Freebase, Wikipedia). Examples of this work include
Smoothie (NeurIPS 2024),
Embroid (NeurIPS 2023), and
Bootleg (CIDR 2021).