Millicent Li

Millicent Li

I'm a PhD student at Northeastern University where I'm advised by Byron Wallace. I've been mostly thinking about evaluations in interpretability, namely whether current evaluations measure what we actually care to measure and how this translates into practical use cases of interpretability (such as for auto-interpretability tools, like verbalization, probing, etc.). I also have broad interests in the science of training LMs, e.g. understanding their training dynamics when training them with synthetic data. I'm grateful to be supported by a Khoury PhD Fellowship and the NSF GRFP.

Before my PhD, I spent time at FAIR/Meta AI as an AI Resident, working with Marjan Ghazvininejad and Mike Lewis, and at Microsoft Research, working with Tristan Naumann. And even before this, I was an undergrad at the University of Washington working with Shwetak Patel on ubiquitous computing and Noah Smith on natural language processing.

Links: CV

News

April 2026

New preprint on understanding how capabilities of language models emerge during pre-training, What do Language Models Learn and When? The Implicit Curriculum Hypothesis, led by Emmy Liu!

October 2025

We investigate existing interpretability methods on decoding activations into natural language in our new preprint, Do Natural Language Descriptions of Model Activations Convey Privileged Information?

January 2025

Our paper, Multi-Field Adaptive Retrieval, done during my internship at Microsoft Semantic Machines, was accepted to ICLR 2025 as a spotlight (top 5%). Thanks to my amazing co-authors!

October 2024

New preprint on our paper, Multi-Field Adaptive Retrieval, done during my internship at Microsoft Semantic Machines!

August 2024

We've released a new preprint on causal interpretability, The Quest for the Right Mediator: A History, Survey, and Theoretical Grounding of Causal Interpretability - work done with David Bau's interpretability group.

February 2024

I've accepted an internship offer with Microsoft Semantic Machines for the upcoming summer, working with Patrick Xia and Tongfei Chen.

January 2024

Our paper, Function Vectors in Large Language Models, was accepted to ICLR 2024!

May 2023
April 2022

I was awarded a 2022 NSF Graduate Research Fellowship. Northeastern wrote an article about it here.

August 2021

Started as an AI Resident with Fundamental AI Research (FAIR) at Meta in Seattle, working on natural language processing and human-computer interaction research for a year.

May 2021

Started my internship at Microsoft Research working with Tristan Naumann on the intersection of natural language processing and healthcare!

April 2021

Excited to announce that I’ll be starting my PhD in the Khoury College of Computer Sciences at Northeastern University in Boston, fall of 2022. Thanks to everyone who has supported me on this journey thus far!

March 2021

I was awarded an Honorable Mention for the 2021 NSF Graduate Research Fellowship competition.


Publications

2026

  1. What do Language Models Learn and When? The Implicit Curriculum Hypothesis
    Emmy Liu, Kaiser Sun, Millicent Li, Isabelle Lee, Lindia Tjuatja, Jen-tse Huang, Graham Neubig
    arXiv, 2026

2025

  1. Do Natural Language Descriptions of Model Activations Convey Privileged Information?
    Millicent Li, Alberto Mario Ceballos Arroyo, Giordano Rogers, Naomi Saphra, Byron C. Wallace
    arXiv, 2025
  2. The Quest for the Right Mediator: A History, Survey, and Theoretical Grounding of Causal Interpretability
    Aaron Mueller, ... Millicent Li ... Yonatan Belinkov
    Computational Linguistics (CL), 2025
  3. Multi-Field Adaptive Retrieval
    Millicent Li, Tongfei Chen, Benjamin Van Durme, Patrick Xia
    International Conference on Learning Representations (ICLR), 2025
    Spotlight, Top 5%

2024

  1. Function Vectors in Large Language Models
    Eric Todd, Millicent Li, Arnab Sen Sharma, Aaron Mueller, Byron C Wallace, David Bau
    International Conference on Learning Representations (ICLR), 2024

2023

  1. Summarizing, Simplifying, and Synthesizing Medical Evidence using GPT-3 (with Varying Success)
    Chantal Shaib, Millicent L. Li, Sebastian Joseph, Iain Marshall, Junyi Jessy Li, Byron C. Wallace
    Annual Meeting of the Association for Computational Linguistics (ACL), 2023

2022

  1. A Review on Language Models as Knowledge Bases
    Badr AlKhamissi*, Millicent Li*, Asli Celikyilmaz^, Mona Diab^, Marjan Ghazvininejad^
    arXiv
    * denotes equal contribution
    ^ denotes equal supervision

2020

  1. Multi-Channel Facial Photoplethysmography Sensing
    Parker S. Ruth, Jerry Cao, Millicent Li, Jacob E. Sunshine, Edward J. Wang, and Shwetak N. Patel
    International Conference of the IEEE Engineering in Medicine Biology Society (EMBC 2020)