Sally Zhu

I'm an undergraduate at Stanford University studying math and computer science. I do research with the Stanford NLP group and am very fortunate to work with Percy Liang and Rohith Kuditipudi.

I'm broadly interested in language models, reasoning, robustness, and synthetic data. Please reach out if you'd like to chat!

Email  /  Scholar  /  Twitter  /  Github

profile photo

Publications


* denotes equal contribution.
Independence Tests for Language Models
Sally Zhu*, Ahmed M. Ahmed*, Rohith Kuditipudi*, Percy Liang
International Conference on Machine Learning (ICML), 2025
⭐️ spotlight award
Paper / Code / Tweet
Model tracing figure
Blackbox Model Provenance via Palimpsestic Membership Inference
Rohith Kuditipudi*, Jing Huang*, Sally Zhu*, Diyi Yang, Christopher Potts, Percy Liang
Neural Information Processing Systems (NeurIPS), 2025
⭐️ spotlight award
Paper / Code / Tweet
Model tracing figure

Other Projects

LLM-basics
Implementations of the Transformer architecture, a BPETokenizer, an AdamW optimizer, w/ training scripts, and more (from scratch). Using this code, we can train a small model for 1.5 hours on an H100 to get a validation loss of 3.55 on OpenWebText! (Assignment 1, from CS336)
Code
Analyzing Encoder Embeddings for Detecting LLM-generated Text, SemEval 2024
w/ Kevin Li, Kenan Hasanaliyev, George Altshuler, Alden Eberts, Eric Chen, Kate Wang, Emily Xia, Eli Browne, Ian Chen
18th International Workshop on Semantic Evaluation, 2024
Paper

Many thanks to Jon Barron for the source code!