I’m interested in how multi-agent LLM systems exhibit failure modes that are invisible to single-agent evaluation — emergent collusion, cross-agent goal drift, deceptive coordination — and in building auditing infrastructure that can catch and steer them before deployment. The work runs along three threads: studying failure itself, especially the misalignment that emerges under conditions where standard evaluation looks clean (Betley et al., 2025); auditing mechanistically, on what these systems are actually doing inside when they coordinate, deceive, or drift; and agents and society, on how multi-agent systems get embedded in human institutions and the technical analysis that makes that coexistence work (Reuel et al., 2024).
Currently a Master’s student in CS at Dartmouth College, co-advised by Nikhil Singh (Science and Art of Human-AI Systems Lab) and Soroush Vosoughi (Minds, Machines, and Society Lab). Before this, I did my undergrad at New York University, and undergrad thesis with Multimodal Agentic Personalization Systems (MAPS) Group.
News
- Jan 03, 2026 Joined SEHAS Lab at Dartmouth for alignment and mechanistic interpretability research.
- Nov 10, 2025 Best Poster Presentation Award at Technigala, Dartmouth College.
- Sep 02, 2025 Started M.S. in Computer Science (Research Track) at Dartmouth College.
- May 20, 2025 Selected as AI/ML Spark Fellow at Gobi Partners (12 of 2,000+ applicants).
- Apr 05, 2025 Launched NEVA — a multimodal AI system for personalized education (80+ college students).
- Nov 15, 2024 Most Popular Research Poster, NYU Undergraduate Research Symposium.