research
In reversed chronological order.
Linearly Decoding Refused Knowledge in Aligned Language Models
Aryan Shrivastava and Ari Holtzman
Under Review, 2025
AbsenceBench: Language Models Can’t Tell What’s Missing
Harvey Yiyun Fu, Aryan Shrivastava, Jared Moore, Peter West, Chenhao Tan, Ari Holtzman
Under Review, 2025
DICE: A Framework for Dimensional and Contextual Evaluation of Language Models
Aryan Shrivastava and Paula Akemi Aoyagui
CHI 2025 Human-Centered Evaluation and Auditing of Language Models (HEAL) Workshop, 2025
Moving Beyond Medical Exam Questions: A Clinician-Annotated Dataset of Real-World Tasks and Ambiguity in Mental Healthcare
Max Lamparth, Declan Grabb, Amy Franks, Scott Gershan, Kaitlyn N. Kunstman, Aaron Lulla, Monika Drummond Roots, Manu Sharma, Aryan Shrivastava, Nina Vasan, Colleen Waickman
Under Review, 2025
Measuring Free-Form Decision-Making Inconsistency of Language Models in Military Crisis Simulations
Aryan Shrivastava, Jessica Hullman, Max Lamparth
NeurIPS 2024 Socially Responsible Language Modelling Research (SoLaR) Workshop; MILA 2024 Harms and Risks of AI in Military Workshop, 2024