I am an MIT EECS PhD student working on human-AI interaction, advised by Mitchell Gordon. I want to build AI that is , by which I mean better at aiding thoughtful human judgement about real problems in the world (e.g., in Philosophy, law, politics, personal life). My work is supported by the NSF GRFP and PD Soros Fellowship.
What am I interested in and what is my research vision?
I am interested in philosophical problems. You could also approximately call them “open-ended” or “semi-structured” problems. I take philosophical problems to involve systematic inquiry into foundational and/or abstract questions dealing with values and interpretations, that we make progress on with human judgement. These include questions in the discipline of Philosophy (e.g., are moral statements truth-apt? what is the basis for knowledge?), law (e.g., how should we interpret this case?), politics (e.g., what should my views on this political issue be? what political action should I take?), therapy & personal life (e.g., how should we understand death? should I have children?), and AI Alignment (e.g. how do humans supervise AI that may become vastly more intelligent than us, possibly through other AIs?). Such problems also arise in science (e.g., how might this drug which fixes one issue with a patient create other issues that adversely affect their wellness? how do we make sense of the reproducability crisis in science?) and math (there are an infinite number of true mathematical statements, but which ones are the interesting ones?).
I think a key feature of philosophical problems is that human judgement is needed to make meaningful progress. As a thought experiment, imagine Dario’s “country of geniuses in a datacenter” running for a thousand years to prove (or disprove) a bunch of mathematical conjectures. It is conceivable that they will have made genuine progress at the end of these 1000 years (although they may also prove conjectures that they propose but we don’t think are that interesting or useful). Now imagine a “country of geniuses in a datacenter” running for a thousand years to discover “philosophical truths”. They come back to us humans and give us the list of truths they have discovered: 1. Humans actually don’t have free will, 2. ought implies can, 3. … I think this seems a little bit ridiculous, not because these AI geniuses may not be right, but because without my human judgement — my capacity to reflect upon and endorse statements that I take myself to be responsible for — these 1000 years of “multi-agent debate” don’t mean anything (to me, which is the point). (My language and thinking here is taken heavily from the philosophers Robert Brandom and Wilfred Sellars.)
Okay, so human judgement is essential to making progress on these open-ended questions. So let’s throw AI out of the window because it, in principle, lacks the capacity for this kind of judgement. No: judgement has a strong relationship with what philosopher Brian Cantwell Smith calls “reckoning” — a calculating prowess for uncovering patterns and reasoning out. Good judgement emerges from reflecting on the fruits of reckoning, or even doing reckoning ourselves. Great friends don’t tell us what to think, but rather help us see other ways of reckoning that we had not reckoned with before, and which we may judge to be ways of reckoning that we want to endorse. Now ideally everyone would have access to these kinds of friends, but we don’t, and even if we do, wouldn’t we want more?
To realize this vision, I think we need to give AI more conceptual structure. That is, we need to allow people and AI to collaborate on “shapes of thinking” — structures or scaffolds to set the terms and goals of inquiry into an open-ended question — and, with the iteratively developed endorsement of human judgement, realize what kinds of ideas those shapes of thinking will lead to. As AI models get better and better, they will better be able to realize the shapes of thinking that we judge to be worth pursuing.
I recently proposed the as a better kind of AI output for open-ended questions like in philosophy or AI alignment, in which models represent, evaluate, and improve on how different conceptual decisions condition outputs.
I’ve also thought about and how the “woke Gemini controversy” could instruct us in .
See my and a pdf CV.
Many of the open-ended questions we bring to AI involve a cascade of decisions that spills into a landscape of possibilities — how to frame the question, what to value, what assumptions to make — but we typically receive a single result, presented as if it were the only one. We have no way to know how much hinges on decisions we never saw. In data analysis, “multiverse analysis” addresses this by systematically varying these choices and reporting how the result changes. We ask: what would multiverse analysis look like for the kinds of value-laden, open-ended problems people are bringing to AI?
We introduce the conceptual multiverse: an interactive system that surfaces the conceptual decisions an AI’s answer depends on, and the alternatives each decision opens up. Instead of taking a single answer on faith, a person can navigate a tree of branch points — transparently inspecting, intervenably changing, and checking each step against the reasoning norms of the domain. We develop a verification framework that enforces properties of good decision structures (like unambiguity and completeness) calibrated by expert-level reasoning, and we represent the multiverse concretely as a Python program that agents can build, verify, and regenerate.
We calibrated and built the system in three domains — philosophy (an open-ended question), AI alignment (a sensitive user request), and poetry (a generation prompt) — and ran in-depth studies with 15 participants. Across all three, the conceptual multiverse helped participants develop a working map of the problem, surfacing thinking that uncontextualized answers would have left buried: philosophy students rewrote essays with sharper framings and reversed theses; alignment annotators moved from surface preferences to reasoning about user intent and harm; poets identified compositional patterns that clarified their taste. arXiv and website coming soon!
Current work in language models (LMs) helps us speed up or even skip thinking by accelerating and automating cognitive work. But can LMs help us with critical thinking — thinking in deeper, more reflective ways which challenge assumptions, clarify ideas, and engineer new concepts? We treat philosophy as a case study in critical thinking, and interview 21 professional philosophers about how they engage in critical thinking and on their experiences with LMs. We find that philosophers do not find LMs to be useful because they lack a sense of selfhood (memory, beliefs, consistency) and initiative (curiosity, proactivity). We propose the selfhood-initiative model for critical thinking tools to characterize this gap. Using the model, we formulate three roles LMs could play as critical thinking tools: the Interlocutor, the Monitor, and the Respondent. We hope that our work inspires LM researchers to further develop LMs as critical thinking tools and philosophers and other ‘critical thinkers’ to imagine intellectually substantive uses of LMs.
Current image generation paradigms prioritize actualizing user intention — “see what you intend” — but often neglect the sociopolitical dimensions of this process. However, it is increasingly evident that image generation is political, contributing to broader social struggles over visual meaning. This sociopolitical aspect was highlighted by the March 2024 Gemini controversy, where Gemini faced criticism for inappropriately injecting demographic diversity into user prompts. Although the developers sought to redress image generation’s sociopolitical dimension by introducing diversity “corrections,” their opaque imposition of a standard for “diversity” ultimately proved counterproductive. In this paper, we present an alternative approach: an image generation interface designed to embrace open negotiation along the sociopolitical dimensions of image creation. Grounded in the principles of agonistic pluralism (from the Greek agon, meaning struggle), our interface actively engages users with competing visual interpretations of their prompts. Through a lab study with 29 participants, we evaluate our agonistic interface on its ability to facilitate reflection — engagement with other perspectives and challenging dominant assumptions — a core principle that underpins agonistic contestation. We compare it to three existing paradigms: a standard interface, a Gemini-style interface that produces “diverse” images, and an intention-centric interface suggesting prompt refinements. Our findings demonstrate that the agonistic interface enhances reflection across multiple measures, but also that reflection depends on users perceiving the interface as both appropriate and empowering; introducing diversity without grounding it in relevant political contexts was perceived as inauthentic. Our results suggest that diversity and user intention should not be treated as opposing values to be balanced.
Loading papers…
I completed my BS in Computer Science, advised by Amy Zhang and Ranjay Krishna, and my BA in Philosophy with Honors, advised by Rose Novick, at the University of Washington. For my work, I was recognized as an Allen School Outstanding Senior, Philosophy Department Outstanding Graduating Senior, Dean’s Medalist for the Social Sciences, and served as a banner carrier for the College of Arts and Sciences at the 150th UW Commencement, and . I was also covered by the UW Office of Undergraduate Academic Affairs.
I got up to other stuff in undergrad . I greatly appreciate the interdisciplinary environment I was allowed to play in at UW.
I maintain a list of that stick with me in my readings. One of my favorites is this one by Deleuze: “A concept is a brick. It can be used to build a courthouse of reason. Or it can be thrown through the window.”
I also maintain a list of that I think are especially interesting, thought-provoking, or good reading.
Entirely for my intellectual edification, I like to learn enough math to understand what some say but not enough to know their proofs (it ruins the magic!).
I also maintain a to unify various historical events I have an interest in. It does the service of helping us realize how complex the world is, that so much barbarity, genius, strife, and goodness happen at the same time.
In an effort to become a more interesting person, I am learning woodworking at the MIT Hobby Shop and guitar; enjoy playing piano, lifting, using my AMC Stubs membership with Barish Namazov, and running along the Charles with Ryan Bahlous-Boldi; and am currently reading East of Eden with Carmel Schare, A Spirit of Trust with Mark Pock, The Design of Everyday Things with Krishna Panchapagesan, and Empire of AI with Vivek Prakriya. Я тоже учу русский. My day in the life was featured by MIT CSAIL.
See also my PhD SOP, my writing on what it means to study philosophy (p. 6), a very short story, my first published research article in literary analysis, my work on AyahuascaNet (p. 27) which went viral, and how an (admittedly sloppy) blogpost I wrote when I was 14 led to an ML Twitter controversy involving Stephen Pinker, Yann Lecun, Francois Chollet, and others.
I was also mentored by Jared Moore, Quan Ze (Jim) Chen, Sebastin Santy, Matt Wallingford, Carina Fourie, and others.
I wrote two books on deep learning for Apress/Springer Nature: Modern Deep Learning Design and Applications (solo) and Modern Deep Learning for Tabular Data (with Andy Wang).
I was a TA for 7 quarters across intro Python programming (CSE 160), intermediate Python (CSE 163), and discrete math for CS (CSE 311). I also TA’d English Composition and Literary Analysis at the Robinson Center, and Introduction to Machine Learning at The Coding School.
I entered university at 14 through the Robinson Center Early Entrance Program. My worldview was especially transformed during Transition School by Amanda Zink teaching English and Michael Reagan teaching History.
One of my close friends who also went through the Early Entrance Program with me is Vivek Prakriya, who is the youngest city councilmember to win a contested race in a city with over 75k people (Redmond, Washington). In Summer 2025 I managed social media for his campaign — running content for his run for Redmond City Council seat no. 2.
I have reviewed for CSCW, the NeurIPS Moral Psychology & Moral Philosophy Workshop, and the ICML AI + HCI Workshop, and have served as a research panelist and student host at events for the Allen School and the Robinson Center.
I spent summer of 2022 at Deepgram working on speech transcription models, advised by Andrew Seagraves.
An unorganized and growing archive of quotes, ideas, and works that influence or represent my thought. Order is randomized.
Academic papers and articles that influence my research and thinking. By the way, clicking on one of the squares in the background takes you to a random article.
“To everything there is a season, and a time to every purpose under Heaven; a time to be born and a time to die; a time to kill and a time to heal; a time to weep and a time to laugh; a time to love and a time to hate; a time for war and a time of peace.”
Ecclesiastes 3:1–8
In making this website, the brushstroke-like highlighting was inspired by Ryan Yen's website. (By the way... this website looks cooler on desktop!)