I’m a final-year PhD student at Stanford CS, advised by Chelsea Finn. My research is supported by OpenAI and KFAS.
My research focuses on building systems that can learn from rich textual feedback. Standard learning approaches reduce learning to optimizing a scalar “loss” or “reward”, but this single number necessarily discards useful information originally present in environmental feedback.
In many real settings, much richer feedback is available: stack traces, natural-language corrections, explanations in pairwise comparisons, or long-form reflections on failed attempts. I develop methods that leverage such signals to drive continual improvement, enabling models to refine their behavior based on such rich information.
For a technical overview, see my blog post or the selected papers below.
ICLR 2026 Workshop on Memory for LLM-Based Agentic Systems (MemAgents)
ICLR 2026 Workshop on AI with Recursive Self-Improvement
ICLR 2026 · ES-FoMo @ ICML 2025 (Spotlight) · Ram2 @ CoLM 2025 (Oral)
ICML 2025 Workshop PUT
UIST 2024, NeurIPS 2023 workshops XAIA and ICBINB
ICLR 2023