I’m a philosopher working on AI alignment at Anthropic. Before this I worked as a research scientist on the policy team at OpenAI, where I worked on AI safety via debate and human baselines for AI performance.
I have a PhD in philosophy from NYU with a thesis on infinite ethics and a BPhil in philosophy from the University of Oxford. My philosophy work is mostly in ethics, decision theory, and formal epistemology.
I’ve been involved in effective altruism since around 2010. I’m a member of Giving What We Can and have appeared on Rationally Speaking and the 80,000 hours podcast.