Aligned to Flourish
Aligning AI and Flourishing

Latest

Jan
23

My current research and request for collaborators

I wrote this as a bio for EAG Bay Area 2024. I'm sharing this here because it gives an overview
3 min read
Dec
29

But is it really in Rome? Limitations of the ROME model editing technique

I just published a new post on LessWrong. It's about the causal tracing and model editing paper (ROME). Here's the
2 min read
Dec
29

An incomplete list of projects I'd like to work on in 2023

Wrote up a short (incomplete) bullet-point list of the projects I'd like to work on in 2023. Here's the link.
Dec
19

(Linkpost) Results for a survey of tool use and workflows in alignment research

In March 22nd, 2022, we released a survey with an accompanying post for the purpose of getting more insight into
1 min read
Dec
16

How learning efficiently applies to alignment research

As we are trying to optimize for actually solving the problem, we should not fall into the trap of learning
2 min read
Dec
07

Differential Training Process: Delaying capabilities until inner aligned

I've been ruminating on an idea ever since I read the section on deception in "The Core of the Alignment
3 min read
Dec
07

Near-Term AI capabilities probably bring low-hanging fruits for global poverty/health

I'm an alignment researcher, but I still think we should be vigilant about how models like GPT-N could potentially be
1 min read
Dec
05

Is the "Valley of Confused Abstractions" real?

Epistemic Status: Quite confused. Using this short post as a signal for discussion. Here's a link to the LessWrong post
3 min read
Dec
05

Foresight for AGI Safety Strategy

For discussion: Link to LessWrong post. Link to EA Forum post. This post is about I think why we should
10 min read
Nov
28

Notes on Cicero

Link to YouTube explanation by Yannic Kilcher: Link to paper (sharing on GDrive since it's behind a paywall on Science)
3 min read