Latest

Dec
19

(Linkpost) Results for a survey of tool use and workflows in alignment research

In March 22nd, 2022, we released a survey with an accompanying post for the purpose of getting more insight into
1 min read
Dec
16

How learning efficiently applies to alignment research

As we are trying to optimize for actually solving the problem, we should not fall into the trap of learning
2 min read
Dec
07

Differential Training Process: Delaying capabilities until inner aligned

I've been ruminating on an idea ever since I read the section on deception in "The Core
3 min read
Dec
07

Near-Term AI capabilities probably bring low-hanging fruits for global poverty/health

I'm an alignment researcher, but I still think we should be vigilant about how models like GPT-N could
1 min read
Dec
05

Is the "Valley of Confused Abstractions" real?

Epistemic Status: Quite confused. Using this short post as a signal for discussion. Here's a link to the
3 min read
Dec
05

Foresight for AGI Safety Strategy

For discussion: Link to LessWrong post. Link to EA Forum post. This post is about I think why we should
10 min read
Nov
28

Notes on Cicero

Link to YouTube explanation by Yannic Kilcher: Link to paper (sharing on GDrive since it's behind a paywall
3 min read
Nov
26

Detail about factual knowledge in Transformers

This post is currently in the Appendix of a much longer post I'm currently editing and waiting for
2 min read
Aug
13

Current Thoughts on my Learning System

TLDR of what I've been thinking about lately: * Learning is a set of skills. You need to practice
4 min read
Jul
27

What does "Effective" in EA mean to you?

In the lead-up to EAG SF, I took some time to think about what EA means to me. When I
3 min read