Discussion about this post

User's avatar
Daniel Kokotajlo's avatar

I'm looking forward to the rest of this series! Here are my thoughts so far:

(1) I think I basically agree with you re precautionary principle paralysis and many safety efforts being in fact counterproductive. I propose we do something like

--Divide risks into those for which iterative development works (i.e. problems that, if they happen, we'll notice and fix before major catastrophe occurs) and those for which it doesn't (i.e. problems such that, if they happen, we either plausibly won't notice or plausibly won't fix before major catastrophe occurs)

--and then have fairly permissive / move-fast-and-break-things policies for the former and more restrictive precautionary-principle-esque policies for the latter.

(2) So far this series has been long on questions and considerations, and short on answers and proposals. Fair enough I guess, but it would be nice to see some more of the latter IMO. What are your opinions about e.g. AGI timelines, takeoff speeds, AGI governance strategies, AGI alignment strategies, likely failure modes of likely-to-be-used AGI alignment strategies, etc.?

(3) The seven essays promised in this series don't seem to contain anything about misalignment, superintelligence, or loss-of-control. Just putting in a vote here in case you care what I think (I won't take offense if you don't) that I'd love to hear your thoughts on those topics!

Expand full comment
1 more comment...

No posts