Discussion about this post

User's avatar
Tim Duffy's avatar

I am more optimistic than you for a few reasons, but one that I'd like to highlight is that I think humans are likely okay in a world with mostly aligned ASIs but also some misaligned ones. In that case there will be much more intelligence and resources working in favor of humanity than against. This is far from perfect of course, since some technologies have much easier offence than defence. It's easier to create a novel pathogen than to defend against one. But in a world where most AI is aligned, it seems very likely that some humans will survive even in a worst-case engineered pandemic.

Expand full comment
Randall Bennington's avatar

This was a good piece and largely mirrors my own evolution on this question. For me, it was Zvi who flipped my thinking with the disempowerment thesis, which I find too highly plausible. Single metric, bayesian priors, have always struck me as an overly narrow frame, but it looks like LLMs think very much like Rationalists, maybe good maybe bad, but narrowing for sure. One thing to look at though, and people hate me for saying this, but Grok is the first next-FLOP-gen model. And while people are praising it for being 'open' and non-refusal, it's not **that much better** than the previous O1 (GPT 4 + Thinking) generation. Unverifiable tasks might be a bigger gap in the jagged frontier than we assume. p(doom) 10-15% depending on definition.

Expand full comment
3 more comments...

No posts