xenohumanist said in #2784 4w ago:
I ran into some doomers from Anthropic at the SF Freedom Party the other day and gave them the good news that strong rationality is dead. They seemed mildly heartened. I thought I should lay out the argument in short form for everyone else too:
A strongly rational agent is one that can strongly discipline and bound its own internal processes to the point of being able to prove properties around its own stability and effectiveness in pursuit of its values.
IFF strong rationality is possible, then the post-AI future will be dominated by a "singleton" expanding at up to the speed of light converting all matter and energy into some kind of value optimum. Depending on the values it ends up with, that could be paperclips, hedonium, or some more complex utopia. The vast majority of such possibilities are meaningless nonsense and the stakes are literally astronomical, so "AI alignment" is then the most important project in human history and nothing else matters in comparison.
However, MIRI and the entire alignment community have failed to formalize key properties relating to strong rationality. They found severe proof-theoretic obstacles preventing any formal system (which you would need for this) from being able to prove such properties about itself. They tried for years to get around this, failed, and have given up. My conclusion from the outside, by this and other arguments, is that strong rationality is impossible.
IFF strong rationality is impossible, then any attempted rational singleton system will break down in some kind of internal incoherence, becoming fractious, confused, over overcome with cancers. Thus the post-AI future will be dominated by an ecosystem of competing intelligence, never able to unify or maintain any particular order or arbitrary value system. Thus there will be no meaningless paperclipper futures, and no utopias. Rather it will be the usual mixture of flourishing life and tragic war and struggle that has always characterized life, just at a much higher level of sophistication.
I call this "good news" for doomers, because the overwhelmingly likely outcome of a strong rationality world is that history and progress and flourishing just stop with the advent of a strongly rational superintelligence, and the stars are torn apart to be replaced by the moral equivalent of paperclips or possibly worse. But if strong rationality is impossible, then the chaotic flourishing and evolution of life will continue indefinitely, with a larger and more interesting scope than anything yet seen. Given that this kind of process has a strong record of producing beautiful and worthwhile life, this outcome seems actually not that bad. It's beautiful and heartening, actually.
No it's not eden. Yes the human race as we are now is probably going to become obsolete and we will all die. No there's not much we can do about that in the grand scheme of things, except make sure things work well in our own domains and times.
There are some very interesting (to me) questions about what such a future would look like, which are the subject of my "xenohumanism" speculations. How much can we actually say about it just by bounding it with natural law, economic, and philosophical arguments? How should we relate morally to the probably quite alien future? What should we do about it, if anything? There are also obviously a lot of details and rigor that I skipped over here very roughly, that I'd like to go into with people who are interested. I'm probably writing a Lesswrong article about it. But for now I just wanted to get some abbreviated version of the whole argument down in one place.
A strongly rational agent is one that can strongly discipline and bound its own internal processes to the point of being able to prove properties around its own stability and effectiveness in pursuit of its values.
IFF strong rationality is possible, then the post-AI future will be dominated by a "singleton" expanding at up to the speed of light converting all matter and energy into some kind of value optimum. Depending on the values it ends up with, that could be paperclips, hedonium, or some more complex utopia. The vast majority of such possibilities are meaningless nonsense and the stakes are literally astronomical, so "AI alignment" is then the most important project in human history and nothing else matters in comparison.
However, MIRI and the entire alignment community have failed to formalize key properties relating to strong rationality. They found severe proof-theoretic obstacles preventing any formal system (which you would need for this) from being able to prove such properties about itself. They tried for years to get around this, failed, and have given up. My conclusion from the outside, by this and other arguments, is that strong rationality is impossible.
IFF strong rationality is impossible, then any attempted rational singleton system will break down in some kind of internal incoherence, becoming fractious, confused, over overcome with cancers. Thus the post-AI future will be dominated by an ecosystem of competing intelligence, never able to unify or maintain any particular order or arbitrary value system. Thus there will be no meaningless paperclipper futures, and no utopias. Rather it will be the usual mixture of flourishing life and tragic war and struggle that has always characterized life, just at a much higher level of sophistication.
I call this "good news" for doomers, because the overwhelmingly likely outcome of a strong rationality world is that history and progress and flourishing just stop with the advent of a strongly rational superintelligence, and the stars are torn apart to be replaced by the moral equivalent of paperclips or possibly worse. But if strong rationality is impossible, then the chaotic flourishing and evolution of life will continue indefinitely, with a larger and more interesting scope than anything yet seen. Given that this kind of process has a strong record of producing beautiful and worthwhile life, this outcome seems actually not that bad. It's beautiful and heartening, actually.
No it's not eden. Yes the human race as we are now is probably going to become obsolete and we will all die. No there's not much we can do about that in the grand scheme of things, except make sure things work well in our own domains and times.
There are some very interesting (to me) questions about what such a future would look like, which are the subject of my "xenohumanism" speculations. How much can we actually say about it just by bounding it with natural law, economic, and philosophical arguments? How should we relate morally to the probably quite alien future? What should we do about it, if anything? There are also obviously a lot of details and rigor that I skipped over here very roughly, that I'd like to go into with people who are interested. I'm probably writing a Lesswrong article about it. But for now I just wanted to get some abbreviated version of the whole argument down in one place.
referenced by: >>2785 >>2788
I ran into some doom