(this is an expanded, edited version of an x.com post) It is easy to interpret Eliezer Yudkowsky’s main goal as creating a friendly AGI. Clearly, he has failed at this goal and has little hope of achieving it. That’s not a particularly interesting analysis, however. A priori, creating a machine that makes things ok forever … Continue reading Executable philosophy as a failed totalizing meta-worldview