Doom from a Solution to the Alignment Problem
lesswrong.com·17h
Flag this post

Published on November 2, 2025 4:37 PM GMT

Suppose that the alignment problem is solvable, and that it is possible for one to specify and upload goals into an agent that make it do what you want while avoiding outcomes you don’t want.

Unfortunately for you, you were not the one to discover it, and now find yourself in a situation where you’re pushing granite stones with your fellow humans in service of the one who did, and that someone (or something) desires a pyramid.

In the relatively benign version of this scenario, you might not even be aware of your misfortune—your actions in service of the goal you have been tasked with give you more pleasure than you felt doing anything in your life before The Event. Not that you remember much of it, since remembering it just …

Similar Posts

Loading similar posts...