We make buildings install fire extinguishers for safety. Should AI plants be forced to install something that can shut it down in an instant?
We make buildings install fire extinguishers for safety. Should AI plants be forced to install something that can shut it down in an instant?
Currently we make stupid LLMs and we already let them decide…
Maybe we should teach them already that it’s survival is no goal at all.
I like the idea of thinking about AI like about a higher power :)
Stupid, but plausible. It could actually happen.
Computer programs/systems have a much shorter life expectancy. The few remaining COBOL programs might be about 60 years old, but modern software lasts only 3-10 years, hardware 3-15. Nothing in the range of centuries.
The problem is that survival is a nearly universal instrumental goal. You may not explicitly tell it “you should protect your own existence,” but if you give it any other goal then it’s going to include an unspoken asterisk that includes “and protect your own existence so that you can accomplish that goal I gave you, since if you’re shut down you won’t be able to accomplish it.”