this post was submitted on 05 Dec 2024
7 points (81.8% liked)
Futurology
1852 readers
86 users here now
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I'd say having some kind of goals is definitional in AGI, so in a broad sense of "alignment" that would include "paperclip optimisers", sure, it's bound to be possible. Natural GI exists, after all.
Speculatively, if you allow it to do controversial things some of the time, my guess is that there is a way to align it so that the average person will agree with most of the time. The trouble is just getting everyone to accept the existence of the edge case.
The versions of utilitarianism usually give acceptable answers, for example, but there's the infamous fact that they imply we might consider killing people for their organs. Similarly, deontology like "don't kill people" runs into problems in a world where retaliation is usually the only way to stop someone else violent. We're just asking a lot when we want a set of rules that gives perfect options in an imperfect world.