Yeah, I find that scenario not very plausible as well. If it understands the subgoals and the effects of those subgoals and can infer that this is not what humans would want, then why can't humans specify to the AI system to actually care for subgoals too?
Yeah, I find that scenario not very plausible as well. If it understands the subgoals and the effects of those subgoals and can infer that this is not what humans would want, then why can't humans specify to the AI system to actually care for subgoals too?
I'm listening in and appreciating the discussion and the post! It seems like it might be necessary to articulate all of the end goals of human life and how to prioritize them in order to keep it from following instructions with side effects we don't find acceptable?
Yeah, I find that scenario not very plausible as well. If it understands the subgoals and the effects of those subgoals and can infer that this is not what humans would want, then why can't humans specify to the AI system to actually care for subgoals too?
I'm listening in and appreciating the discussion and the post! It seems like it might be necessary to articulate all of the end goals of human life and how to prioritize them in order to keep it from following instructions with side effects we don't find acceptable?