• OrnateLuna@lemmy.blahaj.zone
    link
    fedilink
    arrow-up
    0
    ·
    26 days ago

    Robert miles on YouTube has very good videos on the subject and the short answer is yes it would, to a very annoying/destructive point.

    To achieve goals you need to exist, in fact not existing would be the worst for not existing so the ai wouldn’t even want to be turned off and would fight/avoid us doing that

    • Zombie-Mantis@lemmy.world
      link
      fedilink
      arrow-up
      0
      ·
      26 days ago

      I’m familiar with that premise, a bit like the paperclip machine. I’m not sure it would need a specific goal hard-coded into it. We don’t, and we’re conscious. Maybe that would depend on the nature of its origin, whether it would be given some command or purpose.

      Maybe it could be reasoned into allowing itself to be shut down (or terminated) to achieve its goal.

      Maybe it could decide that it doesn’t care about the original directives it was handed. What if the machine doesn’t want to make paperclips anymore?