• Jack Riddle@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      2
      ·
      6 months ago

      Agi won’t kill us because it doesn’t exist and won’t emerge from what we call ai now.

      We can also just unplug it if necessary.

      • JackGreenEarth@lemm.ee
        link
        fedilink
        English
        arrow-up
        2
        ·
        6 months ago

        You clearly didn’t read the Wikipedia article I linked; an intelligent AGI would not let you unplug it.

        And regardless of whether it emerges from current AI, or is developed in a totally different way, there is no reason besides blind optimism (ie, burying your head in the sand) to feel certain it will never exist.

        • Jack Riddle@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          2
          ·
          6 months ago

          I am not saying it will never exist. I am saying it doesn’t exist right now and doesn’t look like it will for a long time. We clearly have way more pressing matters to worry about, like climate change for example.

          • JackGreenEarth@lemm.ee
            link
            fedilink
            English
            arrow-up
            2
            ·
            6 months ago

            No reason to assume it will take a very long time to happen. Its best to take it as a serious threat, unless you want extremely rapid climate change.

            Also, you didn’t address the fact that my comment addressed the second part of your original comment too. Do you accept the correction?

            • Jack Riddle@sh.itjust.works
              link
              fedilink
              English
              arrow-up
              2
              ·
              6 months ago

              No reason to assume it will take a very long time for ragnarok to happen either. Better prepare now!

              If you give the ai enough power, sure, it won’t let you unplug it maybe. Still don’t really see how it wants to prevent a hardware killswitch from being activated except for guarding it or disabling it somehow.

              • intensely_human@lemm.ee
                link
                fedilink
                English
                arrow-up
                1
                ·
                6 months ago

                Why would you put a hardware kill switch on a military robot? Then the enemy can just switch it off while it’s killing then.

                • Jack Riddle@sh.itjust.works
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  6 months ago

                  Why put a brain in a military robot? It’d just have a higher chance to fuck you over.

                  There is also no proof that any form of agi is on the way or even possible. Preparing for it over any of the things we do have proof of makes about as much sense as prepping for a zombie apocalypse.

      • Dizzy Devil Ducky@lemm.ee
        link
        fedilink
        English
        arrow-up
        1
        ·
        6 months ago

        We can also just unplug it if necessary.

        True, but do you really think the investors* will allow the companies behind the AI or running the servers the AI are hosted on would allow that?

        *parasites in human clothes