• thepianistfroggollum@lemmynsfw.com
    link
    fedilink
    English
    arrow-up
    5
    ·
    1 year ago

    That’s not a bad thing. Humans really aren’t good decision makers. Having a system with an incredible amount of input data will be able to draw better conclusions than a person.

    Just look at cars.

    • visak@lemmy.world
      link
      fedilink
      arrow-up
      5
      ·
      edit-2
      1 year ago

      Humans are good decision makers, we’re just not good at paying attention for long periods of time. Which is why I think self-driving cars will eventually be better, but they aren’t yet. And those are expert systems (I refuse to call them AI) trained on a well-curated and limited set of data for a limited and specific purpose. Which is an important difference over the generalized generative models. More data does not make systems, especially more unsorted data.

      But here’s another important difference: I can grab the wheel at any time and take over. If we are going to give these systems decision making authority there needs to be an obvious and intuitive override.

      • thepianistfroggollum@lemmynsfw.com
        link
        fedilink
        English
        arrow-up
        2
        ·
        1 year ago

        Self driving cars are already better than humans. The Waymo cars have a crash rate of 0.59/million miles driven. The national average is 2.98.

        I’m betting that most of the self driving car crashes were caused by humans, too.

    • x4740N@lemmy.world
      link
      fedilink
      English
      arrow-up
      4
      ·
      edit-2
      1 year ago

      AI is just as biased as the data that’s put into them and that data originates from humans who have their own biases so humans are just going to pass their own biases onto the AI that makes the decisions

      I don’t think ai is a good idea

      It just exists as a replacement of the human mind and with the whole population of us on earth that’s a large enough number to contribute any unique ideas to contribute to humanity

      Creating ai would just be making some sort of copy of us

      An AI is similar to an impressionable child

      • seitanic@lemmy.sdf.org
        link
        fedilink
        arrow-up
        3
        ·
        edit-2
        1 year ago

        Bias is a problem, but it can be ameliorated. I don’t agree that because AI can be biased, you should never use it.

        Creating ai would just be making some sort of copy of us

        I don’t know any humans who can munge a ginormous data set like an AI can.

        However, reproducing human intelligence in a computer would be interesting in its own right.

        • x4740N@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          However, reproducing human intelligence in a computer would be interesting in its own right.

          I would not try to replicate that knowing humanity it would probably view us as a threat

          I don’t know any humans who can munge a ginormous data set like an AI can.

          No humans cannot but we use tools made by us to do that

      • thepianistfroggollum@lemmynsfw.com
        link
        fedilink
        English
        arrow-up
        1
        ·
        1 year ago

        Why are you assuming there will be bias in the data, and that the AI couldn’t be made to correct for it? Most of the data for systems like medical AI are basically raw data, and it’s already better than humans at making an accurate diagnoses.

        I’m not sure why people seem to think humans are better than a system that can parse trillions of data points in a few seconds and apply a bunch of statistical models to it almost instantly.

        • x4740N@lemmy.world
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          I wouldn’t trust ai with medical data and neither would medical professionals since your dealong woth someone’s life here to either medical professionals are going to modify the data

          I’m not sure why people seem to think humans are better than a system that can parse trillions of data points in a few seconds and apply a bunch of statistical models to it almost instantly.

          That’s just pre-programmed pattern recognition which has been programmed by rules and data from humans