• MotoAsh@piefed.social
    link
    fedilink
    English
    arrow-up
    7
    ·
    edit-2
    9 hours ago

    Yes it does. “Not as much as training” is a stratospheric bar…

    You may as well say, “my car doesn’t use gas because American semis use way more”. They both still use a resource we should be more careful with.

    • Europellinore@europe.pub
      link
      fedilink
      English
      arrow-up
      2
      ·
      5 hours ago

      I recently tried to calculate this for my company. I wouldn’t call it negligible, but the impact of all video calls turned out to be much greater than the impact of AI.

      • MotoAsh@piefed.social
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        3 hours ago

        Of course that’s going to produce a heavier load on your part of the infrastructure… That stuff is running locally.

        Also, it’s pretty easy to have effective calls and turn off the damn video. Most people don’t need to stare at everyone staring at their computers.

      • lime!@feddit.nu
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        8 hours ago

        that’s old data. inference uses more than training now, since usage has gone up significantly. they traded places in march or april 2025.

        • Elvith Ma'for@feddit.org
          link
          fedilink
          English
          arrow-up
          2
          ·
          5 hours ago

          That’s the problem of reference. Your individual queries might not consume much - especially when compared to the training - but the more people use it, the more the whole consumption is. At some point running those models will consume more than training them