• jorp@lemmy.world
    link
    fedilink
    arrow-up
    12
    ·
    7 months ago

    Brain structures aren’t so dissimilar, unless you believe there’s some metaphysical quantity to consciousness this kind of technology will be how we do achieve general AI

    • AppleTea@lemmy.zip
      link
      fedilink
      arrow-up
      16
      ·
      7 months ago

      Living, growing, changing cells are pretty damn dissimilar to static circuitry. Neural networks are based on an oversimplified model of neuron cells. The model ignores the fact neurons are constantly growing, shifting, and breaking connections with one another, and flat out does not consider structures and interactions within the cells.

      Metaphysics is not required to make the observation that computer programmes are magnitudes less complex than a brain.

      • ChickenLadyLovesLife@lemmy.world
        link
        fedilink
        English
        arrow-up
        17
        ·
        edit-2
        7 months ago

        Neural networks are based on an oversimplified model of neuron cells.

        As a programmer who has studied neuroanatomy and the structure/function of neurons themselves, I remain astonished at how not like real biological nervous systems computer neural networks still are. It’s like the whole field is based on one person’s poor understanding of the state of biological knowledge in the late 1970s. That doesn’t mean it’s not effective in some ways as it is, but you’d think there’d be more experimentation in neural networks based on current biological knowledge.

          • ChickenLadyLovesLife@lemmy.world
            link
            fedilink
            English
            arrow-up
            4
            ·
            7 months ago

            The one thing that stands out to me the most is that programmatic “neurons” are basically passive units that weigh inputs and decide to fire or not. The whole net is exposed to the input, the firing decisions are worked through the net, and then whatever output is triggered. In biological neural nets, most neurons are always firing at some rate and the inputs from pre-synaptic neurons affect that rate, so in a sense the passed information is coded as a change in rate rather than as an all-or-nothing decision to fire or not fire as is the case with (most) programmatic neurons. Implementing something like this in code would be more complicated, but it could produce something much more like a living organism which is always doing something rather than passively waiting for an input to produce some output.

            And TBF there probably are a lot of people doing this kind of thing, but if so they don’t get much press.

            • areyouevenreal@lemm.ee
              link
              fedilink
              arrow-up
              2
              ·
              7 months ago

              Pretty much all artificial neural nets I have seen don’t do all or nothing activation. They all seem to have activation states encoded as some kind of binary number. I think this is to mimic the effects of variable firing rates.

              The idea of a neural network doing stuff in the background is interesting though.

      • NιƙƙιDιɱҽʂ@lemmy.world
        link
        fedilink
        arrow-up
        4
        ·
        7 months ago

        The fact that you believe software based neural networks are, as you put it, “static circuitry” betrays your apparent knowledge on the subject. I agree that many people overblow LLM tech, but many people like yourself grossly underestimate it as well.

    • AProfessional@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      ·
      7 months ago

      This is all theoretical. Today it’s quite basic with billions thrown at the problem. Maybe in decades these ideas can be expanded on.