• @kkj@lemmy.dbzer0.com
      link
      fedilink
      English
      4310 days ago

      And that’s what happens when you spend a trillion dollars on an autocomplete: amazing at making things look like whatever it’s imitating, but with zero understanding of why the original looked that way.

      • @CanadaPlus@lemmy.sdf.org
        link
        fedilink
        6
        edit-2
        9 days ago

        I mean, there’s about a billion ways it’s been shown to have actual coherent originality at this point, and so it must have understanding of some kind. That’s how I know I and other humans have understanding, after all.

        What it’s not is aligned to care about anything other than making plausible-looking text.

        • @Jtotheb@lemmy.world
          link
          fedilink
          119 days ago

          Coherent originality does not point to the machine’s understanding; the human is the one capable of finding a result coherent and weighting their program to produce more results in that vein.

          Your brain does not function in the same way as an artificial neural network, nor are they even in the same neighborhood of capability. John Carmack estimates the brain to be four orders of magnitude more efficient in its thinking; Andrej Karpathy says six.

          And none of these tech companies even pretend that they’ve invented a caring machine that they just haven’t inspired yet. Don’t ascribe further moral and intellectual capabilities to server racks than do the people who advertise them.

          • @CanadaPlus@lemmy.sdf.org
            link
            fedilink
            2
            edit-2
            9 days ago

            Coherent originality does not point to the machine’s understanding; the human is the one capable of finding a result coherent and weighting their program to produce more results in that vein.

            You got the “originality” part there, right? I’m talking about tasks that never came close to being in the training data. Would you like me to link some of the research?

            Your brain does not function in the same way as an artificial neural network, nor are they even in the same neighborhood of capability. John Carmack estimates the brain to be four orders of magnitude more efficient in its thinking; Andrej Karpathy says six.

            Given that both biological and computer neural nets very by orders of magnitude in size, that means pretty little. It’s true that one is based on continuous floats and the other is dynamic peaks, but the end result is often remarkably similar in function and behavior.

              • @CanadaPlus@lemmy.sdf.org
                link
                fedilink
                17 days ago

                I actually was going to link the same one I always do, which I think I heard about through a blog or talk. If that’s not good enough, it’s easy to devise your own test and put it to an LLM. The way you phrased that makes it sound like you’re more interested in ignoring any empirical evidence, though.

                • @Jtotheb@lemmy.world
                  link
                  fedilink
                  13 days ago

                  That’s unreal. No, you cannot come up with your own scientific test to determine a language model’s capacity for understanding. You don’t even have access to the “thinking” side of the LLM.

                  • @CanadaPlus@lemmy.sdf.org
                    link
                    fedilink
                    1
                    edit-2
                    5 hours ago

                    You can devise a task it couldn’t have seen in the training data, I mean. Building a comprehensive argument out of them requires a lot more work and time.

                    You don’t even have access to the “thinking” side of the LLM.

                    Obviously, that goes for the natural intelligences too, so it’s not really a fair thing to require.

    • @petey@aussie.zone
      link
      fedilink
      29 days ago

      It needs good feedback. Agentic systems like Roo Code and Claude Code run compilers and tests until it works (just gotta make sure to tell it to leave the tests alone)