• @jsomae@lemmy.ml
      link
      fedilink
      42 days ago

      what do you mean by spell fine? They’re just emitting the tokens for the words. Like, it’s not writing “strawberry,” it’s writing tokens <302, 1618, 19772>, which correspond to st, raw, and berry respectively. If you ask it to put a space between each letter, that will disrupt the tokenization mechanism, and it’s going to be quite liable to making mistakes.

      I don’t think it’s really fair to say that the lookup 19772 -> berry counts as the LLM being able to spell, since the LLM isn’t operating at that layer. It doesn’t really emit letters directly. I would argue its inability to reliably spell words when you force it to go letter-by-letter or answer queries about how words are spelled is indicative of its poor ability to spell.

      • @__dev@lemmy.world
        link
        fedilink
        32 days ago

        what do you mean by spell fine?

        I mean that when you ask them to spell a word they can list every character one at a time.

        • @jsomae@lemmy.ml
          link
          fedilink
          12 days ago

          Well that’s a recent improvement. GPT3 was very bad at that, and GPT4 still makes mistakes.

    • @buddascrayon@lemmy.world
      link
      fedilink
      22 days ago

      The problem is that it’s not actually counting anything. It’s simply looking for some text somewhere in its database that relates to that word and the number of R’s in that word. There’s no mechanism within the LLM to actually count things. It is not designed with that function. This is not general AI, this is a Generative Adversarial Network that’s using its vast vast store of text to put words together that sound like they answer the question that was asked.