• @APassenger@lemmy.world
        link
        fedilink
        English
        5010 months ago

        It’s this. When boards and non-tech savvy managers start making decisions based on a slick slide deck and a few visuals, enough will bite that people will be laid off. It’s already happening.

        There may be a reckoning after, but wall street likes it when you cut too deep and then bounce back to the “right” (lower) headcount. Even if you’ve broken the company and they just don’t see the glide path.

        It’s gonna happen. I hope it’s rare. I’d argue it’s already happening, but I doubt enough people see it underpinning recent lay offs (yet).

    • @tias@discuss.tchncs.de
      link
      fedilink
      English
      3910 months ago

      AI as a general concept probably will at some point. But LLMs have all but reached the end of the line and they’re not nearly smart enough.

      • @li10@feddit.uk
        link
        fedilink
        English
        1510 months ago

        LLMs have already reached the end of the line 🤔

        I don’t believe that. At least from an implementation perspective we’re extremely early on, and I don’t see why the tech itself can’t be improved either.

        Maybe it’s current iteration has hit a wall, but I don’t think anyone can really say what the future holds for it.

        • @jacksilver@lemmy.world
          link
          fedilink
          English
          25
          edit-2
          10 months ago

          LLMs have been around since roughly 2016 2017 (comment below corrected me that Attention paper was 2017). While scaling the up has improved their performance/capabilities, there are fundamental limitations on the actual approach. Behind the scenes, LLMs (even multimodal ones like gpt4) are trying to predict what is most expected, while that can be powerful it means they can never innovate or be truth systems.

          For years we used things like tf-idf to vectorize words, then embeddings, now transformers (supped up embeddings). Each approach has it limits, LLMs are no different. The results we see now are surprisingly good, but don’t overcome the baseline limitations in the underlying model.

        • @mashbooq@infosec.pub
          link
          fedilink
          English
          510 months ago

          I’m not trained in formal computer science, so I’m unable to evaluate the quality of this paper’s argument, but there’s a preprint out that claims to prove that current computing architectures will never be able to advance to AGI, and that rather than accelerating, improvements are only going to slow down due to the exponential increase in resources necessary for any incremental advancements (because it’s an NP-hard problem). That doesn’t prove LLMs are end of the line, but it does suggest that additional improvements are likely to be marginal.

          Reclaiming AI as a theoretical tool for cognitive science

        • @Wooki@lemmy.world
          link
          fedilink
          English
          1
          edit-2
          10 months ago

          we’re extremely early on

          Oh really! The analysis has been established since the 80’s. Its so far from early on that statement is comical

      • Optional
        link
        fedilink
        English
        610 months ago

        “at some point” being like 400 years in the future? Sure.

        Ok that’s probably a little bit of an exaggeration. 250 years.

    • @assembly@lemmy.world
      link
      fedilink
      English
      410 months ago

      The one thing that LLMs have done for me is to make summarizing and correlating data in documents really easy. Take 20 docs of notes about a project and have it summarize where they are at so I can get up to speed quickly. Works surprisingly well. I haven’t had luck with code requests.