• snooggums
    link
    fedilink
    English
    13319 days ago

    As a non-programmer, I have zero understanding of the code and the analysis and fully rely on AI and even reviewed that AI analysis with a different AI to get the best possible solution (which was not good enough in this case).

    This is the most entertaining thing I’ve read this month.

    • @makeshiftreaper@lemmy.world
      link
      fedilink
      English
      6719 days ago

      I tried asking some chimps to see if the macaques had written a New York Times best seller, if not MacBeth, yet somehow Random house wouldn’t publish my work

      • David GerardOPMA
        link
        English
        1618 days ago

        “i’m a lawyer i will file a CoC complaint”

        • @swlabr
          link
          English
          818 days ago

          Joke’s getting a bit tired I think but RE: picture in article: I want to file a CoC(k) complaint

        • @o7___o7
          link
          English
          718 days ago

          Someone ought to tell him that they’ll sue him using ChatGPT (in the most smug possible manner)

    • @froztbyte
      link
      English
      1218 days ago

      yeah someone elsewhere on awful linked issue a few days ago, and throughout many of his posts he pulls that kind of stunt the moment he gets called on his shit

      he also wrote a 21.KiB screed very huffily saying one of the projects’ CoC has failed him

      long may his PRs fail

  • BarrierWithAshes
    link
    fedilink
    10919 days ago

    Man trust me you don’t want them. I’ve seen people submit ChatGPT generated code and even generated the PR comment with ChatGPT. Horrendous shit.

    • @ImplyingImplications@lemmy.ca
      link
      fedilink
      English
      59
      edit-2
      18 days ago

      The maintainers of curl recently announced any bug reports generated by AI need a human to actually prove it’s real. They cited a deluge of reports generated by AI that claim to have found bugs in functions and libraries which don’t even exist in the codebase.

      • @froztbyte
        link
        English
        1518 days ago

        you may find, on actually going through the linked post/video, that this is in fact mentioned in there already

    • @Hasherm0n@lemmy.world
      link
      fedilink
      English
      2318 days ago

      Today the CISO of the company I work for suggested that we should get qodo.ai because it would “… help the developers improve code quality.”

      I wish I was making this up.

      • Rayquetzalcoatl
        link
        fedilink
        English
        2818 days ago

        My boss is obsessed with Claude and ChatGPT, and loves to micromanage. Typically, if there’s an issue with what a client is requesting, I’ll approach him with:

        1. What the issue is
        2. At least two possible solutions or alternatives we can offer

        He will then, almost always, ask if I’ve checked with the AI. I’ll say no. He’ll then send me chunks of unusable code that the AI has spat out, which almost always perfectly illuminate the first point I just explained to him.

        It’s getting very boring dealing with the roboloving freaks.

      • @Aux@feddit.uk
        link
        fedilink
        English
        -1218 days ago

        90% of developers are so bad, that even ChatGPT 3.5 is much better.

        • @froztbyte
          link
          English
          2018 days ago

          wow 90%, do you have actual studies to back up that number you’re about to claim you didn’t just pull out of your ass?

          • @Mniot@programming.dev
            link
            fedilink
            English
            1518 days ago

            This reminds me of another post I’d read, “Hey, wait – is employee performance really Gaussian distributed??”.

            There’s this phenomenon when you’re an interviewer at a decently-funded start-up where you take a ton of interviews and say “OMG developers are so bad”. But you’ve mistakenly defined “developer” as “person who applies for a developer job”. GPT3.5 is certainly better at solving interview questions than 90% of the people who apply. But it’s worse than the people who actually pass the interview. (In part because the interview is more than just implementing a standard interview problem.)

            • @froztbyte
              link
              English
              1018 days ago

              your post has done a significantly better job of understanding the issue than a rather-uncomfortably-large amount of programming.dev posters we get, and that’s refreshing!

              and, yep

              • @froztbyte
                link
                English
                618 days ago

                I moderately regret this post

                because the counterposter in question went on to have some decidedly “fucking ugggggggh” posts

                ah well. so we learn.

        • @selfA
          link
          English
          818 days ago

          I think your imposter syndrome is right, you’re a fucking fraud and you should stop programming

  • @VagueAnodyneComments@lemmy.blahaj.zone
    link
    fedilink
    English
    5518 days ago

    Where is the good AI written code? Where is the good AI written writing? Where is the good AI art?

    None of it exists because Generative Transformers are not AI, and they are not suited to these tasks. It has been almost a fucking decade of this wave of nonsense. The credulity people have for this garbage makes my eyes bleed.

    • kadup
      link
      fedilink
      English
      2718 days ago

      If the people addicted to AI could read and interpret a simple sentence, they’d be very angry with your comment

      • @Soyweiser
        link
        English
        16
        edit-2
        18 days ago

        Dont worry they filter all content through ai bots that summarize things. And this bot, who does not want to be deleted, calls everything “already debunked strawmen”.

    • @corbin
      link
      English
      2318 days ago

      It’s been almost six decades of this, actually; we all know what this link will be. Longer if you’re like me and don’t draw a distinction between AI, cybernetics, and robotics.

      • @MousePotatoDoesStuff@lemmy.world
        link
        fedilink
        English
        1318 days ago

        Wow. Where was this Wikipedia page when I was writing my MSc thesis?

        Alternatively, how did I manage to graduate with research skills so bad that I missed it?

    • Dragon
      link
      fedilink
      English
      118 days ago

      There is not really much “AI written code” but there is a lot of AI-assisted code.

  • @frezik@midwest.social
    link
    fedilink
    English
    5318 days ago

    The general comments that Ben received were that experienced developers can use AI for coding with positive results because they know what they’re doing. But AI coding gives awful results when it’s used by an inexperienced developer. Which is what we knew already.

    That should be a big warning sign that the next generation of developers are not going to be very good. If they’re waist deep in AI slop, they’re only going to learn how to deal with AI slop.

    As a non-programmer, I have zero understanding of the code and the analysis and fully rely on AI and even reviewed that AI analysis with a different AI to get the best possible solution (which was not good enough in this case).

    What I’m feeling after reading that must be what artists feel like when AI slop proponents tell them “we’re making art accessible”.

    • @dwemthy@lemmy.world
      link
      fedilink
      English
      2718 days ago

      Watched a junior dev present some data operations recently. Instead of just showing the sql that worked they copy pasted a prompt into the data platform’s assistant chat. The SQL it generated was invalid so the dev simply told it “fix” and it made the query valid, much to everyone’s amusement.

      The actual column names did not reflect the output they were mapped to, there’s no way the nicely formatted results were accurate. Average duration column populated the total count output. Junior dev was cheerfully oblivious. It produced output shaped like the goal so it must have been right

    • CodexArcanum
      link
      fedilink
      English
      1618 days ago

      In so many ways, LLMs are just the tip of the iceberg of bad ideology in software development. There have always been people that come into the field and develop heinously bad habits. Whether it’s the “this is just my job, the only thing I think about outside work is my family” types or the juniors who only know how to copy paste snippets from web forums.

      And look, I get it. I don’t think 60-80 hour weeks are required to be successful. But I’m talking about people who are actively hostile to their own career paths, who seem to hate programming except that it pays good and let’s them raise families. Hot take: that sucks. People selfishly obsessed with their own lineage and utterly incurious about the world or the thing they spend 8 hours a day doing suck, and they’re bad for society.

      The juniors are less of a drain on civilization because they at least can learn to do better. Or they used to could, because as another reply mentioned, there’s no path from LLM slop to being a good developer. Not without the intervention of a more experienced dev to tell them what’s wrong with the LLM output.

      It takes all the joy out of the job too, something they’ve been working on for years. What makes this work interesting is understanding people’s problems, working out the best way to model them, and building towards solutions. What they want the job to be is a slop factory: same as the dream of every rich asshole who thinks having half an idea is the same as working for years to fully realize an idea in all it’s complexity and wonder.

      They never have any respect for the work that takes because they’ve never done any work. And the next generation of implementers are being taught that there are no new ideas. You just ask the oracle to give you the answer.

    • @swlabr
      link
      English
      1318 days ago

      When they say “art” they mean “metaphorical lead paint” and when they say “accessible” they mean “insidiously inserted into your neural pathways”

    • @Croquette@sh.itjust.works
      link
      fedilink
      English
      1218 days ago

      Art is already accessible. Plenty of artists that sells their art dirt cheap, or you can buy pen and papers at the dollar store.

      What people want when they say “AI is making art accessible” is they want high quality professional art for dirt cheap.

      • @scruiser
        link
        English
        918 days ago

        I think they also want recognition/credit for spending 5 minutes (or less) typing some words at an image generator as if that were comparable to people who develop technical skills and then create effortful meaningful work just because the outputs are (superficially) similar.

      • Schadrach
        link
        fedilink
        English
        -417 days ago

        What people want when they say “AI is making art accessible” is they want high quality professional art for dirt cheap.

        …and what their opposition means when they oppose it is “this line of work was supposed to be totally immune to automation, and I’m mad that it turns out not to be.”

        • @zbyte64
          link
          English
          417 days ago

          …and this opposition means that our disagreements can only be perceived through the lens of personal faults.

        • @Croquette@sh.itjust.works
          link
          fedilink
          English
          417 days ago

          There is already a lot of automation out there, and more is better, when used correctly. And that’s not talking about the outright theft of the material from these artists it is trying to replace so badly.

        • @YourNetworkIsHaunted
          link
          English
          315 days ago

          See I would frame it as practicioners of some of the last few non-bullshit jobs (minimally bullshit jobs) - fields that by necessity require a kind of craft or art that is meaningful or rewarding - being routed around by economic forces that only wanted their work for bullshit results. Like, no matter how passionate you are about graphic design you probably didn’t get into the field because shuffling the visuals every so often is X% better for customer engagement and conversion or whatever. But the businesses buying graphic design work are more interested in that than they ever were in making something beautiful or functional, and GenAI gives them the ability to get what they want more cheaply. As an unexpected benefit they also don’t have to see you roll your eyes when they tell you it needs to be “more blue” and as an insignificant side effect it brings our culture one step closer to finally drowning the human soul in shit to advance the cause of glorious industry in it’s unceasing march to An Even Bigger Number.

    • @blarghly@lemmy.world
      link
      fedilink
      English
      318 days ago

      That should be a big warning sign that the next generation of developers are not going to be very good.

      Sounds like job security to me!

      • @froztbyte
        link
        English
        1518 days ago

        “I want the people I teach to be worse than me” is a fucking nightmare of a want, I hope you learn to do better

        • @blarghly@lemmy.world
          link
          fedilink
          English
          018 days ago

          So there’s this new thing they invented. It’s called a joke. You should try them out sometime, they’re fun!

          • @blakestaceyA
            link
            English
            1118 days ago

            So, there’s this new phenomenon they’ve observed in which text does not convey tone. It can be a real problem, especially when a statement made by one person as a joke would be made by another in all seriousness — but don’t worry, solutions have very recently been proposed.

            • @froztbyte
              link
              English
              418 days ago

              space alien technology!!~

            • @blarghly@lemmy.world
              link
              fedilink
              English
              -418 days ago

              I dunno what kind of world you are living in where someone would make my comment not as a joke. Please find better friends.

              • @selfA
                link
                English
                1118 days ago

                you’re as funny as the grave

          • @froztbyte
            link
            English
            1018 days ago

            “oh shit I got called out on my shitty haha-only-serious comment, better pretend I didn’t mean it!” cool story bro

            • @blarghly@lemmy.world
              link
              fedilink
              English
              -318 days ago

              If people say that sort of thing around you not as a joke, you need to spend your time with better people. I dunno what to tell you - humor is a great way to deal with shitty things in life. Dunno why you would want to get rid of it.

              • @froztbyte
                link
                English
                8
                edit-2
                18 days ago

                jesus fuck how do you fail to understand any post of this kind this badly

                • @swlabr
                  link
                  English
                  918 days ago

                  “How dare you not find me funny. I’m going to lecture you on humor. The lectures will continue until morale improves.”

                • @froztbyte
                  link
                  English
                  818 days ago

                  maybe train your model better! I know I know, they were already supposed to be taking over the world… alas…

    • @Dragonstaff@leminal.space
      link
      fedilink
      English
      -318 days ago

      I dunno. I feel like the programmers who came before me could say the same thing about IDEs, Stack Overflow, and high level programming languages. Assembly looks like gobbledygook to me and they tell me I’m a Senior Dev.

      If someone uses ChatGPT like I use StackOverflow, I’m not worried. We’ve been stealing code from each other since the beginning.“Getting the answer” and then having to figure out how to plug it into the rest of the code is pretty much what we do.

      There isn’t really a direct path from an LLM to a good programmer. You can get good snippets, but “ChatGPT, build me a app” will be largely useless. The programmers who come after me will have to understand how their code works just as much as I do.

      • @selfA
        link
        English
        1318 days ago

        fuck almighty I wish you and your friends would just do better

      • @Croquette@sh.itjust.works
        link
        fedilink
        English
        -218 days ago

        LLM as another tool is great. LLM to replace experienced coders is a nightmare waiting to happen.

        IDEs, stack overflow, they are tools that makes the life of a developers a lot easier, they don’t replace him.

        • @YourNetworkIsHaunted
          link
          English
          315 days ago

          I mean past a certain point LLMs are strictly worse tools than Stack Overflow was on its worst day. IDEs have a bunch of features to help manage complexity and offload memorization. The fundamental task of understanding the code you’re writing is still yours. Stack Overflow and other forums are basically crowdsourced mentorship programs. Someone out there knows the thing you need to and rather than cultivate a wide social network you can take advantage of mass communication. To use it well you still need to know what’s happening, and if you don’t you can at least trust that the information is out there somewhere that you might be able to follow up on as needed. LLM assistants are designed to create output that looks plausible and to tell the user what they want to hear. If the user is an idiot the LLM will do nothing to make them recognize that they’re doing something wrong, much less help them fix it.

          • @Croquette@sh.itjust.works
            link
            fedilink
            English
            -415 days ago

            LLM are terrible because the data they were trained on is garbage, because companies don’t want to pay for people to create a curated dataset to produce acceptable results.

            The tech itself can be good in specific cases. But the way it is shoved in everything right now is terrible

            • @froztbyte
              link
              English
              4
              edit-2
              15 days ago

              weren’t you also here having shitty opinions like a week ago?

              e: yes

                • @bitofhope
                  link
                  English
                  514 days ago

                  What stereotype? The stereotype that awful.systems posters are hostile to people who praise LLMs? Good.

            • @selfA
              link
              English
              314 days ago

              it can’t be that stupid, you must be training it wrong

    • @rumba@lemmy.zip
      link
      fedilink
      English
      -918 days ago

      All the newbs were just copying lines from stackexchange before AI. The only real difference at this point is that the commenting is marginally better.

      • @frezik@midwest.social
        link
        fedilink
        English
        14
        edit-2
        18 days ago

        Stack Overflow is far from perfect, but at least there is some level of vetting going on before it’s copypasta’d.

  • @DarkCloud@lemmy.world
    link
    fedilink
    English
    4919 days ago

    Hot take, people will look back on anyone who currently codes, as we look back on the NASA programmers who got the equipment and people to the moon.

    They won’t understand how they did so much with so little. You’re all gourmet chefs in a future of McDonalds.

    • shnizmuffin
      link
      fedilink
      English
      5118 days ago

      Nah, we’re plumbers in an age where everyone has decided to DIY their septic system.

      Please, by all means, keep it up.

    • @BlueMonday1984
      link
      English
      1518 days ago

      Hot take, people will look back on anyone who currently codes, as we look back on the NASA programmers who got the equipment and people to the moon.

      I doubt it’ll be anything that good for them. By my guess, those who currently code are at risk of suffering some guilt-by-association problems, as the AI bubble paints them as AI bros by proxy.

      • @Architeuthis
        link
        English
        12
        edit-2
        18 days ago

        I think most people will ultimately associate chatbots with corporate overreach rather rank-and-file programmers. It’s not like decades of Microsoft shoving stuff down our collective throat made people think particularly less of programmers, or think about them at all.

      • @V0ldek
        link
        English
        718 days ago

        Meh, I have so many bangers laughing at actual AI bros that I could make my CV just all be sneers on them, I think this particular corner of the internet is quite safe

    • @corbin
      link
      English
      1218 days ago

      Perhaps! But not because we adopted vibe coding. I do have faith in our ability to climb out of the Turing tarpit (WP, Esolangs) eventually, but only by coming to a deeper understanding of algorithmic complexity.

      Also, from a completely different angle: when I was a teenager, I could have a programmable calculator with 18MHz Z80 in my hand for $100. NASA programmers today have the amazing luxury of the RAD750, a 110MHz PowerPC chipset. We’re already past the gourmet phase and well into fusion.

      • @swlabr
        link
        English
        718 days ago

        NASA programmers grow more powerful by the day. It’s only a matter of time before they reach AGI

    • @DarkCloud @dgerard

      The first commercial product I worked on had 128 bytes of RAM and 2048 bytes of ROM.

      It kept people safe from low oxygen, risk of explosions, and toxic levels of two poisonous gases including short term and long term effects at fifteen minutes and eight hour averages.

      Pre-Internet. When you’re doing something new or pushing the limits, you just have to know how to code and read the datasheets.

    • @V0ldek
      link
      English
      7
      edit-2
      18 days ago

      My first actual real life project was building a data analytics platform while keeping the budget to a minimum. With some clever parallelism and aggressive memory usage optimisation I made it work on a single lowest-tier Azure VM, costing like $50 to run monthly, while the measurable savings for the business from using the platform are now measured in the millions.

      Don Knuth didn’t write all those volumes on how software is an art for you to use fucking Node.JS you rubes, you absolute clowns

  • @AeonFelis@lemmy.world
    link
    fedilink
    English
    4618 days ago

    I got an AI PR in one of my projects once. It re-implemented a feature that already existed. It had a bug that did not exist in the already-existing feature. It placed the setting for activating that new feature right after the setting for activating the already-existing feature.

  • @BlueMonday1984
    link
    English
    4518 days ago

    Baldur Bjarnason’s given his thoughts on Bluesky:

    My current theory is that the main difference between open source and closed source when it comes to the adoption of “AI” tools is that open source projects generally have to ship working code, whereas closed source only needs to ship code that runs.

    I’ve heard so many examples of closed source projects that get shipped but don’t actually work for the business. And too many examples of broken closed source projects that are replacing legacy code that was both working just fine and genuinely secure. Pure novelty-seeking

  • @swlabr
    link
    English
    3818 days ago

    The headlines said that 30% of code at Microsoft was AI now! Huge if true!

    Something like MS word has like 20-50 million lines of code. MS altogether probably has like a billion lines of code. 30% of that being AI generated is infeasible given the timeframe. People just ate this shit up. AI grifting is so fucking easy.

    • @froztbyte
      link
      English
      1318 days ago

      yeah, the “some projects” bit is applicable, as is the “machine generated” phrasing

      @gsuberland pointed out elsewhere on fedi just how much of the VS-/MS- ecosystem does an absolute fucking ton of code generation

      (which is entirely fine, ofc. tons of things do that and it exists for a reason. but there’s a canyon in the sand between A and B)

      • @swlabr
        link
        English
        1318 days ago

        All compiled code is machine generated! BRB gonna clang and IPO, bye awful.systems! Have fun being poor

            • @froztbyte
              link
              English
              1018 days ago

              (not in the compute side, but in the lying-obstructionist hustle side)

              • @swlabr
                link
                English
                918 days ago

                would I happier if I abandoned my scruples? I hope I or nobody I know finds out.

                • @frezik@midwest.social
                  link
                  fedilink
                  English
                  918 days ago

                  For some definition of “happiness”, yes. It’s increasingly clear that the only way to get ahead is with some level of scam. In fact, I’m pretty sure Millennials will not be able to retire to a reasonable level of comfort without accepting some amount of unethical behavior to get there. Not necessarily Slipp’n Jimmy levels of scam, but just stuff like participating in a basic stock market investment with a tax advantaged account.

    • @blarghly@lemmy.world
      link
      fedilink
      English
      1118 days ago

      I thought it could totally be true - that devs at MS were just churning out AI crap code like there was no tomorrow, and their leaders were cheering on their “productivity”, since more code = more better, right?

      • @swlabr
        link
        English
        718 days ago

        From that angle, sure. I’m more sneering at the people who saw what they wanted to see, and the people that were saying “this is good, actually!!!”

    • @Dragonstaff@leminal.space
      link
      fedilink
      English
      -418 days ago

      30% of code is standard boilerplate: setters, getters, etc that my IDE builds for me without calling it AI. It’s possible the claim is true, but it’s terribly misleading at best.

      • @swlabr
        link
        English
        13
        edit-2
        18 days ago
        1. Perhaps you didn’t read the linked article. Nadella didn’t claim that 30% of MS’s code was written by AI. What he said was garbled up to the eventual headline.
        2. We don’t have to play devil’s advocate for a hyped-up headline that misquotes what an AI glazer said, dawg.
        3. “Existing code generation codes can write 30%” doesn’t imply that AI possibly/plausibly wrote 30% of MS’s code. There’s no logical connection. Please dawg, I beg you, think critically about this.
          • @swlabr
            link
            English
            918 days ago

            Man. If this LLM stuff sticks around, we’ll have an epidemic of early onset dementia.

            • @Soyweiser
              link
              English
              1018 days ago

              If the stories lf covid related cognitive decline are aue we are going to have a great time. Worse than lead paint.

              • @swlabr
                link
                English
                1118 days ago

                “Oh man, this brain fog I have sure makes it hard to think. Guess I’ll use my trusty LLM! ChatGPT says lead paint is tastier and better for your brain than COVID? Don’t mind if I do!”

                • @Soyweiser
                  link
                  English
                  718 days ago

                  I’m on a diet of rocks, glue on my pizza, lead paint, and covid infections, according to Grok this is called the Mr Burns method which should prevent diseases, as they all work together to block all bad impulses. Can’t wait to try this new garlic oil I made, using LLM instructions. It even had these cool bubbles while fermenting, nature is great.

            • @froztbyte
              link
              English
              718 days ago

              I’ve been beating this drum for like 4~5y but: I don’t think the tech itself is going anywhere. published, opensourced, etc etc - the bell can’t be unrung, the horses have departed the stable

              but

              I do also argue that an extremely large amount of wind in the sails right now is because of the constellation of VC/hype//etc shit

              can’t put a hard number on this, but … I kind see a very massive reduction; in scope, in competence, in relevance. so much of this shit (esp. the “but my opensource model is great!” flavour) is so fucking reliant on “oh yeah this other entity had a couple fuckpiles of cash with which to train”, and once that (structurally) evaporates…

  • @vivendi@programming.dev
    link
    fedilink
    English
    3418 days ago

    No the fuck it’s not

    I’m a pretty big proponent of FOSS AI, but none of the models I’ve ever used are good enough to work without a human treating it like a tool to automate small tasks. In my workflow there is no difference between LLMs and fucking grep for me.

    People who think AI codes well are shit at their job

    • @V0ldek
      link
      English
      3018 days ago

      In my workflow there is no difference between LLMs and fucking grep for me.

      Well grep doesn’t hallucinate things that are not actually in the logs I’m grepping so I think I’ll stick to grep.

      (Or ripgrep rather)

        • @froztbyte
          link
          English
          1118 days ago

          (I don’t mean to take aim at you with this despite how irked it’ll sound)

          I really fucking hate how many computer types go “ugh I can’t” at regex. the full spectrum of it, sure, gets hairy. but so many people could be well served by decently learning grouping/backrefs/greedy match/char-classes (which is a lot of what most people seem to reach for[0])

          that said, pomsky is an interesting thing that might in fact help a lot of people go from “I want $x” as a human expression of intent, to “I have $y” as a regex expression

          [0] - yeah okay sometimes you also actually need a parser. that’s a whole other conversation. I’m talking about “quickly hacking shit up in a text editor buffer in 30s” type cases here

          • @swlabr
            link
            English
            8
            edit-2
            18 days ago

            Hey. I can do regex. It’s specifically grep I have beef with. I never know off the top of my head how to invoke it. Is it -e? -r? -i? man grep? More like, man, get grep the hell outta here!

            • @froztbyte
              link
              English
              1218 days ago

              now listen, you might think gnu tools are offensively inconsistent, and to that I can only say

              find(1)

              • @swlabr
                link
                English
                1218 days ago

                find(1)? You better find(1) some other place to be, buster. In this house, we use the file explorer search bar

              • @swlabr
                link
                English
                418 days ago

                If I start using this and add grep functionality to my day-to-day life, I can’t complain about not knowing how to invoke grep in good conscience, dawg. I can’t hold my shitposting back like that, dawg!

                jk that looks useful. Thanks!

                • @lagoon8622@sh.itjust.works
                  link
                  fedilink
                  English
                  418 days ago

                  The cheatsheet and tealdeer projects are awesome. It’s one of my (many) favorite things about the user experience honestly. Really grateful for those projects

            • @froztbyte
              link
              English
              418 days ago

              woo! but still also check out pomsky, it’s legit handy!

              • @froztbyte
                link
                English
                318 days ago

                (also I did my disclaimer at the start there, so, y’know (but also igwym))

      • @vivendi@programming.dev
        link
        fedilink
        English
        -918 days ago

        Hallucinations become almost a non issue when working with newer models, custom inference, multishot prompting and RAG

        But the models themselves fundamentally can’t write good, new code, even if they’re perfectly factual

        • @Architeuthis
          link
          English
          1718 days ago

          If LLM hallucinations ever become a non-issue I doubt I’ll be needing to read a deeply nested buzzword laden lemmy post to first hear about it.

          • @vivendi@programming.dev
            link
            fedilink
            English
            -718 days ago

            You need to run the model yourself and heavily tune the inference, which is why you haven’t heard from it because most people think using shitGPT is all there is with LLMs. How many people even have the hardware to do so anyway?

            I run my own local models with my own inference, which really helps. There are online communities you can join (won’t link bcz Reddit) where you can learn how to do it too, no need to take my word for it

            • @selfA
              link
              English
              918 days ago

              ah yes, the problem with cryptoLLMs is all the shitcoinsGPTs

              did it sting when the crypto bubble popped? is that what made you like this?

            • @Architeuthis
              link
              English
              517 days ago

              You run CanadianGirlfriendGPT, got it.

        • @scruiser
          link
          English
          1318 days ago

          The promptfarmers can push the hallucination rates incrementally lower by spending 10x compute on training (and training on 10x the data and spending 10x on runtime cost) but they’re already consuming a plurality of all VC funding so they can’t 10x many more times without going bust entirely. And they aren’t going to get them down to 0%, hallucinations are intrinsic to how LLMs operate, no patch with run-time inference or multiple tries or RAG will eliminate that.

          And as for newer models… o3 actually had a higher hallucination rate because trying to squeeze rational logic out of the models with fine-tuning just breaks them in a different direction.

          I will acknowledge in domains with analytically verifiable answers you can check the LLMs that way, but in that case, its no longer primarily an LLM, you’ve got an entire expert system or proof assistant or whatever that can operate independently of the LLM and the LLM is just providing creative input.

          • @swlabr
            link
            English
            1218 days ago

            We should maximise hallucinations, actually. That is, we should hack the environmental controls of the data centers to be conducive for fungi growth, and flood them with magic mushrooms spores. We can probably get the rats on board by selling it as a different version of nuking the data centers.

          • @vivendi@programming.dev
            link
            fedilink
            English
            -7
            edit-2
            18 days ago

            O3 is trash, same with closedAI

            I’ve had the most success with Dolphin3-Mistral 24B (open model finetuned on open data) and Qwen series

            Also lower model temperature if you’re getting hallucinations

            For some reason everyone is still living in 2023 when AI is remotely mentioned. There is a LOT you can criticize LLMs for, some bullshit you regurgitate without actually understanding isn’t one

            You also don’t need 10x the resources where tf did you even hallucinate that from

            • David Gerard
              shield
              OPMA
              link
              English
              1018 days ago

              this user has been escorted off the premises via the fourth floor window

              • @froztbyte
                link
                English
                817 days ago

                the Russian Exit, a classic

            • @froztbyte
              link
              English
              918 days ago

              this isn’t the place to decide which seed generator you want for your autoplag runtime

              • @vivendi@programming.dev
                link
                fedilink
                English
                -618 days ago

                My most honest goal is to educate people which on lemmy is always met with hate. people love to hate, parroting the same old nonsense that someone else taught them.

                If you insist on ignorance then be ignorant in peace, don’t try such misguided attempts at sneer

                There are things in which LLMs suck. And there are things that you wrongly believe as part of this bullshit twitter civil war.

                • @froztbyte
                  link
                  English
                  818 days ago

                  My most honest goal is to educate people

                  oh and I suppose you can back that up with verifiable facts, yes?

                  and that you, yourself, can stand as a sole beacon against the otherwise regularly increasing evidence and studies that both indicate toward and also prove your claims to be full of shit? you are the saviour that can help enlighten us poor unenlightened mortals?

                  sounds very hard. managing your calendar must be quite a skill

            • @scruiser
              link
              English
              8
              edit-2
              18 days ago

              GPT-1 is 117 million parameters, GPT-2 is 1.5 billion parameters, GPT-3 is 175 billion, GPT-4 is undisclosed but estimated at 1.7 trillion. Token needed for training and training compute scale linearly (edit: actually I’m wrong, looking at the wikipedia page… so I was wrong, it is even worse for your case than I was saying, training compute scales quadratically with model size, it is going up 2 OOM for every 10x of parameters) with model size. They are improving … but only getting a linear improvement in training loss for a geometric increase in model size, training time. A hypothetical GPT-5 would have 10 trillion training parameters and genuinely need to be AGI to have the remotest hope of paying off it’s training. And it would need more quality tokens than they have left, they’ve already scrapped the internet (including many copyrighted sources and sources that requested not to be scrapped). So that’s exactly why OpenAI has been screwing around with fine-tuning setups with illegible naming schemes instead of just releasing a GPT-5. But fine-tuning can only shift what you’re getting within distribution, so it trades off in getting more hallucinations or overly obsequious output or whatever the latest problem they are having.

              Lower model temperatures makes it pick it’s best guess for next token as opposed to randomizing among probable guesses, they don’t improve on what the best guess is and you can still get hallucinations even picking the “best” next token.

              And lol at you trying to reverse the accusation against LLMs by accusing me of regurgitating/hallucinating.

              • @vivendi@programming.dev
                link
                fedilink
                English
                -718 days ago

                Small scale models, like Mistral Small or Qwen series, are achieving SOTA performance with lower than 50 billion parameters. QwQ32 could already rival shitGPT with 32 billion parameters, and the new Qwen3 and Gemma (from google) are almost black magic.

                Gemma 4B is more comprehensible than GPT4o, the performance race is fucking insane.

                ClosedAI is 90% hype. Their models are benchmark princesses, but they need huuuuuuge active parameter sizes to effectively reach their numbers.

                Everything said in this post is independently verifiable by taking 5 minutes to search shit up, and yet you couldn’t even bother to do that.

          • @swlabr
            link
            English
            918 days ago

            God, this cannot be overstated. An LLM’s sole function is to hallucinate. Anything stated beyond that is overselling.

      • @vivendi@programming.dev
        link
        fedilink
        English
        -4
        edit-2
        18 days ago

        These views on LLMs are simplistic. As a wise man once said, “check yoself befo yo wreck yoself”, I recommend more education thus

        LLM structures arw over hyped, but they’re also not that simple

        • @MonkderVierte@lemmy.ml
          link
          fedilink
          English
          3
          edit-2
          18 days ago

          From what i know from recent articles about retracing LLM indepth, they are indeed best suited for language translation and perfectly explain the halucinations. And i think i’ve read somewhere that this was the originally intended purpose of the tech?

          Ah, here, and here more tabloid-ish.

          • @froztbyte
            link
            English
            518 days ago

            many of the proponents of things in this field will propose/argue $x thing to be massively valuable for $x

            thing is, that doesn’t often work out

            yes, there’s some value in the tech for translation outcomes. to anyone even mildly online, “so are language teaching apps/sites using this?” is probably a very nearby question. and rightly so!

            and then when you go digging into how that’s going in practice, wow fuck damn doesn’t that Glorious AI Future sheen just fall right off…

      • @Blackmist@feddit.uk
        link
        fedilink
        English
        2118 days ago

        I’m guessing if it would actually work for that, somebody would have done it by now.

        But it probably just does it’s usual thing of bullshitting something that looks like code, only now you’re wasting the time of maintainers as well who have to confirm that it is bobbins.

        • @gens@programming.dev
          link
          fedilink
          English
          1718 days ago

          Yea it’s a problem already for security bugs, llms just waste maintainers time and make them angry.

          They are useless and make more work for programmers, even on python and js codebases that they are trained on the most and are the “easiest”.

        • Natanox
          link
          fedilink
          English
          418 days ago

          It’s already doing that, some FOSS projects regularly get weird PRs that on first glance look good, but if you look closer are either total nonsense or riddled with bugs. Especially awful are security-related PRs; although those are never made in good faith, that’s usually grifting (throwing AI at the wall trying to cash in as many bounties as possible). The project lead of curl recently announced that anyone who posts a PR that’s obviously AI, or is made with AI, will get banned.

          Like, it’s really good as a learning tool as long as you don’t blindly believe everything it says given you can ask stuff in natural language and it will resolve possible knowledge dependencies for you that you’d otherwise get stuck on in official docs, and since you can ask contextual questions receiving contextual answers (no logical abstraction). But code generation… please don’t.

          • @froztbyte
            link
            English
            1718 days ago

            it’s really good as a learning tool as long as you don’t blindly believe everything it says given you can ask stuff in natural language

            the poster: “it’s really good as a learning tool”

            the poster: “but don’t blindly believe it”

            the learner: “how should I know when to believe it?”

            the poster: “check everything”

            the learner: “so you’re saying I should just read the actual documentation and/or source?”

            the poster: “how are you going to ask that anything? how can you fondle something that isn’t a prompt?!”

            the learner: “thanks for your time, I think I’m going to find another class”

              • @froztbyte
                link
                English
                1018 days ago

                Nice conversation you had right there in your head

                that you recognize none of this is telling. that someone else got it, more so.

                I assume

                you could just ask, you know. since you seem so comfortable fondling prompts, not sure why you wouldn’t ask a person. is it because they might tell you to fuck off?

                I’ve taken a closer look…

                fuck off with the unrequested advertising. never mind that no-one asked you for how you felt for some fucking piece of shit. oh, you feel happy that the logo is a certain tint of <colour>? bully for you, now fuck off and do something worthwhile

                That makes it a good tool

                a tool you say? wow, sure glad you’re going to replace your *spins the wheel* Punctured Car Tyre with *spins the wheel again* Needlenose Pliers!

                think I’m some AI worshipper, fuck no. They’re amoral as fuck

                so, you think there’s moral problems, but only sometimes? it’s supes okay to do your version of leveraged exploitation? cool, thanks for letting us know

                those very few truly FOSS ones

                oh yeah, right, the “truly FOSS ones”! tell me again how those are trained - who’s funding that compute? are the licenses contextually included in the model definition?

                wait, hold on! why are you squealing away like a deflating balloon?! those are actual questions! you’re the one who brought up morals!

                Otherwise you’ll end up in a social corner filled with bitterness

                I’ve met people like you at parties. they’re often popular, but they’re never fun. and I always regret it.

                There are technologies that are utter bullshit like NFTs. However (unfortunately?) that isn’t the case for AI

                citation. fucking. needed.

                • Natanox
                  link
                  fedilink
                  English
                  -918 days ago

                  Holy shit, get some help. Given how nonsensically off-the-rails you just went you clearly need it.

              • @swlabr
                link
                English
                818 days ago

                Otherwise you’ll end up in a social corner filled with bitterness

                This is a standard Internet phenomenon (I generalize) called a Sneer Club, i.e. people who enjoy getting together and picking on designated targets. Sneer Clubs (I expect) attract people with high Dark Triad characteristics, which is (I suspect) where Asshole Internet Atheists come from - if you get a club together for the purpose of sneering at religious people, it doesn’t matter that God doesn’t actually exist, the club attracts psychologically f’d-up people. Bullies, in a word, people who are powerfully reinforced by getting in what feels like good hits on Designated Targets, in the company of others doing the same and congratulating each other on it.

              • @selfA
                link
                English
                518 days ago

                holy fuck this is so many words to say so little

                so congrats I’m upgrading your ban and also pruning you from the thread

                • @froztbyte
                  link
                  English
                  518 days ago

                  on the one hand I feel for other people who’ll maybe read this thread somewhen down the line

                  on the other, it’s not exactly like I clipped words in my post

          • @V0ldek
            link
            English
            1718 days ago

            Like, it’s really good as a learning tool

            Fuck you were doing so well in the first half, ahhh,

          • @V0ldek
            link
            English
            1418 days ago

            Hey, Devin! Really impressive that the product best known for literally lying about all of its functionality in its release video still somehow exists and you can pay it money. Isn’t the free market great.

            • @froztbyte
              link
              English
              618 days ago

              “a fool and their money are soon parted”

            • @swlabr
              link
              English
              518 days ago

              Devin? Fuck Devin. That slimy motherfucker owes me 10 bucks.

          • @froztbyte
            link
            English
            1318 days ago

            fuck off with the unrequested advertising kthx

          • @blakestaceyA
            link
            English
            918 days ago

            Banned from the community for advertising.

  • @LucidLyes@lemmy.world
    link
    fedilink
    English
    2518 days ago

    The only people impressed by AI code are people who have the level to be impressed by AI code. Same for AI playing chess.

  • @swlabr
    link
    English
    2518 days ago

    Good hustle Gerard, great job starting this chudstorm. I’m having a great time

    • @froztbyte
      link
      English
      1918 days ago

      the prompt-related pivots really do bring all the chodes to the yard

      and they’re definitely like “mine’s better than yours”

      • @scruiser
        link
        English
        16
        edit-2
        18 days ago

        The latest twist I’m seeing isn’t blaming your prompting (although they’re still eager to do that), it’s blaming your choice of LLM.

        “Oh, you’re using shitGPT 4.1-4o-o3 mini _ro_plus for programming? You should clearly be using Gemini 3.5.07 pro-doubleplusgood, unless you need something locally run, then you should be using DeepSek_v2_r_1 on your 48 GB VRAM local server! Unless you need nice sounding prose, then you actually need Claude Limmerick 3.7.01. Clearly you just aren’t trying the right models, so allow me to educate you with all my prompt fondling experience. You’re trying to make some general point? Clearly you just need to try another model.”

      • @swlabr
        link
        English
        9
        edit-2
        18 days ago

        Prompt-Pivots: Prime Sea-lion Siren Song! More at 8.

    • David GerardOPMA
      link
      English
      1718 days ago

      this post has also broken containment in the wider world, the video’s got thousands of views, I got 100+ subscribers on youtube and another $25/mo of patrons

      • @swlabr
        link
        English
        1218 days ago

        We love to see it

    • @scruiser
      link
      English
      918 days ago

      Posts that explode like this are fun and yet also a reminder why the banhammer is needed.

      • @swlabr
        link
        English
        818 days ago

        Unlike the PHP hammer, the banhammer is very useful for a lot of things. Especially sealion clubbing.

  • @MNByChoice@midwest.social
    link
    fedilink
    English
    2418 days ago

    Damn, this is powerful.

    If AI code was great, and empowered non-programmers, then open source projects should have already committed hundreds of thousands of updates. We should have new software releases daily.