• @apfelwoiSchoppen@lemmy.world
    link
    fedilink
    98
    edit-2
    6 days ago

    I ordered some well rated concert ear protection from the maker’s website. The order waited weeks to ship after a label was printed and likely forgotten. I went to find a place to call or contact a human there, all they had was a self-described AI chat robot that just talked down to me condescendingly. It simply would not believe my experience.

    I eventually got the ear protection but I won’t be buying from them again. Can’t even staff some folks to check email. I eventually found their PR email address but even that was outsourced to a PR firm that never got back to me. Utter shit, AI.

    • @zod000@lemmy.ml
      link
      fedilink
      385 days ago

      I’m glad you mentioned the company directly as I also want to steer clear of companies like this.

    • @Dozzi92@lemmy.world
      link
      fedilink
      315 days ago

      That would’ve been such an easy disputed charge and get the plugs somewhere else. I’m not wasting a second on something like that, just telling my credit card company they didn’t uphold their end of the deal, and that’s that. I will lose hearing out of spite if this happened to me, because I’m an idiot.

      • misterdoctor
        link
        fedilink
        125 days ago

        I will lose hearing out of spite if this happened to me

        Genuinely admire your self awareness

      • @brbposting@sh.itjust.works
        link
        fedilink
        English
        24 days ago

        I’ll waste a few moments. It becomes a puzzle. Assuming you managed to make it through the maze, you retrospectively analyze where would 99% of the country have dropped out of the flow and given up?

        Then it’s an email to the attorney general if necessary! (I mean that’s been rare but when something is egregious)

        🤓

    • @crank0271@lemmy.world
      link
      fedilink
      English
      116 days ago

      That’s really good to know about these things. They’ve been on sale through Woot. I guess there’s a good reason for that.

      • @Andonyx@lemmy.world
        link
        fedilink
        65 days ago

        A number of companies make “tuned” ear plugs to allow some sound through with a desired frequency curve, but reduce SPL to safe levels. I’ve used Etymotic, which sound great but I personally like a little more reduction. Alpine which I thought had enough reduction but too much coloring, and I settled on Earpeace, for like $25 on-line. Silicone, re-usable and easy to clean and they come with three filters to swap in or out depending on your needs / tastes.

    • Wow, that’s extremely disappointing. I had a really positive experience with them a few years ago when I wanted to exchange what I got (it was too quiet for me), and they just sent me a free pair after I talked to an actual person on their chat thing. It’s good to know that’s not how they are anymore if I ever need to replace them.

    • @lapping6596@lemmy.world
      link
      fedilink
      35 days ago

      Oh man, sad that’s the customer service cause I deeply love my loops. I was already carrying them with me everywhere I went so I grabbed a pill keychain thing and attached them to my keys so I’d never forget to grab them.

      • Yeah this happened back earlier this year. I had lost a pair from a purchase years ago and replaced them. Guessing they are laying off people/support contracts like so many stupid business owners. I was sure that my order would be stuck in limbo forever after the experience, but they eventually showed up. Never again.

  • @roude@lemmynsfw.com
    link
    fedilink
    English
    50
    edit-2
    5 days ago

    Alright I don’t like the direction of AI same as the next person, but this is a pretty fucking wild stance. There are multiple valid applications of AI that I’ve implemented myself: LTV estimation, document summary / search / categorization, fraud detection, clustering and scoring, video and audio recommendations… "Using AI” is not the problem, “AI charlatan-ing” is. Or in this guy’s case, “wholesale anti-AI stanning”. Shoehorning AI into everything is admittedly a waste, but to write off the entirety of a very broad category (AI) is just silly.

    • @jjjalljs@ttrpg.network
      link
      fedilink
      505 days ago

      I don’t think AI is actually that good at summarizing. It doesn’t understand the text and is prone to hallucinate. I wouldn’t trust an AI summary for anything important.

      Also search just seems like overkill. If I type in “population of london”, i just want to be taken to a reputable site like wikipedia. I don’t want a guessing machine to tell me.

      Other use cases maybe. But there are so many poor uses of AI, it’s hard to take any of it seriously.

      • I don’t think AI is actually that good at summarizing. It doesn’t understand the text and is prone to hallucinate. I wouldn’t trust an AI summary for anything important.

        This right here. Whenever I’ve tried using an LLM to summarize, I spent more time fact-checking it (and finding the inevitable misunderstandings and outright hallucinations—they’re always there for anything of substance!) than I’d spend writing my own damned summary.

        There is, however, one use case I’ve found where LLMs work better than alternatives … provided you do due diligence. To put it bluntly, Google Translate and its ilk of similar slop from Bing, Baidu, etc. suck. They are god-awful at translation of anything but straightforward technical writing or the most tediously dull prose. LLMs are far better translators (and can be instructed to highlight cultural artifacts, possible transcription errors, etc.) …

        … as long as you back-translate in a separate session to check for hallucination.

        Oh, and Google Translate-style translators really suck at Classical Chinese. LLMs do much better (provided you do the back-translation check for hallucination).

        • @brognak@lemm.ee
          link
          fedilink
          35 days ago

          If I understand how AI works (predictive models), kinda seems perfectly suited for translating text. Also exactly how I have been using it with Gemini, translate all the memes in ich_iel 🤣. Unironically it works really well, and the only ones that aren’t understandable are cultural not linguistic.

            • @brognak@lemm.ee
              link
              fedilink
              1
              edit-2
              4 days ago

              Oh that’s the best part, since it’s memes honestly never know if it’s even meant to be completely sensible. So even if it does hallucinate, just adds a bit of spice 🤌🤌

              I also like the thought that probably billions was spent to make something that is best suited for deep frying memes

      • ArchRecord
        link
        fedilink
        English
        5
        edit-2
        5 days ago

        I don’t think AI is actually that good at summarizing.

        It really depends on the type and size of text you want it to summarize.

        For instance, it’ll only give you a very, very simplistic overview of a large research paper that uses technical terms, but if you want to to compress down a bullet point list, or take one paragraph and turn it into some bullet points, it’ll usually do that without any issues.

        Edit: I truly don’t understand why I’m getting downvoted for this. LLMs are actually relatively good at summarizing small, low-context-necessary pieces of information into bullet points. They’re quite literally made as code that interprets the likelihood of text based on an input. Giving it a small amount of text to rewrite or recontextualize is one of its best strengths. That’s why it was originally mostly implemented as a tool to reword small isolated sections in articles, emails, and papers, before the technology was improved.

        It’s when they get to larger pieces of information, like meetings, books, wikipedia articles, etc, that they begin to break down, due to the nature of the technology itself. (context windows, lack of external resources that humans are able to integrate into their writing, but LLMs can’t fully incorporate on the same level)

          • ArchRecord
            link
            fedilink
            English
            35 days ago

            It can, but I don’t see that happen often in most places I see it used, at least by the average person, although I will say I’ve deliberately insulated myself a bit from the very AI bro type of people who use it regularly throughout their day, and mostly interact with people who are using it occasionally during research for an assignment, rewriting part of their email, etc, so I recognize that my opinion here might just be influenced by the type of uses I personally see it used for.

            In my experience, when it’s used to summarize, say, 4-6 sentences of text, in a general-audience readable text (i.e. not a research paper in a journal) that doesn’t explicitly rely on a high level of context from the rest of the text (e.g. a news article relies on information it doesn’t currently have, so a paragraph out of context would be bad, vs instructions on how to use a tool, which are general knowledge) then it seems to do pretty well, especially within the confines of an existing conversation about the topic where the intent and context has been established already.

            For example, a couple months back, I was having a hard time understanding subnetting, but I decided to give it a shot, and by giving it a bit of context on what was tripping me up, it was successfully able to reword and re-explain the topic in such a way that I was able to better understand it, and could then continue researching it.

            Broad topic that’s definitely in the training data + doesn’t rely on lots of extra context for the specific example = reasonably good output.

            But again, I also don’t frequently interact with the kind of people that like having AI in everything, and am mostly just around very casual users that don’t use it for anything very high stakes or complex, and I’m quite sure that anything more than extremely simple summaries of basic information or very well-known topics would probably have a lot of hallucinations.

              • ArchRecord
                link
                fedilink
                English
                25 days ago

                See, when I have 4-6 sentences to summarize, I don’t see the value-add of a machine doing the summarizing for me.

                Oh I completely understand, I don’t often see it as useful either. I’m just saying that a lot of people I see using LLMs occasionally are usually just shortening their own replies to things, converting a text based list of steps to a numbered list for readability, or just rewording a concept because the original writer didn’t word it in a way their brain could process well, etc.

                Things that don’t necessarily require a huge amount of effort on their part, but still save them a little bit of time, which in my conversations with them, seems to prove valuable to them, even if it’s in a small way.

                • @jjjalljs@ttrpg.network
                  link
                  fedilink
                  55 days ago

                  I feel like letting your skills in reading and communicating in writing atrophy is a poor choice. And skills do atrophy without use. I used to be able to read a book and write an essay critically analyzing it. If I tried to do that now, it would be a rough start.

                  I don’t think people are going to just up and forget how to write, but I do think they’ll get even worse at it if they don’t do it.

        • @jjjalljs@ttrpg.network
          link
          fedilink
          75 days ago

          But if the text you’re working on is small, you could just do it yourself. You don’t need an expensive guessing machine.

          Like, if I built a rube-goldberg machine using twenty rubber ducks, a diesel engine, and a blender to tie my shoes, and it gets it right most of the time, that’s impressive. but also kind of a stupid waste, because I could’ve just tied them with my hands.

          • ArchRecord
            link
            fedilink
            English
            25 days ago

            you could just do it yourself.

            Personally, I think that wholly depends on the context.

            For example, if someone’s having part of their email rewritten because they feel the tone was a bit off, they’re usually doing that because their own attempts to do so weren’t working for them, and they wanted a secondary… not exactly opinion, since it’s a machine obviously, but at least an attempt that’s outside whatever their brain might currently be locked into trying to do.

            I know I’ve gotten stuck for way too long wondering why my writing felt so off, only to have someone give me a quick suggestion that cleared it all up, so I can see how this would be helpful, while also not always being something they can easily or quickly do themselves.

            Also, there are legitimately just many use cases for applications using LLMs to parse small pieces of data on behalf of an application better than simple regex equations, for instance.

            For example, Linkwarden, a popular open source link management software, (on an opt-in basis) uses LLMs to just automatically tag your links based on the contents of the page. When I’m importing thousands of bookmarks for the first time, even though each individual task is short to do, in terms of just looking at the link and assigning the proper tags, and is not something that takes significant mental effort on its own, I don’t want to do that thousands of times if the LLM will get it done much faster with accuracy that’s good enough for my use case.

            I can definitely agree with you in a broader sense though, since at this point I’ve seen people write 2 sentence emails and short comments using AI before, using prompts even longer than the output, and that I can 100% agree is entirely pointless.

        • @Thisiswritteningerman@midwest.social
          link
          fedilink
          English
          75 days ago

          Our plant manager likes to use it to summarize meetings (Copilot). It in fact does not summarize to a bullet point list in any useful way. Breakes the notes into a headers for each topic then bullet points The header is a brief summary. The bullet points? The exact same summary but now broken by sentences as individual points. Truly stunning work. Even better with a “Please review the meeting transcript yourself as AI might not be 100% accurate” disclaimer.

          Truely worthless.

          That being said, I’ve a few vision systems using an “AI” to recognize product that doesn’t meet the pre taught pattern. It’s very good at this

          • ArchRecord
            link
            fedilink
            English
            25 days ago

            This is precisely why I don’t think anybody should be using it for meeting summaries. I know someone who does at his job, and even he only uses it for the boring, never acted upon meetings that everyone thinks is unnecessary but the managers think should be done anyways, because it just doesn’t work well enough to justify use on anything even remotely important.

            Even just from a purely technical standpoint, the context windows of LLMs are so small relative to the scale of meetings, that they will almost never be able to summarize it in its entirety without repeating points, over-explaining some topics and under-explaining others because it doesn’t have enough external context to judge importance, etc.

            But if you give it a single small paragraph from an article, it will probably summarize that small piece of information relatively well, and if you give it something already formatted like bullet points, it can usually combine points without losing much context, because it’s inherently summarizing a small, contextually isolated piece of information.

          • Lemminary
            link
            fedilink
            25 days ago

            I think your manager has a skill issue if his output is being badly formatted like that. I’d tell him to include a formatting guideline in his prompt. It won’t solve his issues but I’ll gain some favor. Just gotta make it clear I’m no damn prompt engineer. lol

            • @Thisiswritteningerman@midwest.social
              link
              fedilink
              English
              65 days ago

              I didn’t think we should be using it at all, from a security standpoint. Let’s run potentially business critical information through the plagiarism machine that Microsoft has unrestricted access to. So I’m not going to attempt to help make it’s use better at all. Hopefully if it’s trash enough, it’ll blow over once no one reasonable uses it. Besides, the man’s derided by production operators and non-kool aid drinking salaried folk He can keep it up. Lol

              • Lemminary
                link
                fedilink
                35 days ago

                Right, I just don’t want him to think that, or he’d have me tailor the prompts for him and give him an opportunity to micromanage me.

      • @roude@lemmynsfw.com
        link
        fedilink
        English
        3
        edit-2
        5 days ago

        I guess this really depends on the solution you’re working with.

        I’ve built a voting system that relays the same query to multiple online and offline LLMs and uses a consensus to complete a task. I chunk a task into smaller more manageable components, and pass those through the system. So one abstract, complex single query becomes a series of simpler asks with a higher chance of success. Is this system perfect? No, but I am not relying on a single LLM to complete it. Deficiencies in one LLM are usually made up for in at least one other LLM, so the system works pretty well. I’ve also reduced the possible kinds of queries down to a much more limited subset, so testing and evaluation of results is easier / possible. This system needs to evaluate the topic and sensitivity of millions of websites. This isn’t something I can do manually, in any reasonable amount of time. A human will be reviewing websites we flag under very specific conditions, but this cuts down on a lot of manual review work.

        When I said search, I meant offline document search. Like "find all software patents related to fly-by-wire aircraft embedded control systems” from a folder of patents. Something like elastic search would usually work well here too, but then I can dive further and get it to reason about results surfaced from the first query. I absolutely agree that AI powered search is a shitshow.

    • @Mustakrakish@lemmy.world
      link
      fedilink
      25 days ago

      I have ADHD and I have to ask A LOT of questions to get my brain around concepts sometimes, often cause I need to understand fringe cases before it “clicks”, AI has been so fucking helpful to be able to just copy a line from a textbook and say “I’m not sure what they meen by this, can you clarify” or “it says this, but also this, aren’t these two conflicting?” and having it explain has been a game changer for me. I still have to be sure to have my bullshit radar on, but thats solved by actually reading to understand and not just taking the answer as is. In fact, scrutinizing the answer against what I’ve learned and asking further questions has felt like its made me more engaged with the material.

      Most issues with AI are issues with capitalism.

      • @parody@lemmings.world
        link
        fedilink
        34 days ago

        Congratulations to the person who downvoted this

        They use a tool to improve their life?! Screw them!


        Here’s hoping over the next few years we see little baby-sized language models running on laptops entirely devour the big tech AI companies, and that those models are not only open source but ethically trained. I think that will change this community here.

        I get why they’re absolutist (AI sucks for many humans today) but above your post as well you see so much drive-by downvoting, which will obviously chill discussion.

    • Detun3d
      link
      fedilink
      1
      edit-2
      20 hours ago

      Edit for clarity: Don’t hate the science behind the tech, hate the people corrupting the tech for quick profit.

  • @romanticremedy@lemmy.blahaj.zone
    link
    fedilink
    English
    255 days ago

    I think this problem will get worse because many websites that’s used for “your own research” will lose human traffic to watch ads and more bots just scraping their data, reducing motivation to keep the websites running. Most people just take the least resistant path so AI search will be the default soon I think

    Yes, I hate this timeline

      • @romanticremedy@lemmy.blahaj.zone
        link
        fedilink
        English
        55 days ago

        Omg I can see it happening. Instead of annoying intrusive ads, this new type will be so natural as if your close friend is suggesting it.

        More dystopian future. Yes we need it /s

  • @TheObviousSolution@lemm.ee
    link
    fedilink
    115 days ago

    Using AI is telling people they shouldn’t care about your IP because you clearly don’t care about theirs when it passes through the AI lens.

  • @anachrohack@lemmy.world
    link
    fedilink
    215 days ago

    I use claude to ask it coding questions. I don’t use it to generate my code; I mostly use it to do a kind of automated code review to look for obvious pitfalls. It’s pretty neat for that

    I don’t use any other AI-powered products. I don’t let it generate emails, I don’t let it analyze data. If your site comes with a built in LLM powered feature, I assume

    1. It sucks
    2. You are a con artist

    AI is the new Crypto. If you are vaguely associated with it, I assume there’s something criminal going on

    • Lemminary
      link
      fedilink
      45 days ago

      I mostly use it to do a kind of automated code review

      Same here, especially when I’m working with plain JS. Just yesterday I was doing some benchmarking and it fixed a variable reference in my code unprompted by commenting the small fix as part of the answer when I asked it something else. I copy-pasted and it worked perfectly. It’s great for small scope stuff like that.

      But then again, I had to turn off Codeium that same day when writing documentation because it kept giving me useless and distracting, paragraph-long suggestions restating the obvious. I know it’s not meant for that, but jeez, it reminded me so much of Bing’s awfully distracting autocomplete.

      I’ve never felt this sort of technology before that, when it works, it feels like you’re gliding on ice, and when it doesn’t, it feels like ice skating on a dirt road.

    • @ameancow@lemmy.world
      link
      fedilink
      English
      15 days ago

      I use AI to script code.

      For my minecraft server.

      I rely on expert humans to do tech work for my team and their tools.

      I am not anti-AI per-say, I just know what works best and what leads to best results.

  • @zebidiah@lemmy.ca
    link
    fedilink
    75 days ago

    i use AI every day in my daily work, it writes my emails, performance reviews, project updates etc.

    …and yeah, that checks out!

  • Sunkblake
    link
    fedilink
    5
    edit-2
    4 days ago

    I used to work in a software architecture team that used AI to write retrospectives, and upcoming projects, and everything needed to have a positive spin, that sounds good but mean nothing.

    Extra funny when I find out people use AI to summarize it. So the comical cycle of bullet points to text and back again is real.

    I had enough working at the company when my team was working on the new “fantastic” platform, cut corners to reach the deadline on something that will not be used by anyone… and its being built for the explicit purpose of making a better development and working environment.

  • @gmtom@lemmy.world
    link
    fedilink
    10
    edit-2
    5 days ago

    Cool, my work for my company with AI for medical scans has detected thousands upon thousands of tumors and respiratory diseases, long before even the most well trained doctor could have spotted them, and as a result saved many of those people’s lives. But it’s good to know we’re all just lazy pieces of shit because we use AI.

    • @jjjalljs@ttrpg.network
      link
      fedilink
      175 days ago

      Assuming what you’re describing works (and i have no particular reason to doubt, beyond the generally poor reputation of AI), that’s a different beast than “lol i fired all the copywriters, artists, and support staff so I, the owner, could keep more profits for myself!”. Or, “I didn’t pay attention in English 101 and don’t know how to write, so I’ll have expensive auto suggest do it for me”

      • ArchRecord
        link
        fedilink
        English
        65 days ago

        that’s a different beast

        I think what was being implied though was that the original poster was saying that any use or talk of AI by a company immediately invalidates it, regardless of there being any specific traits like firing workers present. (e.g. “Using AI” was the only prerequisite they mentioned)

        So it seems like, based on the original wording, if they saw a hospital going “we use top of the line AI to identify tumors and respiratory diseases early” they would just disregard that hospital entirely, without actually caring how the AI works, is implemented, or affects the employment of the other people working there, even though it’s wholly beneficial.

        At least, that’s just my reading of it though.

      • @gmtom@lemmy.world
        link
        fedilink
        15 days ago

        Yeah that’s my point. AI has a lot of problems that need to be addressed, put people are getting so mad about AI that conversation around it is getting more and more extreme to the point people are talking about all AI being bad.

    • @JandroDelSol@lemmy.world
      link
      fedilink
      65 days ago

      When people talk about “AI” nowadays, they’re usually talking about LLMs and other generative AI, especially if it’s used to replace workers or human effort. Analytical AI is perfectly valid and is a wonderful tool!

      • @gmtom@lemmy.world
        link
        fedilink
        15 days ago

        No they are the same thing.

        The core algorithm we built upon is practically the same one used by AI image generators, the main difference is that we have deeper convolutional layers and more of them and we don’t do any GAN stuff that the newer image generators use.

      • ArchRecord
        link
        fedilink
        English
        15 days ago

        The problem is that anything even remotely related to AI is just being called “AI,” whether it’s by the average person or marketing people.

        So when you go to a company’s website and you see “powered by AI,” they could be talking about LLMs, or an ML model to detect cancer, and the average person won’t know the difference between the technologies.

        So if someone universally rejects anything that says it “uses AI” just because what’s usually called “AI” is just badly implemented LLMs that make the experience worse, they’re going to inevitably catch nearly every ML model in the crossfire too, since most companies are calling their ML use cases “AI powered,” and that means rejecting companies that develop models like those that detect tumors, predict protein folding patterns, identify anomalies in other health characteristics, optimize traffic routes in cities, etc, even if those use cases aren’t even related to LLMs and all the flaws they often bring.

  • @Zacryon@feddit.org
    link
    fedilink
    44 days ago

    LLMs != AI
    LLMs strict subset of AI

    Pls be a bit more specific about what you hate about the wide field of AI. Otherwise it’s almost like saying you hate computers, because they can run applications that you don’t like.

  • @GoodOleAmerika@lemmy.world
    link
    fedilink
    34 days ago

    I use AI as a tool. AI should be a tool to help with job, not to take jobs. Same as calculator. Yep people will be able to code faster with AIs help, so that might mean less demand, at least for IT. But u still gotta know what the exact prompt u need to ask

    • @ameancow@lemmy.world
      link
      fedilink
      English
      4
      edit-2
      5 days ago

      I’ve been using a cartoonish profile picture for my work emails, teams portrait and other communications for many years. There is almost no way to tell that kind of icon apart from AI generated icons at that size anyway.

      And even if it was, that’s not the point of the conversation. Fixating on that is such bad faith it betrays a defensiveness about AI generated content, so it’s particularly important that someone like you get this message, let me reiterate clearly:

      I have a role of responsibility, I hire people and use company budget to make decisions on other companies and products we’ll be paying for. When making these decisions I don’t look at the email signatures of people or the icons they use. I look at their presentation materials and if that shit is AI generated I know immediately it’s just a couple people pretending to be an agency or company, or some company that doesn’t quality-control their slides and presentation decks. It shows laziness. I would rather go with a company that has data and specs rather than lean on graphics anyway. So if those graphics are also lazy AF that’s a hard pass. Not my first rodeo, I’ve learned to listen to experience.

      • @Xatolos@reddthat.com
        link
        fedilink
        15 days ago

        I was pointing out the irony, nothing more. Not every comment needs to be read too deeply into. Even my avatar picture is AI generated, it was from those blockchain AI generated reddit ones years ago.

        • @ameancow@lemmy.world
          link
          fedilink
          English
          55 days ago

          I don’t think anyone thought it was that funny or interesting of a comment, it reads the same kind of petty AI-bro comment that people make who are absolutely hooked on blindly defending AI in all capacities.

          • @Xatolos@reddthat.com
            link
            fedilink
            1
            edit-2
            4 days ago

            Sure it did… and judging from your previous rant, I’m sure it has nothing to do with you maybe reading much deeper into things that aren’t really there.

  • @zerofk@lemm.ee
    link
    fedilink
    35 days ago

    Ironically, an LLM could’ve made his post grammatically correct and understandable.

  • @felykiosa@sh.itjust.works
    link
    fedilink
    35 days ago

    I have to disagree with that one but not completely. It really depends on what type of company I interact with . is that an independent small company or a big corp . also what type of AI (generate picture or generate summary etc…) And is the application fullfit or not . ex if you generate a logo or a picture in a small business is the style of the picture correct or is it the same as everyone , also did you check if the image was correct etc… But for big corps yeah they can go fuck themselves, they have the budget to pay artist