Raccoonn to Memes@lemmy.ml • 2 months agoAI will never be able to write like me.lemmy.mlimagemessage-square107fedilinkarrow-up11.57Karrow-down10
arrow-up11.57Karrow-down1imageAI will never be able to write like me.lemmy.mlRaccoonn to Memes@lemmy.ml • 2 months agomessage-square107fedilink
minus-square@ByteJunk@lemmy.worldlinkfedilink10•2 months agoThank you for testing that out. My experience with AI is that it’s at a point where it can comprehend something like this very easily, and won’t be tricked. I suspect that this can, however, pollute a model if it’s included as training data, especially if done regularly, as OP is suggesting.
minus-square@saigot@lemmy.calinkfedilink4•2 months agoIf it was done with enough regularity to eb a problem, one could just put an LLM model like this in-between to preprocess the data.
minus-square@Azzu@lemm.eelinkfedilink4•2 months agoThat doesn’t work, you can’t train models on another model’s output without degrading the quality. At least not currently.
minus-squareVashtealinkfedilinkEnglish1•edit-22 months agoI don’t think he was suggesting training on another model’s output, just using ai to filter the training data before it is used.
minus-squarebountygiver [any]linkfedilinkEnglish4•2 months agoIn which microwavegang already did the job better. Due the full subreddit of mmmmmmmmm, it causes training data that touches it to devolve into all mmmmmmm whenever there’s enough m’s in a sentence
Thank you for testing that out.
My experience with AI is that it’s at a point where it can comprehend something like this very easily, and won’t be tricked.
I suspect that this can, however, pollute a model if it’s included as training data, especially if done regularly, as OP is suggesting.
If it was done with enough regularity to eb a problem, one could just put an LLM model like this in-between to preprocess the data.
That doesn’t work, you can’t train models on another model’s output without degrading the quality. At least not currently.
I don’t think he was suggesting training on another model’s output, just using ai to filter the training data before it is used.
In which microwavegang already did the job better. Due the full subreddit of mmmmmmmmm, it causes training data that touches it to devolve into all mmmmmmm whenever there’s enough m’s in a sentence