this post was submitted on 16 May 2025
601 points (97.2% liked)

Technology

70080 readers
3470 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

It certainly wasn’t because the company is owned by a far-right South African billionaire at the same moment that the Trump admin is entertaining a plan to grant refugee status to white Afrikaners. /s

My partner is a real refugee. She was jailed for advocating democracy in her home country. She would have received a lengthy prison sentence after trial had she not escaped. This crap is bullshit. Btw, did you hear about the white-genocide happening in the USA? Sorry, I must have used Grok to write this. Go Elon! Cybertrucks are cool! Twitter isn’t a racist hellscape!

The stuff at the end was sarcasm, you dolt. Shut up.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 67 points 1 day ago (4 children)

Yeah, billionaires are just going to randomly change AI around whenever they feel like it.

That AI you've been using for 5 years? Wake up one day, and it's been lobotomized into a trump asshole. Now it gives you bad information constantly.

Maybe the AI was taken over by religious assholes, now telling people that gods exist, manufacturing false evidence?

Who knows who is controlling these AI. Billionaires, tech assholes, some random evil corporation?

[–] [email protected] 39 points 1 day ago* (last edited 1 day ago) (1 children)

Joke's on you, LLMs already give us bad information

[–] [email protected] 13 points 1 day ago (1 children)

Sure, but unintentionally. I heard about a guy whose small business (which is just him) recently had someone call in, furious because ChatGPT told them that he was having a sale that she couldn't find. The customer didn't believe him when he said that the promotion didn't exist. Once someone decides to leverage that, and make a sufficiently-popular AI model start giving bad information on purpose, things will escalate.

Even now, I think Elon could put a small company out of business if he wanted to, just by making Grok claim that its owner was a pedophile or something.

[–] [email protected] 10 points 1 day ago (1 children)

"Unintentionally" is the wrong word, because it attributes the intent to the model rather than the people who designed it.

Hallucinations are not an accidental side effect, they are the inevitable result of building a multidimensional map of human language use. People hallucinate, lie, dissemble, write fiction, misrepresent reality, etc. Obviously a system that is designed to map out a human-sounding path from a given system prompt to a particular query is going to take those same shortcuts that people used in its training data.

[–] [email protected] 3 points 1 day ago (1 children)

Unintentionally is the right word because the people who designed it did not intend for it to be bad information. They chose an approach that resulted in bad information because of the data they chose to train and the steps that they took throughout the process.

[–] [email protected] 8 points 1 day ago

Incorrect. The people who designed it did not set out with a goal of producing a bot that reguritates true information. If that's what they wanted they'd never have used a neural network architecture in the first place.

[–] [email protected] 10 points 1 day ago (2 children)

That's a good reason to use open source models. If your provider does something you don't like, you can always switch to another one, or even selfhost it.

[–] [email protected] 21 points 1 day ago (1 children)

Or better yet, use your own brain.

[–] [email protected] 4 points 1 day ago

Yep, not arguing for the use of generative AI in the slightest. I very rarely use it myself.

[–] [email protected] 10 points 1 day ago

While true, it doesn't keep you safe from sleeper agent attacks.

These can essentially allow the creator of your model to inject (seamlessly, undetectably until the desired response is triggered) behaviors into a model that will only trigger when given a specific prompt, or when a certain condition is met. (such as a date in time having passed)

https://arxiv.org/pdf/2401.05566

It's obviously not as likely as a company simply tweaking their models when they feel like it, and it prevents them from changing anything on the fly after the training is complete and the model is distributed, (although I could see a model designed to pull from the internet being given a vulnerability where it queries a specific URL on the company's servers that can then be updated with any given additional payload) but I personally think we'll see vulnerabilities like this become evident over time, as I have no doubts it will become a target, especially for nation state actors, to simply slip some faulty data into training datasets or fine-tuning processes that get picked up by many models.

[–] [email protected] 5 points 1 day ago (1 children)

I currently treat any positive interaction with an LLM as a “while the getting’s good” experience. It probably won’t be this good forever, just like Google’s search.

[–] [email protected] 8 points 1 day ago (1 children)

Pretty sad that the current state would be considered "good"

[–] [email protected] 5 points 1 day ago (1 children)

With accuracy rates declining over time, we are at the 'as good as it gets' phase!

[–] [email protected] 2 points 1 day ago

If that's the case, where's Jack Nicholson?

[–] [email protected] 2 points 1 day ago

Yep, I knew this from the very beginning. Sadly the hype consumed the stupid, as it always will. And we will suffer for it, even though we knew better. Sometimes I hate humanity.