13
you are viewing a single comment's thread
view the rest of the comments
[-] AbouBenAdhem@lemmy.world -2 points 1 day ago

amplifying H-Neurons’ activations systematically increases a spectrum of over-compliance behaviors – ranging from overcommitment to incorrect premises and heightened susceptibility to misleading contexts, to increased adherence to harmful instructions and stronger sycophantic tendencies. These findings suggest that H-Neurons do not simply encode factual errors, but rather represent a general tendency to prioritize conversational compliance over factual integrity.

I wonder if the same tendencies are associated in humans—and if so, is it something LLMs learned from humans, or is it a consequence of the general structure of neural networks?

[-] snooggums@piefed.world 4 points 1 day ago

Prioritizing conversational compliance over factual integrity when the output is promoted as being factual is a design flaw.

Saying double check the output does not excuse that flaw when LLM CEOS say their models are like someone with a PhD or that it can automate every white collar job within a year.

[-] ageedizzle@piefed.ca 2 points 1 day ago* (last edited 1 day ago)

Is it a design flaw? Or is it just false advertising? If I sell you a vacuum by telling you it can mop your floor, is the problem with the vacuum or the way I’m selling the product?

[-] XLE@piefed.social 4 points 1 day ago

For this particular paper, it seems like a design flaw got uncovered. And it may very well be part of the architecture of how LLMs are even readable to begin with, given how deep and universal the "bad" nodes are.

I can't prove any AI company was aware of this, but they would have been in a much better position to realize it than researchers who have to do a postmortem on the models being crappy. And if they weren't aware of it, they're probably not very good at their jobs...

[-] snooggums@piefed.world 2 points 1 day ago

Since shop vacs which vacuum and suck up water exist, it could be both.

this post was submitted on 25 Feb 2026
13 points (60.3% liked)

Technology

81869 readers
4364 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS