16
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
this post was submitted on 19 Jan 2026
16 points (94.4% liked)
TechTakes
2412 readers
127 users here now
Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.
This is not debate club. Unless it’s amusing debate.
For actually-good tech, you want our NotAwfulTech community
founded 2 years ago
MODERATORS
TracingWoodgrains's hit piece on David Gerard (the 2024 one, not the more recent enemies list one, where David Gerard got rated above the Zizians as lesswrong's enemy) is in the top 15 for lesswrong articles from 2024, currently rated at #5! https://www.lesswrong.com/posts/PsQJxHDjHKFcFrPLD/deeper-reviews-for-the-top-15-of-the-2024-review
It's nice to see that with all the lesswrong content about AI safety and alignment and saving the world and human rationality and fanfiction, an article explaining about how terrible David Gerard is (for... checks notes, demanding proper valid sources about lesswrong and adjacent topics on wikipedia) won out to be voted above them! Let's keep up our support for dgerard!
The #5 article of the year was a crock of a few kinds of shit, and I have already spent too much time thinking about why
Picking a few that I haven't read but where I've researched the foundations, let's have a party platter of sneers:
To add to your sneers... lots of lesswrong content fits you description of #9, with someone trying to invent something that probably exists in philosophy, from (rationalist, i.e. the sequences) first principles and doing a bad job at it.
I actually don't mind content like #25 where someone writes an explainer topic? If lesswrong was less pretentious about it and more trustworthy (i.e. cited sources in a verifiable way and called each other out for making stuff up) and didn't include all the other junk and just had stuff like that it would be better at its stated goal of promoting rationality. Of course, even if they tried this, they would probably end up more like #47 where they rediscover basic concepts because they don't know how to search existing literature/research and cite it effectively.
45 is funny. Rationalists and rationalist adjacent people started OpenAI, ultimately ignored "AI safety". Rationalist spun off anthropic, which also abandoned the safety focus pretty much after it had gotten all the funding it could with that line. Do they really think a third company would be any better?
Wonder if that was because it basically broke containment (still was not widely spread, but I have seen it at a few places, more than normal lw stuff) and went after one of their enemies (And people swallowed it uncritically, wonder how many of those people now worry about NRx/Yarvin and don't make the connection).