TinyTimmyTokyo

joined 1 year ago
[–] [email protected] 1 points 1 year ago

I see him more as a dupe than a Cassandra. I heard him on a podcast a couple months ago talking about how he's been having conversations with Bay Area AI researchers who are "really scared" about what they're creating. He also spent quite a bit of time talking up Geoffrey Hinton's AI doomer tour. So while I don't think Ezra's one of the Yuddite rationalists, he's clearly been influenced by them. Given his historical ties to effective altruism, this isn't surprising to me.

[–] [email protected] 1 points 1 year ago* (last edited 1 year ago) (1 children)

I mean, of course he loves unfettered technology and capitalism. He's a fucking billionaire. He hit the demographic lottery.

EDIT: I just noticed his list of "techno-optimist" patrons. On the list? John Galt. LMAO. The whole list is pretty much an orgy of libertarians.

[–] [email protected] 3 points 1 year ago* (last edited 1 year ago) (1 children)

Roko's authoritative-toned "aktshually..." response to Annie's claims have me fuming. I don't know why. I mean I've known for years that this guy is a total boil on the ass of humanity. And yet he still manages to shock with the worst possible take on a topic -- even when the topic is sexual abuse of a child. If, like Roko, I were to play armchair psychiatrist, I'd diagnose him as a sociopath with psychopathic tendencies. But I'm not. So I won't.

[–] [email protected] 1 points 1 year ago (1 children)

My attention span is not what it used to be, and I couldn't force myself to get to the end of this. A summary or TLDR (on the part of the original author) would have been helpful.

What is it with rationalists and their inability to write with concision? Is there a gene for bloviation that also predisposes them to the cult? Or are they all just mimicking Yud's irritating style?

[–] [email protected] 1 points 1 year ago

Is it wrong to hope they manage to realize one of these libertarian paradise fantasies? I'd really love to see how quickly it devolves into a Mad Max Thunderdome situation.

[–] [email protected] 0 points 1 year ago (1 children)

What's it like to be so good at PR?

[–] [email protected] 1 points 1 year ago (1 children)

Stephen Jay Gould's The Mismeasure of Man is always a good place to start.

[–] [email protected] 1 points 1 year ago* (last edited 1 year ago) (4 children)

This is good:

Take the sequence {1,2,3,4,x}. What should x be? Only someone who is clueless about induction would answer 5 as if it were the only answer (see Goodman’s problem in a philosophy textbook or ask your closest Fat Tony) [Note: We can also apply here Wittgenstein’s rule-following problem, which states that any of an infinite number of functions is compatible with any finite sequence. Source: Paul Bogossian]. Not only clueless, but obedient enough to want to think in a certain way.

Also this:

If, as psychologists show, MDs and academics tend to have a higher “IQ” that is slightly informative (higher, but on a noisy average), it is largely because to get into schools you need to score on a test similar to “IQ”. The mere presence of such a filter increases the visible mean and lower the visible variance. Probability and statistics confuse fools.

And:

If someone came up w/a numerical“Well Being Quotient” WBQ or “Sleep Quotient”, SQ, trying to mimic temperature or a physical quantity, you’d find it absurd. But put enough academics w/physics envy and race hatred on it and it will become an official measure.

[–] [email protected] 1 points 1 year ago* (last edited 1 year ago)

Now that his alter ego has been exposed, Hanania is falling back on the "stupid things I said my youth" chestnut. Here's a good response to that.

[–] [email protected] 1 points 1 year ago* (last edited 1 year ago) (1 children)

In theory, a prediction market can work. The idea is that even though there are a lot of uninformed people making bets, their bad predictions tend to cancel each other out, while the subgroup of experts within that crowd will converge on a good prediction. The problem is that prediction markets only work when they're ideal. As soon as the bettor pool becomes skewed by a biased subpopulation, they stop working. And that's exactly what happens with the rationalist crowd. The main benefit rationalists obtain from prediction markets and wagers is an unfounded confidence that their ideaas have merit. Prediction markets also have a long history in libertarian circles, which probably also helps explain why rationalists are so keen on them.

[–] [email protected] 1 points 1 year ago

"TempleOS on the blockchain"

Ok that's some quality sneer. A bit obscure and esoteric, but otherwise perfect for those who know anything about Temple OS.

[–] [email protected] 1 points 1 year ago

Yeah, Behe's one of the leading lights (dimmest bulbs?) of the so-called "Intelligent Design" movement: a molecular biologist who knows just enough molecular biology to construct strawmen arguments about evolution. Siskind being impressed by him tells me everything I need to know about Siskind's susceptibility to truly stupid ideas.

view more: ‹ prev next ›