220
"Sam Altman is one of the dullest, most incurious and least creative people to walk this earth."
(deadsimpletech.com)
Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.
This is not debate club. Unless it’s amusing debate.
For actually-good tech, you want our NotAwfulTech community
In this and other use cases I call it a pretty effective search engine, instead of scrolling through stackexchange after clicking between google ads, you get the cleaned up example code you needed. Not a Chat with any intelligence though.
That ChatGPT can be more useful than a web search is really more indicative of how bad the web has got, and can only get worse as fake text invades it. It's not actually better than a functional search engine and a functional web, but the companies making these things have no interest in the web being usable. Pretty depressing.
Remember when you could read through all the search results on Google rather than being limited to the first hundred or so results like today? And boolean search operators actually worked and weren't hidden away behind a "beware of leopard" sign? Pepperidge Farm remembers.
"despite the many people who have shown time and time and time again that it definitely does not do fine detail well and will often present shit that just 10000% was not in the source material, I still believe that it is right all the time and gives me perfectly clean code. it is them, not I, that are the rubes"
The problem with stuff like this is not knowing when you dont know. People who had not read the books SSC Scott was reviewing didnt know he had missed the points (or not read the book at all) till people pointed it out in the comments. But the reviews stay up.
Anyway this stuff always feels like a huge motte bailey, where we go from 'it has some uses' to 'it has some uses if you are a domain expert who checks the output diligently' back to 'some general use'.
A lot of the "I'm a senior engineer and it's useful" people seem to just assume that they're just so fucking good that they'll obviously know when the machine lies to them so it's fine. Which is one, hubris, two, why the fuck are you even using it then if you already have to be omniscient to verify the output??
"If you don't know the subject, you can't tell if the summary is good" is a basic lesson that so many people refuse to learn.
Ahah I'm totally with you, I just personally know people that love it because they have never learned how to use a search engine. And these generalist generative AIs are trained on gobbled up internet basically, while also generating so many dangerous mistakes, I've read enough horror stories.
I'm in science and I'm not interested in ChatGPT, wouldn't trust it with a pancake recipe. Even if it was useful to me I wouldn't trust the vendor lock-in or enshittification that's gonna come after I get dependent on aa tool in the cloud.
A local LLM on cheap or widely available hardware with reproducible input / output? Then I'm interested.