view the rest of the comments
AMUSING, INTERESTING, OUTRAGEOUS, or PROFOUND
This is a page for anything that's amusing, interesting, outrageous, or profound.
♦ ♦ ♦
RULES
❶ Each player gets six cards, except the player on the dealer's right, who gets seven.
❷ Posts, comments, and participants must be amusing, interesting, outrageous, or profound.
❸ This page uses Reverse Lemmy-Points™, or 'bad karma'. Please downvote all posts and comments.
❹ Posts, comments, and participants that are not amusing, interesting, outrageous, or profound will be removed.
❺ This is a non-smoking page. If you must smoke, please click away and come back later.
❻ Don't be a dick.
Please also abide by the instance rules.
♦ ♦ ♦
Can't get enough? Visit my blog.
♦ ♦ ♦
Please consider donating to Lemmy and Lemmy.World.
$5 a month is all they ask — an absurdly low price for a Lemmyverse of news, education, entertainment, and silly memes.
Since I already started typing my hot take reaction I'll just post it as is, but seeing as Ed Zitron has addressed this I'd definitely read his take as it's surely better sourced and professionally written :)
I made it to about here and then decided I couldn't suspend disbelief enough to continue. The way this author describes it we're now at the point where you can just prompt, "I want a native Linux version of Adobe Photoshop/Solidworks/whatever-the-fuck-MS-Office-is-now-called," and it should be able to make that happen.
Now that I think more about it, why should anybody even care if AI can write code? You wouldn't tell it to make accounting software for your business, you'd tell it to do your business accounting and not care how it was doing so.
Prior to 2022 our calculators could reliably do arithmetic, but would also display an error when asked to do nonsensical operations. Are the amazing models the author describes now consistently able to resist efforts to force a response where a correct one is impossible? I suspect with the right prodding it'll still declare a winner in unstoppable vs. immovable etc.
So am I not supposed to assume that a model was trained on the bar exam and then asked to complete a test using the same material? I wouldn't be impressed if you told me a child passed the bar exam by using the answer key either.
I read this as "working software" = capable of "Hello World!", and "explain graduate-level science" = can quote relevant blocks of text sourced from wikis and scholarly docs. Could it attempt to plausibly explain anything novel that is science related, or only repeat already published and understood things?
If the best engineers in the world already handed over most of their coding work then why are any engineers still employed anywhere? How is just replacing customer service reps with AI agents working out so far?
Also, let me point out they didn't properly grade the bar exam: https://www.livescience.com/technology/artificial-intelligence/gpt-4-didnt-ace-the-bar-exam-after-all-mit-research-suggests-it-barely-passed
It did excellent on the multiple choice section, but so would literally any law student using Google.
And that's not the only lie. It can't even repeat stuff we already know. I occasionally give a model one of my own, by new decades old, papers without the abstract and conclusions and asked what it could conclude. It got it completely wrong. Like not-even-funny wrong, wrong conclusions, wrong theory, wrong methodology.
It's pretty fun to see AI boosters get upset at that and blame my paper for the LLM saying literally the opposite of what it says.
Excellent & appreciated, thanks.