1564
Everything's fine. Nothing to see here.
(thelemmy.club)
"We did it, Patrick! We made a technological breakthrough!"
A place for all those who loathe AI to discuss things, post articles, and ridicule the AI hype. Proud supporter of working people. And proud booer of SXSW 2024.
AI, in this case, refers to LLMs, GPT technology, and anything listed as "AI" meant to increase market valuations.
I'm a data analyst and primary authority on the data model of a particular source system. Most questions for figures from that system that can't be answered directly and easily in the frontend end up with me.
I had a manager show me how some new LLM they were developing (which I had contributed some information about the model to) could quickly answer some questions that usually I have to answer manually, as part of a pitch to make me switch to his department so I can apply my expertise for improving this fancy AI instead of answering questions manually.
He entered a prompt, got a figure that I knew wasn't correct and I queried my data model for the same info, with a significantly different answer. Given how much said manager leaned on my expertise in the first place, he couldn't very well challenge my results and got all sheepish about how the AI still in development and all.
I don't know how that model arrived at that figure. I don't know if it generated and ran a query against the data I'd provided. I don't know if it just invented the number. I don't know how the devs would figure out the error and how to fix it. But I do know how to explain my own queries, how to investigate errors and (usually) how to find a solution.
Anyone who relies on a random text generator - no matter how complex that generation method to make it sound human - to generate facts is dangerously inept.
This is like the biggest factor that people don't get when thinking of these models in the context of software. "Oh it got it wrong, but the developers will fix it in an update". Nope, they can fix traditional software mistakes, LLM output and machine learning things... They can throw more training data at it (which sometimes just changes what it gets wrong) and hope for the best, they can do better job at curating the context window to give the model the best shot at outputting the right stuff (e.g. the guy who got Opus to generate a slow crappy buggy compiler had to traditionally write a filter to find and show only the 'relevent' compiler output back to the models), they can try to generate code to do what you want and have you review the code and correct issues. But debugging and fixing the model itself... that's just not a thing at all.
I was in a meeting where a sales executive was bragging about the 'AI sales agent' they were working, but admitting frustration with the developres and a bit confused why the software developers weren't making progress when those same developers always made decent progress before, and they should be able to do this even faster because they have AI tools to help them... It eternally seemed in a state that almost worked but not quite no matter what model or iteration they went to, no matter how much budget they allocated, when it came down to the specific facts and figures it would always screw up.
I cannot understand how long these executives wade in the LLM pool and still believes in capabilities beyond what anyone has experienced.
They leave the actual work to the boots on the ground so they don't see how shitty the output is. They listen to marketing about how great it is and mandate everyone use it and then any feedback is filtered through all the brownnosers that report to them.
This is probably the biggest misunderstanding since "Project Managers think three developers can produce a baby in three months": Just throw more time and money at AI model "development" for better results. It supposes predictable, deterministic behaviour that can be corrected, but LLMs aren't deterministic ny design, since that wouldn't sound human anymore.
Sure, when you're a developer dedicated to advancing the underlying technology, you may actually produce better results in time, but if you're just the consumer, you may get a quick turnaround for an alright result (and for some purposes, "alright" may be enough) but eventually you'll plateau at the limitations of the model.
Of course, executives universally seem to struggle with the concept of upper limits, such as sustainable growth or productivity.