this post was submitted on 18 Sep 2023
492 points (96.9% liked)

Programmer Humor

32479 readers
893 users here now

Post funny things about programming here! (Or just rant about your favourite programming language.)

Rules:

founded 5 years ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 26 points 1 year ago (3 children)

Any task that can be expressed as mostly translation is a good task to try with an LLM.

And you know what? Stakeholders tend to love LLMs, so have fun with your complicated problems while I build them by using the ancient technique of slapping some boilerplate together and combining it with the new ways of pasting error messages into chatgippity.

[–] [email protected] 40 points 1 year ago* (last edited 1 year ago) (2 children)

I don't think that was the point. The thing is, people replace calculators with that...

  • User: Assistant?
  • Assistant: * BEEP *
  • User: What is 21 divided by three?
  • Assistant: 52, my master.

Thing is, they only get some results right and hallucinate others. And you're doing billions of matrix multiplications just to calculate 2+1.

Sure. You can go to a construction site with only your one favorite tool. And use it for everything. And it's impressive to open a glass bottle of beer with a hammer and such. But I can guarantee you, you'll be slower digging that hole than the guys using a proper tool like an excavator.

[–] [email protected] 12 points 1 year ago

And: you don't solve any fundamental problems if you don't have the data for it. If the information isn't in your data, the network will start guessing and it will be horrible.

[–] [email protected] 3 points 1 year ago (1 children)

That's not a translation problem, so LLMs are terrible for it.

Always use the right tool for the job. If there are a lot of nails to be hammered, you need a guy with a hammer.

[–] [email protected] 6 points 1 year ago

Yeah, that was kind of my point. I think the meme picture means people throw it at everything. No matter what. And the next logical thing would be to strip the computer scientist out of the picture. We have Github Copilot now ;) Let AI decide if AI is the proper tool.

[–] [email protected] 9 points 1 year ago

Translation, but not categorization. Trying to get reliable, and more importantly, predictibly accurate, metadata from an LLM without serious training is a pain. ML algorithms are far better for this but certainly take more brainpower (in my experience so far).

[–] [email protected] 3 points 1 year ago

There are a disproportionately large number of people who get one pretty demo and think LLMs are the solution to everything. Even for translations, I'd be interested to see how accurate the major models are in real world scenarios. We've been struggling hard to find any practical usage of LLMs that doesn't require the user to be able to verify the output themselves.