this post was submitted on 17 Apr 2024
109 points (91.6% liked)

Technology

59366 readers
5811 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 5 points 7 months ago (1 children)

This is already worked in through mathematics, it is its own mathematical field. We can optimize packaging through formulas that are very fast and accurate. No need to train a AI for that. Especially not for space flight, AI are prone to hallucinations that is not something you want anywhere near any space mission that requires precision and predictability. I believe Johannes Kepler started this field in the 1600s, it is not something new. It is definitely a complex problem, but not new and not unheard of. Amazon is not exactly inventing something new and amazing here..

[–] [email protected] 5 points 7 months ago (3 children)

AI is not prone to hallucinations, LLMs are. I doubt Amazon is building a chatbot to optimise packaging.

[–] [email protected] 3 points 7 months ago (1 children)

I mean, AI is used in fraud detection pretty often; when it hits a false positive (which happens frequently on a population-level basis), is that not a hallucination of some sort? Obviously LLMs can go off the rails much further because it's readable text, but any machine learning model will occasionally spit out really bad guesses almost any person could have done better with. (To be fair, humans are highly capable of really bad guesses too).

[–] [email protected] 6 points 7 months ago (1 children)

No, false positives and false negatives are not hallucinations. Otherwise things like a blood test not involving any ml would also be halucinating which removes all meaning from the term.

[–] [email protected] 2 points 7 months ago

That's fair. I think fundamentally a false positive/negative isn't that much different. Pretty much all tests—especially those dealing with real world conditions—are heuristic, as are all LLMs by necessity of the design. Hallucination is a pretty specific term given to AI as an attempt to assign agency to a system that doesn't actually have any (by implying it's crazy and making stuff up instead of a black box with deterministic inputs and outputs spitting out something factually wrong but with a similar format to what is trained on). I feel like the nature of any tool where "you can't trust this to be entirely accurate" should have an umbrella term that encompasses both types of providing inaccurate info under certain conditions.

I suppose the difference is that AI is a lot more likely to randomly go off, whereas a blood test is likelier to provide repeated false positives for the same person with their unique biology? There's also the fact that most medical tests represent a true/false dichotomy or lookup table, whereas an LLM is given the entire bounds of language.

Would an AI clustering algorithm (say, K-means for instance) giving an inaccurate diagnosis be a false positive/negative or a hallucination? These models can be programmed on a sliding scale and I feel like there's definitely an area where the line could get pretty blurry.

[–] [email protected] 0 points 7 months ago (1 children)

What do you consider to be an AI?
And do you consider any of the existing systems to be the one?

[–] [email protected] 1 points 7 months ago (1 children)

When I use "AI" I'm using computer science terminology. Artificial intelligence is a subfield of CS, in that sense, any model that comes of that field is, by definition, AI.

[–] [email protected] 0 points 7 months ago* (last edited 7 months ago) (2 children)

Then it's strange that you are separating AI and LLM, because in CS LLM is a type of artificial intelligence.

[–] [email protected] 1 points 7 months ago (1 children)

Some AI, namely, LLMs, can hallucinate, but not AI in general. I just had a bit of fun in how I worded it, I guess I should've expected someone to become annoyingly nitpicky about it.

[–] [email protected] 0 points 7 months ago* (last edited 7 months ago) (1 children)

Technicalities matter in technological matters.

[–] [email protected] 1 points 7 months ago

I don't think I was being wrong, technically, I do think you can write that way if you want to be a bit facetious, but I'm not a native speaker so, maybe not.

[–] [email protected] 1 points 7 months ago

AI as a whole is not subject to the flaws of LLMs

[–] [email protected] -2 points 7 months ago (1 children)

AI in general is definitely prone to hallucinations. It is more commonly seen in LLMs because it is more widely used by the public. It is definitely a problem with all AI

[–] [email protected] 2 points 7 months ago* (last edited 7 months ago) (1 children)

Besides generative AI, which models can hallucinate?

[–] [email protected] 1 points 7 months ago (1 children)

Text to video, automated driving, object detection, language translations. I might be misusing the term, you could argue that the word is describing what LLMs commonly does and that is where the term is derived from. You can also argue that AI is sometimes correct and the human have issues identifying the correct answer. But In my mind it is much the same just different applications. A car completely missing a firetruck approaching or a LLM just spewing out wrong statements is the same to me.

[–] [email protected] 1 points 7 months ago (1 children)

Yeah, well it's not the same. Models are wrong all the time, why use a different term at all when it's just "being wrong"?

[–] [email protected] 1 points 7 months ago (1 children)

The model makes decisions thinking it is right, but for whatever reason can't see a firetruck or stopsign or misidentifies the object.. you know almost like how a human hallucinating would perceive something from external sensory that is not there.

I don't mind giving it another term, but "being wrong" is misleading. But you are correct in the sense that it depends on every given case..

[–] [email protected] 1 points 7 months ago (1 children)

No, the model isn't "thinking", no model in use today has anything resembling an internal cognitive process. It is making a prediction. A covid test is predicting whether you have the Covid-19 virus inside you or not. If its prediction contradicts your biological state, it is wrong. If an object recognition algorithm does not predict there being a firetruck, how is that not being wrong in the same way?

[–] [email protected] 1 points 7 months ago

Predicting? Ok, if you say so.