this post was submitted on 23 Sep 2024
175 points (94.9% liked)

Technology

60047 readers
2811 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 11 points 3 months ago (2 children)

I don't understand why it's so hard to sandbox an LLM's configuration data from it's training data.

[–] MoondropLight 10 points 3 months ago

Because its all one thing. The promise of AI is that you can basically throw anything at it, and you don't need to understand exactly how/why it makes the connections it does; you just adjust the weights until it kinda looks alright.

There are many structural hacks used to give it better results (and in this case some form of reasoning) but ultimately they're mostly relying on connecting multiple nets together and retrying queries and such. There's no human understandable settings. Neural networks are basically one input and one output (unless you're training it).

[–] [email protected] 1 points 3 months ago (1 children)

What do you mean by "configuration data?"

[–] [email protected] 2 points 3 months ago (1 children)

The data used to configure it.

[–] [email protected] 1 points 3 months ago

Do you mean finetune data?

A model's configuration data is training data.