this post was submitted on 17 Aug 2023
486 points (96.0% liked)

Technology

59581 readers
2822 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

cross-posted from: https://nom.mom/post/121481

OpenAI could be fined up to $150,000 for each piece of infringing content.https://arstechnica.com/tech-policy/2023/08/report-potential-nyt-lawsuit-could-force-openai-to-wipe-chatgpt-and-start-over/#comments

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] -1 points 1 year ago (2 children)

This is a misunderstanding on your part. While some neurons are trained this way, word2vec and doc2vec are not these mechanisms. The llms are extensions of these models and while there are certainly some aspects of what you are describing, there is a transcription into vector formats.

This is the power of vectorization of language (among other things). The one to one mapping between vectors and words / sentences to documents and so forth allows models to describe the distance between words or phrases using euclidian geometry.

[–] [email protected] 2 points 1 year ago* (last edited 1 year ago) (1 children)

I was trying to make it as simple as possible. The format is irrelevant. The model is still storing nothing but weights at the end of the day. Storing the relationships between words and sentences is not the same thing as storing works in a different format which is what your original comment implied.

[–] [email protected] -1 points 1 year ago

I'm sorry you failed to grasp how it works in this context.

[–] [email protected] 0 points 1 year ago (1 children)

You made me really interested in this concept so I asked GPT-4 what the furthest word away from the word “vectorization” would be.

Interesting game! If we're aiming for a word that's conceptually, contextually, and semantically distant from "vectorization," I'd pick "marshmallow." While "vectorization" pertains to complex computational processes and mathematics, "marshmallow" is a soft, sweet confectionery. They're quite far apart in terms of their typical contexts and meanings.

It honestly never ceases to surprise me. I’m gonna play around with some more. I do really like the idea that it’s essentially a word calculator.

[–] [email protected] 4 points 1 year ago

Try asking it how the vectorization of king and queen are related.