this post was submitted on 10 Feb 2025
522 points (95.3% liked)

Technology

62073 readers
5004 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 30 points 1 day ago (1 children)

I hate to disagree but IIRC deepseek is not a open-source model but open-weight?

[–] [email protected] 32 points 1 day ago* (last edited 1 day ago) (2 children)

It's tricky. There is code involved, and the code is open source. There is a neural net involved, and it is released as open weights. The part that is not available is the "input" that went into the training. This seems to be a common way in which models are released as both "open source" and "open weights", but you wouldn't necessarily be able to replicate the outcome with $5M or whatever it takes to train the foundation model, since you'd have to guess about what they used as their input training corpus.

[–] [email protected] 2 points 15 hours ago* (last edited 15 hours ago)

Definitions are tricky, and especially for terms that are broadly considered virtuous/positive by the general public (cf. "organic") but I tend to deny something is open source unless you can recreate any binaries/output AND it is presented in the "preferred form for modification" (i.e. the way the GPLv3 defines the "source form").

A disassembled/decompiled binary might nominally be in some programming language--suitable input to a compiler for that langauge--but that doesn't actually make it the source code for that binary because it is not in the form the entity most enabled to make a modified form of the binary (normally the original author) would prefer to make modifications.

[–] [email protected] 8 points 23 hours ago (1 children)

I view it as the source code of the model is the training data. The code supplied is a bespoke compiler for it, which emits a binary blob (the weights). A compiler is written in code too, just like any other program. So what they released is the equivalent of the compiler's source code, and the binary blob that it output when fed the training data (source code) which they did NOT release.

[–] [email protected] 1 points 8 hours ago

This is probably the best explanation I've seen so far and really helped me actually understand what it means when we talk about "weights" for LLMs.