this post was submitted on 17 Mar 2025
174 points (96.3% liked)

Technology

66783 readers
4641 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 8 points 1 day ago (2 children)

Isn't DeepSeek open source? Is there a need to reverse engineer it?

[–] [email protected] 24 points 1 day ago (2 children)

"Open source" in ML is a really bad description for what it is. "Free binary with a bit of metadata" would be more accurate. The code used to create deepseek is not open source, nor is the training datasets. 99% of "open source" models are this way. The only interesting part of the open sourcing is the architecture used to run the models, as it lends a lot of insight into the training process, and allows for derivatives via post-training

[–] [email protected] 8 points 14 hours ago

Deepseek actually released a bunch of their infrastructure code, including the infamous tricks for making training and interference more efficient, a couple of weeks ago.

[–] [email protected] 9 points 15 hours ago

It certainly is a lot more open source than OpenAI, that's for sure.

[–] [email protected] 12 points 1 day ago* (last edited 23 hours ago) (1 children)

Yes, and no. Yes in that they've released the research papers, pretrained parameters and weights of the model itself. Which is more than I can say for "OpenAI." But no in that it doesn't include training data or other critical components. Luckily, they've shown how they did it which makes it easy for anyone else to reverse engineer the process. That's what Altman is afraid of.