Proton working overtime to discourage me from renewing.
Technology
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
I don’t think they are that biased. They say in the article that ai models from all the leading companies are not private and shouldn’t be trusted with your data. The article is focusing on Deepseek given that’s the new big thing. Of course, since it’s controlled by China that makes data privacy even less of a thing that can be trusted.
Should we trust Deepseek? No. Should we trust OpenAI? No. Should we trust anything that is not developed by an open community? No.
I don’t think Proton is biased, they are explaining the risks with Deepseek specifically and mention how Ai’s aren’t much better. The article is not titled “Deepseek vs OpenAI” or anything like that. I don’t get why people bag on proton when they are the biggest privacy focused player that could (almost) replace google for most people!
Exactly.
Also, none of the article applies if you run the model yourself, since the main risk is whatever the host does with your data. The model itself has no logic.
I would never use a hosted AI service, but I would probably use a self hosted one. We are trying a few models out at work and we're hosting it ourselves.
Tutamail is a great email provider that takes security very seriously. Switched a few days ago and I'm very happy.
Yet not great from a privacy perspective. They don't even allow third party email apps.
Why do they even have to give their goddamn opinion? Who asked? Why should they car
It would be fair if ChatGPT or any american service received the same treatment, but the only article I found from 2023 seems quite neutral :/
We actually it seems quite fair-ish 🤷
AI has the potential to be a truly revolutionary development, one that could drive advancement for centuries. But it must be done correctly. These companies stand to make billions of dollars in revenue, and yet they violated our privacy and are training their tools using our data without our permission. Recent history shows we must act now if we’re to avoid an even worse version of surveillance capitalism.
Also from 2023 : https://proton.me/blog/ai-gdpr
I don’t see how what they wrote is controversial, unless you’re a tankie.
Given that you can download Deepseek, customize it, and run it offline in your own secure environment, it is actually almost irrelevant how people feel about China. None of that data goes back to them.
That's why I find all the "it comes from China, therefore it is a trap" rhetoric to be so annoying, and frankly dangerous for international relations.
Compare this to OpenAI, where your only option is to use the US-hosted version, where it is under the jurisdiction of a president who has no care for privacy protection.
TBF you almost certainly can't run R1 itself. The model is way too big and compute intensive for a typical system. You can only run the distilled versions which are definitely a bit worse in performance.
Lots of people (if not most people) are using the service hosted by Deepseek themselves, as evidenced by the ranking of Deepseek on both the iOS app store and the Google Play store.
Yeah the article is mostly legit points that if your contacting the chatpot in China it is harvesting your data. Just like if you contact open AI or copilot or Claude or Gemini they're all collecting all of your data.
I do find it somewhat strange that they only talk about deep-seek hosting models.
It's absolutely trivial just to download the models run locally yourself and you're not giving any data back to them. I would think that proton would be all over that for a privacy scenario.
It might be trivial to a tech-savvy audience, but considering how popular ChatGPT itself is and considering DeepSeek's ranking on the Play and iOS App Stores, I'd honestly guess most people are using DeepSeek's servers. Plus, you'd be surprised how many people naturally trust the service more after hearing that the company open sourced the models. Accordingly I don't think it's unreasonable for Proton to focus on the service rather than the local models here.
I'd also note that people who want the highest quality responses aren't using a local model, as anything you can run locally is a distilled version that is significantly smaller (at a small, but non-trivial overalll performance cost).
You should try the comparison between the larger models and the distilled models yourself before you make judgment. I suspect you're going to be surprised by the output.
All of the models are basically generating possible outcomes based on noise. So if you ask it the same model the same question five different times and five different sessions you're going to get five different variations on an answer.
You will find that an x out of five score between models is not that significantly different.
For certain cases larger models are advantageous. If you need a model to return a substantial amount of content to you. If you're asking it to write you a chapter story. Larger models will definitely give you better output and better variation.
But if you're asking you to help you with a piece of code or explain some historical event to you, The average 14B model that will fit on any computer with a video card will give you a perfectly serviceable answer.
Anyone promoting LLMs without a big side of skepticism is exposing their bias.
Now this is something people can be mad at