this post was submitted on 28 Jul 2023
21 points (95.7% liked)

Technology

59143 readers
2314 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

hmm, strange article tbh, not sure what to say about it, i wouldnt trust the govt with anything though

top 4 comments
sorted by: hot top controversial new old
[–] [email protected] 6 points 1 year ago* (last edited 1 year ago) (1 children)

The article is meh, but what is said in there is pretty much true.

Unregulated technology is something we have now : ai generation for example.

What is happening right now with the ai generation be text or image, is that they are the most privacy invasive thing there could be.

For example chatgpt/Bing chat, or Google bard. They take every bit of text, analyse it and use it for future prompts.

User do not know because it's too long to read their huge privacy policy.

And so people are puting in those prompts private elements, trade secrets, and elements they should not put in there.

Image generation is currently a huge copyright issue.

Worldcoin deployed balls, installed like art in multiple cities, with a camera to scan the eyeballs (and so their identity, as it seems to be unique for each person) of every person who just gazes into that lens. https://news.artnet.com/art-world/worldcoin-orb-ai-2341500

So non regulated tech is a huge anti-people / pro money mess.

For brain interfaces it may even be worse. Companies could just put ads in your brain directly, and everything that was described in that article.

I saw a little part of a movie some long time ago, no idea what it was. There was some brain interface allowing communication and displaying images.

When just enabled it was a huge mess adverts everywhere, noise ads, so bad that it would make that character unable to think and know what he wanted to do.

In order to get back into what he had to do he had to use dampener to remove all those ads and noise, all the mess.

And the first part with so much noise that we cannot do anything from our own will is what will happen for non regulated brain interface after enough people have adopted it.

And we may not have such brain "ad blocking" tech as in the movie.

[–] [email protected] 4 points 1 year ago* (last edited 1 year ago) (1 children)

The data broker industry is a far better example of the dangers of unregulated tech than ai generated works — which is more a rebuke of copyright and a threat to labor than a privacy invasion.

[–] [email protected] 1 points 1 year ago

It may be even worse as you said, however AI currently is more present in the news and maybe easier to understand because of this.

Also chatgpt had a huge amount of personal info leaked to the dark net, not really because they got hacked, but because the users put their login credentials into fishing websites.

But also, as any thing you input into chatgpt/Bing chat/bard is scanned, it can also be a big antitrust/corporate espionage as openai/microsoft and Google may be able to spy on any users who may leak the development of another AI.

[–] [email protected] 4 points 1 year ago

I would trust a for profit company even less than the government.

load more comments
view more: next ›