this post was submitted on 11 Sep 2023
1132 points (96.0% liked)
Technology
59143 readers
2285 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I think the thing isn't just providing access to the content, but using algorithms to promote how likely it is for deranged people to view more and more content that fuel their motives for hateful acts instead of trying to reduce how often that content is seen, all because they make more money if they watch more content, wether it is harmful or not.
This.
I don't know about Reddit, but YouTube 100% drives engagement by feeding users increasingly flammable and hateful content.
Hell they'll even take ad money to promote Jan 6th conspiracies
Yeah, the difference is in whether or not the company is choosing what to put in front of a viewer's eyes.
For the most part an ISP just shows people what they request. If someone gets bomb making directions from YouTube it would be insane to sue AT&T because AT&T delivered the appropriate packets when someone went to YouTube.
On the other end of the spectrum is something like Fox News. They hire every host, give them timeslots, have the opportunity to vet guests, accept advertising money to run against their content, and so on.
Section 512 of the DMCA treats "online service providers" like YouTube and Reddit as if they're just ISPs, merely hosting content that is generated by users. OTOH, YouTube and Reddit use ML systems to decide what the users are shown. In the case of YouTube, the push to suggest content to users is pretty strong. You could argue they're much closer to the Fox News side of things than to the ISP side these days. There's no human making the decisions on what content should be shown, but does that matter?
Yep. I often fall asleep to long YouTube videos that are science or history related. The algorithm is the reason why I wake up at 3am to Joe Rogan. It’s like a terrible autocomplete.
The algorithm is tailored to you. This says more about you. I never get recommended Rogan.
Absolutely. I saw a Google ad the other day from maybe PragerU that was about climate change not being real, while I was searching for an old article that was more optimistic about outcomes. They actually said by the ad that they were showing it as a suggested thing, and thankfully you could report it, which I did immediately. It pissed me off a ton.
A friend recently shared a similar suggested video/ad they got on YouTube, which was saying "Ukrainians are terrorists". PragerU or TPUSA.
I can see the argument for allowing these ads to exist as a freedom of speech thing, fine. But actively promoting these ads is very different. The lawsuit would have merits on this. I'd prefer if this content was actively minimized, but at the very least it shouldn't be promoted.
What if it isn't algorithms but upvotes? What if Lemmy is next?
and we all know what reddit mods do.