I found that idea interesting. Will we consider it the norm in the future to have a "firewall" layer between news and ourselves?
I once wrote a short story where the protagonist was receiving news of the death of a friend but it was intercepted by its AI assistant that said "when you will have time, there is an emotional news that does not require urgent action that you will need to digest". I feel it could become the norm.
EDIT: For context, Karpathy is a very famous deep learning researcher who just came back from a 2-weeks break from internet. I think he does not talks about politics there but it applies quite a bit.
EDIT2: I find it interesting that many reactions here are (IMO) missing the point. This is not about shielding one from information that one may be uncomfortable with but with tweets especially designed to elicit reactions, which is kind of becoming a plague on twitter due to their new incentives. It is to make the difference between presenting news in a neutral way and as "incredibly atrocious crime done to CHILDREN and you are a monster for not caring!". The second one does feel a lot like exploit of emotional backdoors in my opinion.
Why do people, especially here in the fediverse, immediately assume that the only way to do it is to give power of censorship to a third party?
Just have an optional, automatic, user-parameterized, auto-tagger and set parameters yourself for what you want to see.
Have a list of things that should receive trigger warnings. Group things by anger-inducing factors.
I'd love to have a way to filter things out by actionnable items: things I can get angry about but that I have little ways of changing, no need to give me more than a monthly update on.
Because your "auto-tagger" is a third party and you have to trust it to filter stuff correctly.
How about no? You set it up with your parameters, it is optional and open source.