this post was submitted on 25 Jun 2023
61 points (90.7% liked)

No Stupid Questions

39860 readers
1843 users here now

No such thing. Ask away!

!nostupidquestions is a community dedicated to being helpful and answering each others' questions on various topics.

The rules for posting and commenting, besides the rules defined here for lemmy.world, are as follows:

Rules (interactive)


Rule 1- All posts must be legitimate questions. All post titles must include a question.

All posts must be legitimate questions, and all post titles must include a question. Questions that are joke or trolling questions, memes, song lyrics as title, etc. are not allowed here. See Rule 6 for all exceptions.



Rule 2- Your question subject cannot be illegal or NSFW material.

Your question subject cannot be illegal or NSFW material. You will be warned first, banned second.



Rule 3- Do not seek mental, medical and professional help here.

Do not seek mental, medical and professional help here. Breaking this rule will not get you or your post removed, but it will put you at risk, and possibly in danger.



Rule 4- No self promotion or upvote-farming of any kind.

That's it.



Rule 5- No baiting or sealioning or promoting an agenda.

Questions which, instead of being of an innocuous nature, are specifically intended (based on reports and in the opinion of our crack moderation team) to bait users into ideological wars on charged political topics will be removed and the authors warned - or banned - depending on severity.



Rule 6- Regarding META posts and joke questions.

Provided it is about the community itself, you may post non-question posts using the [META] tag on your post title.

On fridays, you are allowed to post meme and troll questions, on the condition that it's in text format only, and conforms with our other rules. These posts MUST include the [NSQ Friday] tag in their title.

If you post a serious question on friday and are looking only for legitimate answers, then please include the [Serious] tag on your post. Irrelevant replies will then be removed by moderators.



Rule 7- You can't intentionally annoy, mock, or harass other members.

If you intentionally annoy, mock, harass, or discriminate against any individual member, you will be removed.

Likewise, if you are a member, sympathiser or a resemblant of a movement that is known to largely hate, mock, discriminate against, and/or want to take lives of a group of people, and you were provably vocal about your hate, then you will be banned on sight.



Rule 8- All comments should try to stay relevant to their parent content.



Rule 9- Reposts from other platforms are not allowed.

Let everyone have their own content.



Rule 10- Majority of bots aren't allowed to participate here. This includes using AI responses and summaries.



Credits

Our breathtaking icon was bestowed upon us by @Cevilia!

The greatest banner of all time: by @TheOneWithTheHair!

founded 2 years ago
MODERATORS
 

Hey! Thanks to the whole Reddit mess, I’ve discovered the fediverse and its increidible wonders and I’m lovin’ it :D

I’ve seen another post about karma, and after reading the comments, I can see there is a strong opinion against it (which I do share). I’d love to hear your opinions, what other method/s would you guys implement? If any ofc

(page 3) 50 comments
sorted by: hot top controversial new old
[–] AnonTwo@kbin.social 1 points 2 years ago

I don't think there needs to be another system, just a lower emphasis on said system. Like others have said, just post karma.

Someone pointed out that if you make your own instance, you can probably just manipulate your reputation anyway, so it won't be in any way a judge of a person. Even then it was really only useful in extremes, aka if the person had a massively negative rep they were probably a troll, and a near-0 was a new account. Anything other than that was more a metric of time than anything else.

Sure, even post karma is an abusable system, but if you have nothing then you'll be wading through people ignoring others or using circular reasoning to make bad faith arguments. The system is good for at least dealing with extremes.

[–] subignition@kbin.social 1 points 2 years ago (1 children)

I think voting has the potential to be very useful but what we want to move away from is the reactionary "I disagree with you / dislike your post, so downvote"

One experiment I would like to see is requiring a reason when downvoting. Factually incorrect, violates this magazine's civility code, trolling, etc. Some reasons might have overlap with the report feature, so a downvote for e.g. illegal content might automatically notify the moderators as well. This might be contingent on a feature that can impose restrictions for abuse of the report feature.

Although it is already relatively easy for anyone to notice when an account is blanket downvoting a thread. In theory, it's already a bit easier for users to sniff out bad actors because they'll either have a clear pattern of misbehavior or a conspicuous lack of account age or participation

[–] VGarK@lemmy.world 1 points 2 years ago

I like the idea of giving more information, however I’d say for both upvotes and downvotes. Me, as a user, I want to do things as quick as possible, so I’d say the action of upvoting/downvoting while giving more information should be embed into a quick wheel.

I like your comment, long press upvote and then select why in the categories. The same for downvoting

[–] Grabthar@lemmy.world 1 points 2 years ago

I think if we feel the need to do anything like it, chasing the number alone will be what makes people insufferable. Maybe track it invisibly on the instance where the account is hosted and display only a colour. Red for negative points, yellow up to maybe 50ish, and anything above is green. Enough to pick out the trolls while providing no reward for karma chasers. I don't think anything beyond that is useful.

[–] Rottcodd@kbin.social 1 points 2 years ago (2 children)

Any shortcut method of mimicking reputation can be and thus will be abused, so they're all toxic.

The only sure way to do it is the good old-fashioned way - by name recognition - actual, earned "reputation."

The way it used to work on all forums and still does on some smaller ones is that people just read posts and write their own posts and over time they come to recognize each other's names and associate them with some impression of each individual's value as a poster.

And yes - that's not very effective in gigantic forums, and it's not accessible to newcomers. You need a relatively small group of posters and new people have to pay attention in order to figure out who are the better or worse posters. That's just the way it is, and is one of the problems with gigantic forums.

[–] FaceDeer@kbin.social 1 points 2 years ago (1 children)

I'm not a fan of this because the main thing I liked about Reddit over all the other various "social media" was that I didn't have to try to recognize names. Names didn't matter, it was all about the conversation and the subject matter. If a post or comment was good I liked it, if it was bad I disliked it, and I moved on.

RES kept track of my personal like/dislike sum for each individual, perhaps since upvotes/downvotes aren't anonymous here in the Fediverse it'd be more straightforward to do that and show that to each user.

[–] schizanon@calckey.social 1 points 2 years ago

@FaceDeer@kbin.social I actively jam this by changing my name every so often like a molting arthropod

[–] AnonTwo@kbin.social 1 points 2 years ago

by name recognition - actual, earned "reputation."

That...is going to mean most people are untrustworthy after a few months. You won't recognize most people you speak to.

Keep in mind forums also usually only had maybe a 100 people at the most. We're dealing with potentially thousands in one thread.

I also find it really strange you gave that as "The Only Way", then by the end said that it doesn't really apply here.

[–] eating3645@lemmy.world 1 points 2 years ago* (last edited 2 years ago)

In my opinion the best alternative is a -1 : +1 scale. Members who contribute little are near 0, members who contribute a lot in a positive way get towards +1, if users contribute a lot in a negative way, their score goes to -1.

There are lots of different particular ways to implement this that isn't up vs. downvote count. Communities created, moderation activity, post count, engagement per post, positive reporting rate, false reporting rate, number of reports against the user, number of communites banned from, etc.

[–] TheBananaKing@lemmy.world 1 points 2 years ago (1 children)

I asked chatgpt, just to see what would filter out...

I would like some advice on designing a content-sorting and content-filtering mechanisms for discussion forums, in order to avoid or mitigates some of the problems that current systems are prone to.

One of the social problems with discussion forums results from the concept of upvoting or downvoting content.

It can be argued that voting content up or down serves to sort content by quality, allowing high-quality content to be seen and engaged with, while discouraging low-quality submissions.

However, in practice, Goodhart's law ends up applying: when a measure becomes a goal, it stops being a useful measure.

By using popularity as a proxy metric to determine quality, and by rewarding popularity with positive attention, this ends up selecting for content with superficial emotional appeal - ragebait, memes, facile/obvious comments pandering to common sentiment, puns, etc. - and not 'useful' content that is thoughtful, incisive, analytical, or important-yet-unpalatable. Ironically, content of this nature is also low-quality in a different way.

Worse, this ends up training users to produce content of this nature at the expense of thoughtful, interesting, incisive discussion, reducing both the production of quality content, and the quality of the collaborative sorting that users perfom by way of voting.

A cumulative 'karma' score for users, being a sum total of their upvotes and downvotes across all of their submissions, while ostensibly encouraging engagement and admirable behaviour... can again end up being gamified, and lead to users making large numbers of popular but low-effort submissions in order to maximise their score.

This can also produce 'filter bubbles' or 'hive minds': subcultures or communities that reject and discourage disagreement or criticism, and end up with a positive feedback loop leading to increasing orthodoxy with little grounding in reality. In some cases, the communities involved can become radicalised or toxic as a result, as more-moderate voices are suppressed from the discussion, and increasingly-extreme views become required in order to gain attention. Of course, it's reasonable and useful to let groups based around common ideals preserve an individual identity separate from the mainstream of discussion, but perhaps a way to regulate excessive self-selection would be useful.

Doing away with any kind of quality-sorting mechanic altogether is not a very useful alternative; a forum where spam, antisocial, or irrelevant submissions are given equal weight and visibility as interesting or thought-provoking posts tends to discourage users from participating or even reading.

As a further complication, users do not generally follow instructions or policy set out for voting on posts, and will tend to use them to indicate approval/agreement regardless.

Human moderation is an important part of any discussion forum and essential to the guidance ans shaping of any community, but the workload does not scale well; it should be considered a limited and intensive resource.

It would seem there are a few challenges to overcome:

1: Better-define the concept of 'quality' for discussion-forum posts beyond immediate, superficial emotional appeal - what is it that makes discussions feel satisfying, and drives deeper engagement? 2: Find a way to programatically ascribe a metric based on this concept, via the engagement of other users, in any form including upvotes, downvotes, tagging, commenting - and statistical or derived data such as the number, depth and overall score of replies to the post, and perhaps even analysis of the text of replies. 3: Ensure that any such system is difficult for individual users to exploit or game; for instance a naive reply-count metric could be trivially exploited by a user posting a large number of replies to a single comment.
4: Limit the effect of self-selection feedback loops, to prevent radicalisation while still allowing common ideals and identity to flourish in their own spaces.

One idea I've had would be to sort all content by 'controversialness' - that is to say, posts with a high number of total votes, but a low-magnitude score once downvotes are subtracted from upvotes. I believe this would tend to deprioritize low-effort posts such as ragebait or pandering to common sentiment, along with junk, spam and boring or irrelevant content, and thus bring 'interesting' content to the top. I believe it would be difficult to game for attention, or for other users to have undue effect on its visibility. However, I'm not sure whether controversialness would be a good metric to drive engagement, or what large-scale effects sorting on this metric might have on a community.

Can you offer any advice or commentary on the situation, or identify any important flaws in my analysis? Can you suggest any alternatives that might work better?

===================================================================================== ChatGPT You've outlined several important challenges and considerations when it comes to designing content-sorting and filtering mechanisms for discussion forums. The issues you've raised are indeed prevalent in many online communities, and finding effective solutions can be complex. I'll provide some advice and commentary on your analysis, as well as suggest alternative approaches that might address the problems you've highlighted.

Better defining quality: You're right in emphasizing the need to define quality beyond superficial emotional appeal. One way to approach this is by incorporating multiple dimensions of quality, such as insightful analysis, thought-provoking content, well-reasoned arguments, and valuable information. Encouraging constructive and meaningful discussions should be a primary goal.

Ascribing metrics based on user engagement: Rather than relying solely on upvotes and downvotes, consider implementing a more comprehensive metric that takes into account various aspects of user engagement. This could include factors such as the number of replies, the depth of discussion, the length of time spent reading the post, and even sentiment analysis of the comments. By incorporating a broader set of metrics, you can obtain a more nuanced understanding of user engagement and use it to evaluate content quality.

Avoiding gaming and exploitation: To mitigate the risk of individuals exploiting the system, it's crucial to design mechanisms that discourage or penalize excessive or low-quality contributions. For example, you could implement rate-limiting measures that prevent users from making an excessive number of submissions within a certain time frame. Additionally, considering the quality of the engagement (e.g., the relevance, thoughtfulness, and contribution to the discussion) rather than just the quantity (e.g., sheer number of replies) can help minimize gaming attempts.

Addressing self-selection feedback loops: To prevent echo chambers and radicalization, it's important to strike a balance between allowing communities with common ideals to flourish and promoting diverse perspectives. You could introduce mechanisms that encourage exposure to alternative viewpoints, such as algorithms that prioritize showing posts to users with different preferences or even implementing mandatory or randomized exposure to different topics or communities. Additionally, fostering a culture of respectful and open-minded discussion can help create an environment where diverse opinions are valued.

Regarding your idea of sorting content by controversialness, it could be an interesting approach to promoting engaging discussions. However, it's important to carefully consider the potential effects of such sorting on the community. Controversial topics and discussions can be divisive, and if not managed well, they may lead to increased conflict and hostility. Implementing safeguards, such as active moderation and community guidelines, would be essential in maintaining a healthy and constructive environment.

In conclusion, tackling the challenges of content sorting and filtering in discussion forums requires a multifaceted approach. By defining quality beyond superficial appeal, incorporating comprehensive engagement metrics, preventing gaming and exploitation, and addressing self-selection feedback loops, you can design a system that promotes meaningful discussions while minimizing the negative effects associated with current mechanisms. It's crucial to iterate and refine these systems based on user feedback and community dynamics to ensure the best possible experience for forum participants.

[–] ultimate_question@lemmy.world 1 points 2 years ago

This is a great example of how aggressively useless chatgpt is at this kind of thing, the response is literally just the prompt rephrased and packaged like an answer lol

[–] mykl@lemmy.world 1 points 2 years ago (2 children)

Every system that can be thought of (and has been suggested here) might sound great but when implemented at scale will no doubt prove to be open to abuse and require an army of mods to oversee. Otherwise every multi-million dollar social media company would have implemented it already.

Upvotes and downvotes and cumulative scores kind of do the job well enough that that’s what we keep ending up with.

That being said though, I would be interested in seeing a system where each downvote you make also counts against your own karma to discourage profligate use of the downvote to mean “I have a different opinion but can’t express it here”.

[–] Aceticon@lemmy.world 1 points 2 years ago* (last edited 2 years ago)

Surely by the same logic upvoting without providing a reason for it should also be decentivised - why should negative feedback require taking the time to explain "why", whilst positive feedback would not - logically either they both require a "why" or none does.

An uneven posture when it comes to receiving feedback only makes sense if one is emotionally impacted by "somebody disagreed with me and didn't told me why" and having such a socially fragile ego is really the problem of that person, not of everybody else.

More generally and as I pointed out in a long post which I made in the other Karma thread (which I will not repeat here), the removing (or punishing) downvotes is just a strategy to incentivise more content posting, no matter how mediocre, which in turn leads to a a lower signal to noise ratio (i.e. more mindles fluff less content) which is bad for everybody - no-work negative criticism (i.e. downvotes without the need to spend time making explanatory posts) are quite an effective way of providing feedback on the shoddiness of something without the artificial barrier against criticism which is to require spending time on an explanation - I mean, if 1 or 2 downvotes get to you, then you definitelly have emotional issues you need to explore with an expert in such things as a handful of anonymous "I don't like that" can be easilly dismissed as "there are a handful of people who disagree with what I wrote (so what?!)", whilst an unexpected 10 or 20+ downvotes are often a pretty good hint to think again about what your published.

It seems to me that it's incredibly selfish and self-centred to demand that everybody else takes the time to write an explanation when you write something they disagree with: other people's time is their own and they do not exist merelly to serve your ego just as you don't exist merelly to serve theirs.

Mind you, I do think it would be fair for there to be some way for people to disable viewing of downvotes on their account, as people with such "sensitivity" to negative feedback deserve to be able to participate in social media just like everybody else and since Lemmy keeps track of both negative and positive votes getting the interface to just show the positive stuff should be reasonably easy and it would protect the ego of those who need such protection.

[–] VGarK@lemmy.world 1 points 2 years ago (1 children)

I think the idea of sacrificing your own “good boi points” to downvote a post bring a new layer of complexity. For sure, the hive-mind effect would be attenuated if users had to sacrifice their own “points”, and, probably, the downvoting as a whole would be less used. What do you guys think about how the use of the downvote would change?

[–] mykl@lemmy.world 1 points 2 years ago

Oh I doubt anyone would implement that as a system, but it is interesting to speculate what it might change. Imagine all the downvote curmudgeons having to regularly post pictures of kittens to /c/aww in order to recharge their karma.

I rarely downvoted on Reddit because I’m pretty sure everyone (no matter how blasé they appear to be) gets more upset by a comment ending up at -1 than they would get pleased by it ending at +5. When I found that some instances here don’t even have the downvote button, I decided that I wasn’t going to use it here at all. If someone says something damaging to the conversation I’ll report it to the mods, if they say something factually incorrect I might correct them, but otherwise I will just move on.

[–] bstix@feddit.dk 1 points 2 years ago* (last edited 2 years ago) (3 children)

How about expanding the votes into multiple kinds of karma. Make it possible to place several votes: "on topic/off topic", "funny/boring", "shit post", "solution to the question", "agree/disagree", "political", "Interesting", "Spam", "Original Content", etc.

Communities could create whatever rating is suitable for their forum.

Sort of like tags, but votable.

It would basically reward everyone for what they do (being a level 7 funny shitposter is something) but at the same time making it possible for others to filter out anything they don't care about. So instead of clicking downvote because of disagreement or upvote because it's funny, there would be an outlet for that in its own vote.

I feel that would make it easier to find quality content whether you're looking for serious debate or the hottest memes.

We'd need better comment filtering on individual communities, and it could/would be abused, but overall it would be facilitate the possibility of having different kinds of conversations on the same topic.

Sometimes I want to read funny stuff in serious topics and sometimes there is serious stuff in funny threads, and sometimes people write clever stuff that I disagree with and so on. One vote is just not enough.

[–] Matcraftou@lemmy.world 1 points 2 years ago

Yes, I like that, there should be like 3 dots with a up/downvote logo on top of it and when you click it you would have the options to like: -downvote: off topic -Upvote: Intersesting -Upvote: Original content

And so on

load more comments (2 replies)
[–] Dick_Justice@lemmy.world 1 points 2 years ago

Other methods of what?

load more comments