Downvotes only so karma whores never comment, and completely random monthly account bannings so no one gets too comfortable.
No Stupid Questions
No such thing. Ask away!
!nostupidquestions is a community dedicated to being helpful and answering each others' questions on various topics.
The rules for posting and commenting, besides the rules defined here for lemmy.world, are as follows:
Rules (interactive)
Rule 1- All posts must be legitimate questions. All post titles must include a question.
All posts must be legitimate questions, and all post titles must include a question. Questions that are joke or trolling questions, memes, song lyrics as title, etc. are not allowed here. See Rule 6 for all exceptions.
Rule 2- Your question subject cannot be illegal or NSFW material.
Your question subject cannot be illegal or NSFW material. You will be warned first, banned second.
Rule 3- Do not seek mental, medical and professional help here.
Do not seek mental, medical and professional help here. Breaking this rule will not get you or your post removed, but it will put you at risk, and possibly in danger.
Rule 4- No self promotion or upvote-farming of any kind.
That's it.
Rule 5- No baiting or sealioning or promoting an agenda.
Questions which, instead of being of an innocuous nature, are specifically intended (based on reports and in the opinion of our crack moderation team) to bait users into ideological wars on charged political topics will be removed and the authors warned - or banned - depending on severity.
Rule 6- Regarding META posts and joke questions.
Provided it is about the community itself, you may post non-question posts using the [META] tag on your post title.
On fridays, you are allowed to post meme and troll questions, on the condition that it's in text format only, and conforms with our other rules. These posts MUST include the [NSQ Friday] tag in their title.
If you post a serious question on friday and are looking only for legitimate answers, then please include the [Serious] tag on your post. Irrelevant replies will then be removed by moderators.
Rule 7- You can't intentionally annoy, mock, or harass other members.
If you intentionally annoy, mock, harass, or discriminate against any individual member, you will be removed.
Likewise, if you are a member, sympathiser or a resemblant of a movement that is known to largely hate, mock, discriminate against, and/or want to take lives of a group of people, and you were provably vocal about your hate, then you will be banned on sight.
Rule 8- All comments should try to stay relevant to their parent content.
Rule 9- Reposts from other platforms are not allowed.
Let everyone have their own content.
Rule 10- Majority of bots aren't allowed to participate here.
Credits
Our breathtaking icon was bestowed upon us by @Cevilia!
The greatest banner of all time: by @TheOneWithTheHair!
I loved the random monthly account banning 😂 I’d do it pemaban and hold a big event to decide the lucky winners~
This may be a dumb Idee so please go ahead and tell me if it is. It seem ppl use it to know who to block in advance. What if you get a red name or some kind of info on the profile if the account has been blocked by other users and it is above 10 or 20 blocks? Would that help? That would suck for that account because it will forever be the ass hole account. But at least no one would really want to farm that except the trolls who want ppl to know they are trolls.
I like how it is now tho. It is good when the mods are responsible first and foremost instead of a system.
I think it would be better to have like a currency system where posts that are kept alive the longest trigger points, not just how many people upvote them. But then, you should be able to use those points to do something instead of hoarding them like a dragon's treasure or maybe turn them in to awards. If OTHER people give you awards, that's what you should have on display, not just how many upvotes you had. This would also give you more points for helping smaller communities create meaningful content instead of what's popular.
I'm not from reddit, what is a reputation and what practical effect does it have? Is it just upvotes minus downvotes?
I found a reputation in my profile of "1" but I don't have a clue where that came from. I'm not sure why we need to have scores associated with our accounts, that in itself seems toxic to me to care about (clout chasing).
Upvoting comments in threads makes sense, I'm just not seeing any actual practical connection with the thing called "reputation" on my profile. What does it do in a best case?
I was thinking maybe some kind of ranking system like Street Fighter 6?
I know everyone seems to be hating karma but I do like that dopamine release. Ofx it will get abused... but what if there are just tiers, rather than seeing a number go up.
And at the highest tier, it doesn't matter anymore. That was you can see who is most active and it kind of gives just a bit of prestige. Furthermore, you won't see a number going up forever, so after awhile it's not like you want to keep gaming the system to see the number go up. But at the same time you can feel some some of progression.
Anyways, it's just a random thought I had as I am grinding on SF6 today haha. I could easily do without karma but it's just a thought.
From this and other posts on this as well as comments I read and discussions that made me think about it, here's my suggestion.
- Upvotes and downvotes but lemmy allows people to only see upvotes in their client if they wish to (be it because they don't like the "negativity" of downvotes or because they're not very good at emotionally dealing with seeing their own comments downvoted)
- Some kind of summary of upvotes/downvotes a user got on his or her posts, per forum and only if enabled in that forum. The objective being to as much as possive avoid the gamification side of karma and its side effects (i.e. people taking it in as a "score" which leads to things like karma farming) whilst preserving the positive side of it as a measure of domain expertise or at least willingness to positivelly participate in domain specific forums.
Thats just like 9gag, reddit, and the king youtube bs. If someone sees downvotes as negativity, it is already to late for him on a normal platform. A downvote should be something the author should see and know about. And other users should see that too. As it is a good content filter against clickbaity/scammer or just chatgpt answers.
Whilst I personally agree with you that it's a good idea for an author to see both negative and positive feedback, I don't see what's the problem to allow the "criticism sensitive" types to protect themselves from criticism by toggling and option in their client which makes downvotes not visible to them (but it does nothing for others).
For clarity, my suggestion is not to allow people to disable downvotes, it's to allow them to not see downvotes in their own client - downvotes are still there and everybody else who hasn't "disabled downvote display" on their clients can still see them.
Think of it as a social interaction equivalent of installing ramps for people with disabilities on wheel chairs: nobody else is force to use the ramps but those who need them have them.
It's very easily abused. Does Karma affect article and comment visibility on Reddit? I don't know the details, but if so I'd suggest that it not do so here. Maybe just have it be a number calculated from boosts, upvotes and downvotes that you can see on the profile if you are a mod trying to determine if someone tends to troll, but not something that has any affect on whether or not your stuff is displayed.
Sort of.
On reddit, moderators can set minimum karma thresholds to control who can post in the sub, admins can use it to control who posts anywhere sitewide.
For example, new users who make a contentious statement and get down voted for it will suddenly find themselves rate limited for posting comments. They could be in the middle of an exchange where the other user is firing back responses, but then suddenly they are restricted to making one post every few minutes in the sub.
Admin can also use that to automate things like shadow banning.
I for one wish we could return to the lawlessness of old phpBB boards from the late 90s. Weren't no karma points to work any back then.
Those boards were rife with toxicity. You’d have secret rooms with people attacking a commenter behind their back. Created a very ‘with us or against us’ vibe. Then it would hit a boiling point and spill out as a public attack on a commenter. It made the bullying premeditated and targeted.
Non-algorithmic ordering, auto-collapse replies after a certain user-preferred setting.
User scores are bad. Up/downvotes are bad.
The whole point of them was to create a flow of content with minimum human intervention. That’s a huge goal and The Dream if you’re making money off social media. If you’re not making money off social media then it’s not doing you any good.
Abolish karma, abolish comment and post scores.
I asked chatgpt, just to see what would filter out...
I would like some advice on designing a content-sorting and content-filtering mechanisms for discussion forums, in order to avoid or mitigates some of the problems that current systems are prone to.
One of the social problems with discussion forums results from the concept of upvoting or downvoting content.
It can be argued that voting content up or down serves to sort content by quality, allowing high-quality content to be seen and engaged with, while discouraging low-quality submissions.
However, in practice, Goodhart's law ends up applying: when a measure becomes a goal, it stops being a useful measure.
By using popularity as a proxy metric to determine quality, and by rewarding popularity with positive attention, this ends up selecting for content with superficial emotional appeal - ragebait, memes, facile/obvious comments pandering to common sentiment, puns, etc. - and not 'useful' content that is thoughtful, incisive, analytical, or important-yet-unpalatable. Ironically, content of this nature is also low-quality in a different way.
Worse, this ends up training users to produce content of this nature at the expense of thoughtful, interesting, incisive discussion, reducing both the production of quality content, and the quality of the collaborative sorting that users perfom by way of voting.
A cumulative 'karma' score for users, being a sum total of their upvotes and downvotes across all of their submissions, while ostensibly encouraging engagement and admirable behaviour... can again end up being gamified, and lead to users making large numbers of popular but low-effort submissions in order to maximise their score.
This can also produce 'filter bubbles' or 'hive minds': subcultures or communities that reject and discourage disagreement or criticism, and end up with a positive feedback loop leading to increasing orthodoxy with little grounding in reality. In some cases, the communities involved can become radicalised or toxic as a result, as more-moderate voices are suppressed from the discussion, and increasingly-extreme views become required in order to gain attention. Of course, it's reasonable and useful to let groups based around common ideals preserve an individual identity separate from the mainstream of discussion, but perhaps a way to regulate excessive self-selection would be useful.
Doing away with any kind of quality-sorting mechanic altogether is not a very useful alternative; a forum where spam, antisocial, or irrelevant submissions are given equal weight and visibility as interesting or thought-provoking posts tends to discourage users from participating or even reading.
As a further complication, users do not generally follow instructions or policy set out for voting on posts, and will tend to use them to indicate approval/agreement regardless.
Human moderation is an important part of any discussion forum and essential to the guidance ans shaping of any community, but the workload does not scale well; it should be considered a limited and intensive resource.
It would seem there are a few challenges to overcome:
1: Better-define the concept of 'quality' for discussion-forum posts beyond immediate, superficial emotional appeal - what is it that makes discussions feel satisfying, and drives deeper engagement?
2: Find a way to programatically ascribe a metric based on this concept, via the engagement of other users, in any form including upvotes, downvotes, tagging, commenting - and statistical or derived data such as the number, depth and overall score of replies to the post, and perhaps even analysis of the text of replies.
3: Ensure that any such system is difficult for individual users to exploit or game; for instance a naive reply-count metric could be trivially exploited by a user posting a large number of replies to a single comment.
4: Limit the effect of self-selection feedback loops, to prevent radicalisation while still allowing common ideals and identity to flourish in their own spaces.
One idea I've had would be to sort all content by 'controversialness' - that is to say, posts with a high number of total votes, but a low-magnitude score once downvotes are subtracted from upvotes. I believe this would tend to deprioritize low-effort posts such as ragebait or pandering to common sentiment, along with junk, spam and boring or irrelevant content, and thus bring 'interesting' content to the top. I believe it would be difficult to game for attention, or for other users to have undue effect on its visibility. However, I'm not sure whether controversialness would be a good metric to drive engagement, or what large-scale effects sorting on this metric might have on a community.
Can you offer any advice or commentary on the situation, or identify any important flaws in my analysis? Can you suggest any alternatives that might work better?
===================================================================================== ChatGPT You've outlined several important challenges and considerations when it comes to designing content-sorting and filtering mechanisms for discussion forums. The issues you've raised are indeed prevalent in many online communities, and finding effective solutions can be complex. I'll provide some advice and commentary on your analysis, as well as suggest alternative approaches that might address the problems you've highlighted.
Better defining quality: You're right in emphasizing the need to define quality beyond superficial emotional appeal. One way to approach this is by incorporating multiple dimensions of quality, such as insightful analysis, thought-provoking content, well-reasoned arguments, and valuable information. Encouraging constructive and meaningful discussions should be a primary goal.
Ascribing metrics based on user engagement: Rather than relying solely on upvotes and downvotes, consider implementing a more comprehensive metric that takes into account various aspects of user engagement. This could include factors such as the number of replies, the depth of discussion, the length of time spent reading the post, and even sentiment analysis of the comments. By incorporating a broader set of metrics, you can obtain a more nuanced understanding of user engagement and use it to evaluate content quality.
Avoiding gaming and exploitation: To mitigate the risk of individuals exploiting the system, it's crucial to design mechanisms that discourage or penalize excessive or low-quality contributions. For example, you could implement rate-limiting measures that prevent users from making an excessive number of submissions within a certain time frame. Additionally, considering the quality of the engagement (e.g., the relevance, thoughtfulness, and contribution to the discussion) rather than just the quantity (e.g., sheer number of replies) can help minimize gaming attempts.
Addressing self-selection feedback loops: To prevent echo chambers and radicalization, it's important to strike a balance between allowing communities with common ideals to flourish and promoting diverse perspectives. You could introduce mechanisms that encourage exposure to alternative viewpoints, such as algorithms that prioritize showing posts to users with different preferences or even implementing mandatory or randomized exposure to different topics or communities. Additionally, fostering a culture of respectful and open-minded discussion can help create an environment where diverse opinions are valued.
Regarding your idea of sorting content by controversialness, it could be an interesting approach to promoting engaging discussions. However, it's important to carefully consider the potential effects of such sorting on the community. Controversial topics and discussions can be divisive, and if not managed well, they may lead to increased conflict and hostility. Implementing safeguards, such as active moderation and community guidelines, would be essential in maintaining a healthy and constructive environment.
In conclusion, tackling the challenges of content sorting and filtering in discussion forums requires a multifaceted approach. By defining quality beyond superficial appeal, incorporating comprehensive engagement metrics, preventing gaming and exploitation, and addressing self-selection feedback loops, you can design a system that promotes meaningful discussions while minimizing the negative effects associated with current mechanisms. It's crucial to iterate and refine these systems based on user feedback and community dynamics to ensure the best possible experience for forum participants.
A blockchain might be fun as long as there is no way of converting it into money. Like just a ledger of how much "karma" everyone gets