Does Transparency in Moderation Really Matter?: User Behavior After Content Removal Explanations on Reddit
When posts are removed on a social media platform, users may or may not receive an explanation. What kinds of explanations are provided? Do those explanations matter? Using a sample of 32 million Reddit posts, we characterize the removal explanations that are provided to Redditors, and link them to measures of subsequent user behaviorsâincluding future post submissions and future post removals.
Adopting a topic modeling approach, we show that removal explanations often provide information that educate users about the social norms of the community, thereby (theoretically) preparing them to become a productive member. We build regression models that show evidence of removal explanations playing a role in future user activity. Most importantly, we show that offering explanations for content moderation reduces the odds of future post removals. Additionally, explanations provided by human moderators did not have a significant advantage over explanations provided by bots for reducing future post removals.
We propose design solutions that can promote the efficient use of explanation mechanisms, reflecting on how automated moderation tools can contribute to this space. Overall, our findings suggest that removal explanations may be under-utilized in moderation practices, and it is potentially worthwhile for community managers to invest time and resources into providing them.
(I havenât read the contents of that paper btw, so this is just a general comment on moderation and moderation transparency.)
Discourse is already excellent and better than most forum software - users get notified of most mod actions. However I feel we could certainly do with expanding moderation features.
To give you an example, we routinely edit out bad language or personal attacks from posts and some users have complained about this - asking that they are asked to make the changes themselves and that we donât touch their posts.
Of course that is not practical - the whole point of a moderator removing personal attacks is so that the person that the attack is aimed at doesnât see it, which could lead to things spiralling. Additionally, moderators donât have the time to PM a user asking them to make the change, make a note of the incident, remember to check, PM the user again or take other action if they havenât made the required changes - itâs just too much of a burden on moderator teams.
An alternative could be to use the flag system which hides a post - but what happens if the user just unhides it or does not make an adequate change? Weâd still have to make a note to check, unhide again, etc. (Additionally, removing the post from view also has other negatives - can slow the flow of a conversation, can make people feel like itâs too much bother posting, etc - which can all avoided usually by a very simple mod edit.)
I do firmly believe that Discourse could take moderation to the next level just like they have in other aspects of forum software It would require a lot more thought than just a single post or thread though - moderation is hard (nobody likes being moderated!)
Iâd be happy to be part of such discussions and I am sure @HAWK would be invaluable in such an effort.
For heavy stuff like racist / sexist / bigoted invective I do not think there is any value in leaving toxic waste lying around to continually poison the community, e.g.
I completely agree - I have a note about this in our staff room:
Hi all,
One of our members has come out as _____ recently, and although it is unlikely that they will suffer any abuse from any genuine member of the community, itâs possible that someone with an axe to grind against [topic of forum] or the community may pick on them as an easy target or a way to create intercommunity conflict.
If a poster is ânewâ (i.e someone who has registered recently and has very few posts) and they make any racist, trans-phobic, homophobic or sexist remarks - please ban their account immediately and also remove any offending posts. The ban notice can simply be âComments contravene our Code of Conductâ.
If a poster is a regular member, please delete the offending post/s and PM the user to let them know the post was deleted because you believe it is against our rules and to please refrain from making any further posts in that thread (or contacting the other user if another user was involved) until a member of the Admin team has reviewed the situation and followed up/contacted them directly. If they ignore this notice and continue to post in the thread or continue to make abusive posts, suspend their account with the message âFailure to comply with Moderator request. Pending Admin review.â
Then please log any incidents in this thread and weâll review each case to see what action might be necessary. If Iâm not around and issues are ongoing (for example, where others are debating the issue in that or another thread) then just lock the thread with a âPending Admin reviewâ notice.
Itâs the other stuff thatâs tricky - Iâm off to bed now but will try to come back to this tomorrow.
(Are those staff notes publicly viewable or just by the person who made the post btw?)
We donât want to dictate how people moderate their communities, we just want to provide the tools so that they can do it as they see fit.
So Iâm curious here â what could be done differently? You can hide, you can unhide, you can delete, you can empower the community to hide, you can add staff notices.
With regards to such notices, we would almost certainly not use them for any kind of user-moderation because we prefer to avoid publicly embarrassing (or chastising) a user if at all possible - only dealing with user issues more publicly if we have been left with little choice (for instance, by their behaviour or conduct increasingly affecting more and more people on the forum).
If those notices could only be seen by the creator of that post, we could use them in some instances so long as we could make obvious that only they can see it.
Why do we take this path? Because people hate being moderated and being âshamedâ publicly even more - it can be infuriating (and somewhat disrespectful - if someone has made a genuine mistake they shouldnât be made to feel like an idiot in front of the whole community). Offending users like this can create huge problems further down the line - if they take umbrage they will never forget, and it will be compound future issues, which could lead to inter-community issues or them falling out with moderators (which unfortunately never ends well). At worst they could leave or get banned and carry out more malicious attacks. It is just not worth it, particularly if you were once reasonably confident they were a genuine user.
I agree - I just think at the moment the tools seem to reflect how moderation was traditionally carried out (with some improvements of course) but I think thereâs probably a much better way (just like you all at DC have demonstrated for other areas of forum software)⌠we just need to work out what that better path is.
Itâs a huge topic, but I think the main thing that requires our focus right now is making moderation as palatable as possible - while - not burdening the mod team.
Unfortunately I donât have the time to go into greater detail atm, but would love to chat about this more in the new year perhaps?
I spent many months thinking on this (while revisiting the way that we moderated SitePoint prior to moving to Discourse) and I feel like we reached a really good place [for that particular forum] using new processes which were driven by the tools available here. We pretty much threw everything out and started again, which was well overdue.
Iâm not sure what moderation problems we need to solve that arenât already being solved so Iâm certainly interested in your thoughts. I do concede that itâs difficult to think outside of the realm of the communities that youâve individually managed, but I do think itâs important to start with real problems, rather than assumptions.
This thread kinda went off on a tangent, but the main point I think we should take away from this article is that it is worth it to privately explain to users who post content against the guidelines why their own stuff is getting removed. The study indicated that the more detail provided, the better.
So, concretely, if somebodyâs content is removed because N people flagged them, I believe right now the automated message to the user says âyour post was removed because N people flagged you;â the message provides no information about why the post may have violated community guidelines.
Thatâs just an example. I feel like the overall gist is: removals should be associated with reasons that the post was removed, reasons that the user could act upon to change/fix the post, even (especially) if theyâre just canned explanations selected from a drop down.