Moderation is top of mind as media organizations aim to drive engagement, increase civility and create a positive user experience.
In a study done by Pew Research Center in 2010, 37% of internet users in the U.S. participated in the creation of news by commenting, social sharing and emailing. For perspective, that’s 85 million people—more than twice the population of Canada! This number has been on the rise ever since, as social networking and messaging apps have gained popularity.
Additionally, research by The Engaging News Project also shows that a large portion of an audience goes online specifically to engage in dialogue and participate in a community. For a media organization, this is great news—having an audience interested in spending time on your properties and sharing your content is essential for growth. But this increased engagement does come with a caveat. As more people frequent your comments section and website, the likelihood of getting a troll infestation rises. Trolls are those bad commenters attracted to the scent of a thriving community, whose main goal is to sow discord in your digital kingdom. These pests can not only derail conversations and drive away loyal subjects, but they can also tarnish your brand.
How do you stop them? Moderation.
Media organizations have relied on human moderation to weed out negative comments on their site, so audiences can participate in valuable conversations. This type of moderation has its inherent benefits, since humans can analyze context to make decisions, but it’s also very expensive and inefficient. That’s where automated moderation comes in. Let’s take a look at both below, and see which one triumphs in the war on trolls.
Beleaguered Knights: Human Moderators
Human moderation can occur either pre- or post-publication of a comment. Pre-moderation involves putting all reader comments in a queue, to be reviewed by a human before being published. This can be helpful on stories that incite a lot of heated opinions, where abusive comments are more likely. As Gulf News, UAE told WAN-IFRA, “The environment compels you to make sure there is not offensive content because we are in the Middle East. We don’t want content that is offensive or inflammable.” Post-moderation on the other hand, involves allowing all comments to publish to your site immediately, and then using humans to sift through and remove any inappropriate posts. This supports real-time conversations around time-sensitive stories and as a result, users are inclined to spend longer on your page, or first party data publishers.
The problem with human moderation is primarily cost and moderator bias. If you’re a publication with a large community, you’ll need a lot of moderators to review all the user-generated content. If you choose to selectively moderate, you open yourself to having potentially tarnishing material on your website. Additionally, while human moderators can account for context while moderating comments, their intrinsic biases may affect their decisions, leading to audience frustration. As a result, human moderation is not sustainable or scalable as your publication and community grows.
Steadfast Sentinel: Automated Moderation
With over 600 media brands using our Viafoura Engagement Cloud, we heard from our clients repeatedly that they needed a better way to moderate the trolls. That’s why we jumped into the fray with automated moderation.
Viafoura Automated Moderation allows you to moderate all comments as they’re submitted (in real-time), and learns from post-moderation changes. It reviews comments as they’re made, and publishes or flags them based on pre-defined criteria. Once flagged, these comments can either be deleted automatically or reviewed and approved/deleted by an in-house moderator. As a result, you can support real-time dialogue, since the automated system ensures that no hostility or vulgarity is published.
Automated moderation also learns from any of the post-moderation actions taken by your team, with self-learning algorithms deployed every day. Because of this, your comment-monitoring capabilities increase the more posts you generate. Each comment provides a new chance to learn the nuances of your community guidelines and adapt to fit your publication. Additionally, our research shows that automated moderation has an accuracy rate of 92%, compared to the human accuracy of 81%. This is supported by our highly adaptable algorithms that can be set to err on the side of caution.
If you do have in-house moderators, automated moderation allows them to focus on messages with flagged content. This reduces their time spent wading through spam, increases the quality of your publication (without incurring large staffing costs), and gives you 24/7 reliable coverage on your stories.
Visit our blog for more content on enriching and industry relevant topics.