BrandBastion Blog

Social media engagement insights, news, and tips.

Specialized Comment Management Vs. Basic Filtering Tools

  • by: Crystal Cha
  • On: 13, Mar 2019
5 min read

Removing harmful comments, reacting promptly to threats, and providing a great customer experience on social media is increasingly important in achieving both brand and performance goals. When it comes to managing harmful comments, we set out to learn how big of a difference there is between what Facebook automatically filters out and what a specialized solution can moderate.

Social media has become a key marketing channel for major brands. But a brand’s ad spend and brand reputation can be significantly impacted when spam, scam or hate speech comments are made on a brand’s posts if these comments are not addressed.

Facebook has made significant efforts, pledging to increase its moderation and safety headcount by 10,000 by the end of 2018 and make changes to improve its spam and profanity filters. This has improved the experience on Facebook at large. However, when it comes to individual brand pages and communities, many large scale brands and advertisers still struggle to manage high comment volumes in-house and detect harmful comments on their brand properties.


What we analyzed

BrandBastion works with large gaming companies to protect them from harmful comments on social media. To understand the extent of harmful comment on Facebook and what Facebook auto-moderates, we analyzed all the engagement received on these company’s posts from for 6 months in 2018.

Note: Due to the nature of gaming, brands in the gaming industry tend to have among the highest levels of hate speech, profanity, and discrimination, as well as rampant posting of spam, hacks, and cheats that violate the games’ terms of use.


The results

906,476 comments were received between January and July 2018, including harmful comments. Some of these were automatically hidden by Facebook. However, there was a large amount of other harmful comments that were not hidden by Facebook that BrandBastion detected and hid.


The major types of harmful comments received for these gaming brands were Extreme Profanity, Hacks + Cheats, Spam + Scams, and Discrimination.

Harmful comments processed over time

Breakdown of harmful comments identified


The benefits of specialized moderation services

There are clear differences between Facebook's native profanity and spam filters and a third-party solution such as BrandBastion when it comes to coverage and accuracy. At BrandBastion, we have built classifiers that use Machine Learning and Natural Language Processing to determine if a comment is likely to belong to a certain categories that are sensitive for our clients, based on a wide range of signals. Signals taken into account include combinations of words and characters, syntactic structures of the comments, emojis used, and multiple other factors.

These classifiers are trained and validated on large language- and industry-specific datasets to guarantee a high level of confidence in classifying the content. We also consider each brand's individual guidelines when applying the classifiers, providing the required flexibility to fulfill the brand's specific needs. Additionally, we analyze new data processed to build a feedback loop that allows for constant improvement and and fine-tuning of the classifiers. This allows us to account for new trends and variations in the ever-evolving social media language.

Examples of comments not hidden by Facebook and detected by BrandBastion


TAKEAWAY 1: Facebook is great at detecting profanity, but not discrimination

Facebook auto-hid 34.74% of all harmful comments classified as Extreme Profanity, but only 16.27% of comments classified as Discrimination, which is often more subtle and contextual compared to profanity.

% of harmful comments detected and hidden by Facebook


TAKEAWAY 2: The accuracy of Facebook’s spam and profanity filters is increasing

As part of the moderation service provided to clients, BrandBastion also reviews content that is auto-hidden by Facebook and has the ability to unhide comments that are incorrectly hidden by Facebook’s filters.

During January to July 2018, BrandBastion unhid on average 44.67% of comments that had been auto-hidden by Facebook, but were not harmful.

However, the graph below shows that over time, the volume of what was unhidden decreased from 67.89% in January to 18.25% in July.

Volume of non harmful comments hidden by BrandBastion and Facebook

This indicates that the accuracy of Facebook’s auto-moderation algorithms is increasing, although the level of coverage seems to remain similar.


In summary

Social media is  an ever-evolving space and changes come from all directions, from algorithm changes to shifts in user behavior. Marketers that manage to ride the waves of change successfully are those who are data-driven, adaptable, and who use an arsenal of built-for-purpose tools as part of a comprehensive marketing stack that takes into account brand protection, community engagement, and performance, rather than looking at each piece in isolation. 

Related Posts

33 min read

The Best Guide to Social Media Brand Protection

Social media brand protection is a proactive approach to defending your brand from potential threats, and reputation dam...

Read full article
26 min read

How to Delete Comments on TikTok (And Why Hiding is Better)

TikTok has rapidly become one of the most popular social media platforms, allowing users to create and share short video...

Read full article
21 min read

How to Hide Comments on TikTok

TikTok has become a popular platform for sharing creative videos, but managing comments on your posts can sometimes be c...

Read full article

Let's Talk 

BrandBastion monitors, moderates, and responds to your brand's comments on social. We take action on your brand's behalf around the clock 24/7.
Book a Meeting