The Case for Automated Content Moderation
Let’s go step by step.
What is moderation?
Moderation, as commonly understood, is the process of guiding a discussion or movement away from extremes. Exercising moderation requires one to be within reasonable limits, as defined by law or by societal norms. It is the act of being calm or measured.What is content?
While the English word ‘content’ has many meanings, in the context in which we are using the word, in the media sense, in the publishing and communication sense, content is “the information and experiences that are directed toward an end user. Content is “something that is to be expressed through some medium, as speech, writing or any of various arts. Content can be delivered via many different media including the Internet, cinema, television, radio, smartphones, audio CDs, books, e-books, magazines, and live events, such as speeches, conferences, and stage performances.” With the internet penetrating to the deepest corners of mankind, content is being created at a mind-boggling rate. Every time we comment on a post, we create content. Each review we post of a new product, we create content. Each image we share of a vacation, we create content. Now multiply that with the several billion people doing similar things, and you can perhaps get an idea of the scale of content being created, and consumed, today. This vast and distributed content creating machinery that the world has become, thanks to the internet and social media, has a free hand with what they can create. While in their private lives they may be free to do what they please, within the confines of civil liberties and local regulations, when their content becomes available widely, it needs to adhere to the basic human and societal guidelines, written or not. It also needs to adhere to the defined, usually written, content rules of the platform they are creating their content on. This gives rise to the need for content moderation. The practice of monitoring and determining suitability of user generated content (UGC) to be visible to all comers, is known as content moderation. It is generally performed against pre-defined and generally accepted rules and guidelines. oWorkers has been moderating content generated on web properties owned and maintained by its clients for over 7 years. As traffic has increased, so has our capability in ensuring that communities adhere to the objectives with which they were created. We have been selected as one of the top three global providers in the space of data services on multiple occasions. We focus relentlessly on our chosen areas of work in data based BPO services. Before we move to automated content moderation, let us take a quick look at the different ways in which moderation can be done.Methods of moderating content
Assuming content moderation to be a given, something that needs to be done, it can be done manually or it can be automated. Of course, a combination of different techniques is always a choice that can be made. And, if one were to assume that moderation is not a given, that could operate at the top level as a third method; no moderation.Manual moderation
Before any process is automated, for reasons of efficiency, volume, standardization, or any other, it is done manually. Manual moderation can be further subdivided, usually based on the stage at which the exercise is being carried out:- Pre-moderation – This works like an approval queue, where content is published only after a moderator has been able to review it and stamp her approval on it. It can also be called pre-publish moderation. Though it has its limitations like content visibility being delayed and stifling of open communication, for sensitive websites and subjects, this may be suitable.
- Post-moderation – As the name might suggest, publishing of the content is permitted, and subsequently reviewed and removed, if found objectionable. It can also be called post-publish moderation. While it permits open communication and healthy sharing, damage could be caused by unsavory content.
- Reactive moderation – Reliance is placed on the community visiting and participating in that online forum to flag content that is out of place in that setting. This can work where members are deeply invested in the community and are keen to ensure its success. Otherwise, this might be better used as an additional check.
- Distributed moderation – This relies on a rating mechanism where members rate content on the basis of its relevance. Content that is rated low gets pushed down in relevance and eventually almost vanishes from view. Like reactive moderation, business-run communities are generally reluctant to leave moderation open to the whims of participating members and hence are used, if at all, in addition to business-guided moderation.
Automated content moderation
As business and transactions grow, some elements of automation become necessary. Content moderation is no different. With the rise and rise in the volume of transactions and content on the internet, automated moderation is gradually becoming a necessity and less of a choice. That being said, automation can mean different things to different people. Also, unique automation tools and processes can be implemented for each company or platform using it. Some of the more common types of automations are:- Filters – Moderation through creating filters is simple and quick and does not require too much technology. Acceptable, or unacceptable words and phrases can be created and the tool will faithfully apply them as required. Human beings can continue to review the outcomes and update the filter lists.
- Blocking IP addresses – Users identified as abusive can be blocked from further interaction with the platform. Of course, users, if intent on creating malicious content on a platform, can continue to acquire new IP addresses and IDs to get through this block.
- Natural Language Processing (NLP) – The human brain processes information and is able to identify context, which a machine is unable to, or able to only to the extent it has been programmed to. With the help of NLP techniques, software applications are able to better identify context, patterns of conversation, relationships, etc. and take actions accordingly, giving an edge to automated content moderation.
- Artificial Intelligence (AI) – AI has been in the making for several years and is trotted out to be the moment when machines learnt to think and behave like humans. While it may make processing tasks and transactions easier and faster, in reality we are far away from such a time. However, AI tools are rapidly expanding the remit of the moderation that tools could handle. From handling textual content based on filters and NLP tools, automation is now able to review and make sense of images, audio and video content as well.