For the last few weeks, users have noticed that particular Chinese expressions have been automatically removed from YouTube comments. After rampant speculation about why Google would want these phrases — both of which are critical of the Chinese government — to be banned, it now claims it was an error by its machine learning software.
YouTube has deleted every comment I ever made about the Wumao (五毛), an internet propaganda division of the Chinese Communist Party. Who at Google decided to censor American comments on American videos hosted in America by an American platform that is already banned in China?
— Palmer Luckey (@PalmerLuckey) May 26, 2020
Users noticed that any YouTube comment containing the phrase “共匪” or “五毛” would purged within seconds of being submitted. One of my colleagues tested this and found it to be true. The former is an insult directed at the Chinese Communist government (it translates as “Communist bandit,” according to activist Jennifer Zeng). The latter is a slang term for online denizens who’ve been paid to deflect criticism of the Communist party.
As you might expect, users immediately suspected there might be an ulterior motive for Google banning the phrase. YouTube is banned in China, so why would its parent company care that anyone would criticize the CCP? This particular phrase had triggered this reaction for months, which is an awfully long time for an error to persist.
#YouTube “automatically” deletes a comment in Chinese, “Gongfei”, which means “communist bandit”, in 15 seconds.
This person tested 3 times, same result. #油管 15秒內自動刪除「共匪」留言，網友連試三次皆如此。
— Jennifer Zeng 曾錚 (@jenniferatntd) May 13, 2020
Perhaps Google was prompted to check into the matter when the phrase’s mysterious banning was pointed out by Oculus founder Palmer Luckey. Either way, it’s finally spoken about it. And it claims it’s not banning the phrase out of some hidden sympathy for the Chinese government, but rather out of an error.
According to a statement to TechCrunch, YouTube says the banned phrase was added to its hate speech filters, which automatically remove comments containing offensive content. That would explain why using the phrase even in a positive way, instantly brought the hammer down.
The question now is why that phrase was added to the filters. All Google would say is that the company is relying more heavily on AI-based moderation as its employees are out of the office thanks to the coronavirus pandemic. A YouTube blog post from March foreshadows the problem:
Machine learning helps detect potentially harmful content and then sends it to human reviewers for assessment. As a result of the new measures we’re taking, we will temporarily start relying more on technology to help with some of the work normally done by reviewers. This means automated systems will start removing some content without human review, so we can continue to act quickly to remove violative content and protect our ecosystem, while we have workplace protections in place.
It wouldn’t be the first time a major tech company ran into issues with machine learning moderation thanks to the coronavirus. Facebook had a similar problem when its AI blocked posts about making face masks.
YouTube claims it is still investigating the error. It invites anyone to “report suspected issues to troubleshoot errors and help us make product improvements.”
via The Verge