Illustration by Alex Castro / The Verge
Facebook and Instagram have instituted a temporary change in policy that allows users in some countries to post content thatâs usually forbidden, including calls for harm or even the death of Russian soldiers or politicians. The change first surfaced in a report by Reuters, citing internal emails to moderators. In them, the outlet reports mods are told that calls for the death of Russian President Vladimir Putin or Belarusian President Alexander Lukashenko will be allowed, as long as they donât contain threats toward others or âindicators of credibilityâ like saying where or how the act will take place.
In a statement sent to The Verge, Meta spokesperson Andy Stone said, âAs a result of the Russian invasion of Ukraine we have temporarily made allowances for forms of political expression that would normally violate our rules like violent speech such as âdeath to the Russian invaders.â We still wonât allow credible calls for violence against Russian civilians.â
The New York Times confirmed this policy applies to people using the service from Ukraine, Russia, Poland, Latvia, Lithuania, Estonia, Slovakia, Hungary, and Romania. The Times also notes that in 2021, Vice reported Facebook moderators received similar temporary instructions about âdeath to Khamaneiâ content and cited a spokesperson saying that Facebook had made that particular exception in certain earlier cases as well.
The Facebook community standards regarding hate speech and violence and incitement have continued to receive updates since the company started publishing them publicly in 2018. This change is just the latest example of how platforms have modified their treatment of content originating from the invading countries or relating to them since the fighting began.
An update to the Reuters report includes the content of the message sent to moderators, which reads as follows:
We are issuing a spirit-of-the-policy allowance to allow T1 violent speech that would otherwise be removed under the Hate Speech policy when: (a) targeting Russian soldiers, EXCEPT prisoners of war, or (b) targeting Russians where itâs clear that the context is the Russian invasion of Ukraine (e.g., content mentions the invasion, self-defense, etc.).
Typically, moderation guidelines would dictate that language dehumanizing or attacking a particular group based on its identity be removed. But the emails cited by Reuters claim the context of the current situation requires reading posts from the listed countries about generic Russian soldiers as a proxy for the Russian military as a whole, and absent credible statements attached, moderators are directed not to take action on them.
Still, itâs unclear whether the posts would be removed even without the direction. The policy already includes many carve-outs and exceptions. It states explicitly that additional information or context is needed before the enforcement of the policy in several cases, including:
Content attacking concepts, institutions, ideas, practices, or beliefs associated with protected characteristics, which are likely to contribute to imminent physical harm, intimidation or discrimination against the people associated with that protected characteristic. Facebook looks at a range of signs to determine whether there is a threat of harm in the content. These include but are not limited to: content that could incite imminent violence or intimidation; whether there is a period of heightened tension such as an election or ongoing conflict; and whether there is a recent history of violence against the targeted protected group. In some cases, we may also consider whether the speaker is a public figure or occupies a position of authority.
The Russian governmentâs reaction to the report is unknown, and there havenât been any updates from its censorship agency Roskomnadzor, which banned Facebook earlier this month.