Israel-Hamas Conflict Sparks Meta Oversight Board’s First Emergency Case

Today, Meta’s Oversight Board announced that it would take on two expedited cases, the first ever, both dealing with the ongoing conflict between Israel and Hamas. The case will look at two posts initially removed and then reinstated from Instagram and Facebook for violating Meta’s policies against sharing graphic imagery and depicting dangerous organizations and individuals, respectively. One of the posts showed the aftermath of the attack on al-Shifa hospital by Israel Defense Forces, and the other was a video of an Israeli hostage being taken by Hamas on October 7.

“The current Israel-Hamas conflict marks a major event where Meta could have applied some of the Board’s more recent recommendations for crisis response and we are evaluating how the company is following through on its commitments,” Thomas Hughes, director of the Oversight Board Administration, told WIRED. “We see this as an opportunity to scrutinize how Meta handles urgent situations.”

Earlier this year, the Board announced it would take on “expedited cases” in what it called “urgent situations.”

The company has been critiqued for how it has handled content around the conflict. In October, 404 Media reported that the Meta’s AI was translating “Palestinian” into “Palestinian terrorist,” and many Palestinians believe that their content has been suppressed, “shadowbanned,” or removed altogether.

Meta, like many social media platforms, uses a combination of automated tools and a stable of human content moderators—many of them outsourced—to decide whether or not a piece of content violates the platform’s rules. The company also maintains a list of what it calls “dangerous organizations and individuals”—which includes organizations and names like the Islamic State, Hitler, the Ku Klux Klan, and Hamas. Sabhanaz Rashid Diya, a former member of Meta’s policy team and the founding director at tech policy think tank the Tech Global Institute, told WIRED that an automated system often might not be able to tell the difference between posts discussing or even condemning Hamas, as opposed to ones expressing support.

“There’s the question of whether even the very mention of Hamas is sufficient for it to lead to further adverse actions or not,” Diya says.

Following the 2021 conflict between Israel and Palestine, a human rights due diligence report requested by the Oversight Board and released in 2022 found that the company had both over- and under-enforced some of its own policies, meaning that, at times, content that should have been removed was left up, and content that didn’t violate the platform’s policies was removed anyway. In particular, researchers found, “Arabic content had greater over-enforcement,” meaning that it was more likely that content in Arabic would be taken down by Meta’s automated content moderation systems.

Marwa Fatafta, MENA policy and advocacy director at the nonprofit Access Now, a digital rights advocacy group, says that she has seen little change in Meta’s systems from 2021, and believes that the company’s content moderation policies still lack transparency for users. For instance, following Russia’s invasion of Ukraine, Meta made an exception to its policies around incitement to violence, allowing users in certain countries to post content, such as “death to the Russian invaders”, referring to the Russian military, that would normally be removed. “​​We still won’t allow credible calls for violence against Russian civilians,” Meta spokesperson Andy Stone told the Verge at the time.

“It’s not clear why some of these exceptions are made for some conflicts and not others,” says Fatafta. “We’re seeing videos and photos, sometimes just from bystanders or journalists, being removed and it’s not clear why. We’re really advocating for more context-specific content moderation.”

For some of these, like the post from al-Shifa, the company will assess whether the post is “newsworthy”, and reinstate images or videos when users appeal a takedown decision. This happens “pretty much in every single crisis,” according to Diya. In the case of the hostage video, the user posted it with a caption encouraging people to watch it to gain a “deeper understanding” of what happened on October 7, violating Meta’s long standing policy of not showing terrorist attacks, and its new policy of showing identifiable images of hostages. (Meta temporarily updated its policies to take down videos in which hostages were identifiable after October 7th).

In a company blog, Meta said the “Oversight Board’s guidance in these cases, along with feedback from other experts, will help us to continue to evolve our policies and response to the ongoing Israel-Hamas War.”

But the bigger issue, Diya says, is that the company continues treating each conflict like a one-off situation that requires a tailored response. “There’s a general reluctance within platforms to preempt or prepare for crises, especially if it’s outside the U.S., even when there is a prolonged history of conflict or violence in that region,” she says. “But we have seen enough crises in the last decade to get some sense of some patterns and what kind of tools should be in place.”

The expedited decisions from the Oversight Board, expected within 30 days, may finally push the company to do just that.

Facebook
Twitter
LinkedIn
Telegram
Tumblr