Facebook Makes Public its Full Content Guidelines

FILE - A picture illustration shows a Facebook logo reflected in a person's eye.

Your browser doesn’t support HTML5

Facebook Makes Public its Full Content Guidelines

Facebook published detailed guidelines Tuesday for what users can and cannot post on the social media service.

Earlier this month, lawmakers called on Facebook chief Mark Zuckerberg to appear in Congress. They demanded that his company take more steps to protect the privacy of its users. Some even raised the possibility of regulating the social media service.

Most of the newly-announced guidelines - which Facebook calls “Community Standards” – have already been in place. But it is the first time Facebook has publicly shared complete and detailed information about all its restrictions on user content.

About 7,600 Facebook moderators use the guidelines in examining Facebook content. The employees work in 40 languages. Facebook says its goal is to answer every report of possible content violation within 24 hours. Facebook removes material that violates the rules.

In some cases, Facebook employees also contact police or other officials to investigate dangerous situations or possible crimes.

The guidelines for posting content cover many areas. These include hate speech, violent threats or incitement of violence, self-harm and human trafficking.

Employees of the Competence Call Center (CCC) work for the Facebook Community Operations Team in Essen, Germany, Thursday, Nov. 23, 2017. (AP Photo/Martin Meissner)

Monika Bickert is Head of Global Policy Management at Facebook. She told reporters that the company felt it was important to be more open about its operations and totally clear about its user rules.

“You should, when you come to Facebook, understand where we draw these lines and what’s OK and what’s not OK,” Bickert said. She added that the guidelines themselves are always being reexamined and the latest changes are sent to employees weekly.

The guidelines demonstrate how complex and difficult decisions on restricting content can be. For example, Bickert said there are “community standards” that differ widely in nations across the world.

Another complex issue is how moderators decide how and when expressions of ideas and opinions violate policy.

For example, the Facebook guidelines permit a user to share another person’s hate speech. But only if it is done “for the purpose of raising awareness or educating others.” Users sharing such content must clearly show such purpose or moderators will remove the content.

Facebook says comments concerning violence must also be carefully examined in an effort to prevent real threats. The guidelines say that many times, people express their disagreement by threatening or calling for violence in “non-serious ways.”

The moderators have to consider “the language, context and details” of posts in order to separate non-serious statements from threatening statements. Facebook says moderators call law enforcement when they believe there is a real risk of physical harm.

In addition to human moderators, Facebook also uses artificial intelligence or AI in an effort to find content violations. Facebook says the machines can help identify some pieces of banned material. But it says they are not equipped to judge the history and context related to it. It is also difficult for AI to correctly identify humor as well as a human can.

Facebook also announced that people will now be permitted to appeal decisions about the removal of individual photos, videos or posts. Until now, appeals could be launched only over the removal of whole accounts, groups or pages. The service is also beginning to provide exact reasons for content removal action.

One of Facebook's new privacy settings will ask users whether they want to block certain features, such as facial recognition, targeted advertising or personal information. (Facebook)

The new guideline information was released one week after Facebook announced a major change to its privacy settings. The change makes it easier for people to approve or deny the company permission to use different kinds of user data.

“Everyone – no matter where they live – will be asked to review important information about how Facebook uses data and make choices about their privacy on Facebook,” the company said in a post on its website.

Users now will be asked directly whether they want Facebook to block facial recognition, some kinds of targeted advertising or personal information. The European Union forced the enactment of those changes. Facebook said its new privacy protections went into effect in Europe last week, but will be extended to users worldwide in the future.

I’m Bryan Lynn.

Bryan Lynn wrote this story for VOA Learning English, based on information from Facebook and reports from the Associated Press and Reuters. Caty Weaver was the editor.

We want to hear from you. Write to us in the Comments section, and visit our Facebook page.

______________________________________________________________

Quiz

_____________________________________________________________

Words in This Story

regulate v. to make rules or laws that control something

moderator n. person who monitors or manages public information

awareness n. feeling, experiencing or noticing of something

context n. the situation something happens within and that can help explain it

artificial intelligence n. the ability of a machine to reproduce human behavior

review v. look at or examine carefully