BUSINESS

Facebook reveals process of removing controversial posts

Facebook CEO Mark Zuckerberg (AGENCIES PHOTO)

Facebook is trying to be more transparent about how it decides what content to take down or leave up.

On Tuesday, the company is making public for the first time its detailed internal community standards policies.

The document , previewed by PML Daily, is what Facebook’s (FB) 7,500 content moderators use when deciding what is and isn’t acceptable content, including hate speech, nudity, gun sales and bullying. A shorter version was previously available online.

Facebook is also adding a way for individuals to appeal when it removes one of their posts because of sexual content, hate speech or violence. Appeals will be reviewed by a moderator within a day, the company promises. Eventually, it will add appeals for more types of content and for people who reported posts that weren’t taken down.

Every week, Facebook sifts through millions of reports from users about inappropriate posts, groups or pages. Additional posts are also flagged by Facebook’s automated systems. A member of the team of moderators — a combination of full-time and contract employees around the world — reviews each post.

Why safer cities are popular with tourists

Discover how safer cities around the world are leveraging technology innovatively and flourishing as a result of their forward-thinking investment.

They include the company’s full definitions for terrorist organizations and hate groups. Hate speech is divided into three levels, and includes “some protections for immigration status.” There’s a detailed policy on the sale of marijuana (not allowed, even where it’s legal) and firearms (only shown to adults aged 21 or older — and no sales between individual people). Bullying rules don’t apply to comments made about public figures.

The document is filled with striking details about very specific issues. For example, you can’t post addresses or images of safe houses, or explicitly expose undercover law enforcement. You can only show victims of cannibalism if there’s a warning screen and age requirement. And photos of breasts are allowed if they depict an act of protest.

Facebook has come under criticism for not being transparent enough about how it decides what is or isn’t banned. And it has at times appeared inconsistent in the applications of its own rules.

Most recently, Facebook fought accusations that it censored conservative personalities like Diamond and Silk in the United States. Human rights groups have complained about its handling of hate-filled posts linked to violence in countries like Myanmar.

“Our enforcement isn’t perfect. We make mistakes because our processes involve people, and people are not infallible,” Monika Bickert, Facebook’s head of product policy, said in a blog post Tuesday.

The guidelines are global and will be released in 40 different languages. Facebook says it has detailed local information to help moderators handle the nuances of different locations and languages. It will not make all of its moderator guides public, such as lists of hate-speech words, as releasing them could make it easier for people to game the system.

To keep up with changes in language and behaviors, the guidelines are updated regularly. A policy team meets every two weeks to review potential additions or edits.

“We’ve promised to do better and we hope that sharing these details will serve as the basis for increased dialogue and input,” Bickert said.

Comments

iInHouse

All rights reserved. The information contained in this website may not be published, broadcast, re-written or re- distributed without the prior written authority of the publishers

Copyright © 2017. PML Daily

To Top