Wikipedia is in the process of launching its first global code of conduct, seeking to address the criticism it has received over allegedly failing to combat harassment and suffers from a lack of diversity.
“We need to be much more inclusive,” said María Sefidari, the chair of board of trustees for the non-profit Wikimedia Foundation.
“We are missing a lot of voices, we’re missing women, we’re missing marginalized groups.”
Online platforms have come under intense scrutiny for abusive behaviour, violent rhetoric and other forms of problematic content, pushing them to revamp content rules and more strictly enforce them.
Unlike Facebook and Twitter which take more top-down approaches to content moderation, the online encyclopedia, which turned 20 years old last month, largely relies on unpaid volunteers to handle issues around users’ behaviour.
Wikimedia said more than 1,500 volunteers from five continents and 30 languages participated in the creation of the new rules after the board of trustees voted in May last year to develop new binding standards for the Wikipedia platform.
“There’s been a process of change throughout the communities,” Katherine Maher, the executive director of the Wikimedia Foundation, said in an interview with journalists at our partner news agency Reuters.
“It took time to build the support that was necessary to do the consultations for people to understand why this is a priority.”
The new code of conduct bans harassment on and off the site, barring behaviours like hate speech, the use of slurs, stereotypes or attacks based on personal characteristics, as well as threats of physical violence and ‘hounding,’ or following someone across different articles to critique their work.
It also bans deliberately introducing false or biased information into content. Wikipedia is a relatively trusted site compared to major social media platforms which have struggled to curb misinformation.
Maher said some users’ concerns that the new rules meant the site was becoming more centralized were unfounded.
Wikipedia has 230,000 volunteer editors who work on crowdsourced articles and more than 3,500 ‘administrators’ who can take actions like blocking accounts or restricting edits on certain pages. Sometimes, complaints are decided on by panels of users elected by the communities.
Wikimedia said the next phase of the project would be working on the rules’ enforcement.
“A code of conduct without enforcement…is not going to be useful,” said Sefidari.
“We’re going to figure this out with the communities”.
Maher said there would be training for communities and interested task-forces of users.
Wikimedia has no immediate plans to beef up its small ‘trust and safety’ team, a group of about a dozen staff which currently acts on urgent matters such as death threats or the sharing of people’s private information, she said.
The team at Platform Executive hope you have enjoyed this news article. Translation from English to other languages via Google Cloud Translation. Initial reporting via our official content partners at Thomson Reuters. Reporting by Elizabeth Culliford. Editing by Kenneth Li and Edwina Gibbs.
Stay on top of the latest developments across the platform economy and gain access to our problem-solving tools, proprietary databases and content sets by becoming a member of our community. For a limited time, premium subscription plans start from just $7 per month.