Â鶹Éçʼþ

Why Wikipedia Keeps Overlooking Women

Entries on women make up less than 30 percent of the website’s biographical coverage. Here’s what people are doing to fix it.

The ConversationMovements like #MeToo are drawing increased attention to the systemic discrimination facing women in a range of professional fields, from Hollywood and journalism to banking and government.

Discrimination is also a problem on user-driven sites like Wikipedia. Wikipedia is the . In January, the English-language version of the online encyclopedia had , more than 2000 percent higher than other online reference sites such as IMDb or Dictionary.com.

The volume of traffic on Wikipedia’s site—coupled with its integration into search results and digital assistants like Alexa and Siri—makes Wikipedia the predominant source of information on the web. YouTube even recently announced that it would start including on highly contested topics. But studies show that Wikipedia underrepresents content on women.

At the Rochester Institute of Technology, we’re taking steps to empower our students and our global community to address issues of gender bias on Wikipedia.

Signs of bias

Driven by , Wikipedia’s content can change in almost real time. That makes it a prime resource for current events, popular culture, sports and other evolving topics.

But relying on volunteers leads to systemic biases—both in content creation and improvement. A 2013 study estimated that women only accounted for . Wikipedia co-founder Jimmy Wales believes much since then, despite several organized efforts.

If women don’t actively edit Wikipedia at the same rate as men, topics of interest to women are at risk of receiving disproportionately low coverage. One study found that Wikipedia’s coverage of women was , but entries on women still constituted less than 30 percent of biographical coverage. Entries on women also and are more likely to include information on romantic relationships and family roles.

Gender bias is also an ongoing issue in content development and search algorithms.

What’s more, Wikipedia’s policies state that all content must be Because women throughout history have been less represented in published literature than men, finding reliable published sources on women can be challenging.

An obituary in a paper of record is often a criterion for inclusion as a biographical entry in Wikipedia. So it should be no surprise that women are underrepresented as subjects in this vast online encyclopedia. As The New York Times itself noted, its obituaries since 1851 —an oversight the paper now hopes to address through its “Overlooked” series.

Categorization can also be an issue. In 2013, a revealed that some editors had moved women’s entries from gender-neutral categories (e.g., “American novelists”) to gender-focused subcategories (e.g., “American women novelists”).

Wikipedia is not the only online resource that suffers from such biases. The user-contributed online mapping service . On GitHub, an online development platform, women’s contributions have a higher acceptance rate than men, but a study showed that the rate drops noticeably when the contributor .

Gender bias is also an ongoing issue in content development and search algorithms. Google Translate has been shown to and, for a time, LinkedIn in search results when users searched for a woman.

What can be done?

The solution to systemic biases that plague the web remains unclear. But libraries, museums, individual editors and the Wikimedia Foundation itself continue to make efforts to improve gender representation on sites such as Wikipedia.

Organized edit-a-thons can create a community around editing and developing underrepresented content. Edit-a-thons aim to increase the number of active female editors on Wikipedia, while empowering participants to edit entries on women during the event and into the future.

Earlier this month, our university library hosted its second annual Women on Wikipedia Edit-a-thon in celebration of Women’s History Month. The goal was to improve the content on at least 100 women in one afternoon.

Organized efforts help to give voice to women previously ignored by other resources.

For the past four years, students in our school’s American Women’s and Gender History course have worked to create new or substantially edit existing Wikipedia entries about women. One student created an entry on deaf-blind pioneer , while another added roughly 1,500 words to jazz artist entry.

This class was supported by , which encourages educators and students to contribute to Wikipedia in academic settings.

Through this assignment, students can immediately see how their efforts contribute to the larger conversation around women’s history topics. One student said that it was .

Other efforts to address gender bias on Wikipedia include ; organized editing communities such as and ; and .

Wikipedia’s dependence on volunteer editors has resulted in several systemic issues, but it also offers an opportunity for self-correction. Organized efforts help to give voice to women previously ignored by other resources.

This article was originally published by . It has been edited for YES! Magazine. 

Share
Yes! Republish




Reprints and reposts: YES! Magazine encourages you to make free use of this article by taking these easy steps.
Republish This
Related Stories