User:Evad37/Wikipedia:Wikipedia Signpost/Single/2021-01-31

Source: Wikipedia, the free encyclopedia.
The Signpost
Single-page Edition
WP:POST/1
31 January 2021

Videos and podcasts
Celebrating 20 years
 

2021-01-31

1,000,000,000 edits, board elections, virtual Wikimania 2021

Edit number 1,000,000,000

Ser Amantio di Nicolao is already well known for being the "most productive editor" on the English language Wikipedia in terms of the number of edits he has made. This month, he added "the billionth edit" made by all editors on enWiki to his laurels. The count was complicated by a change in software and other factors – so let's say it was the billionth edit since the counter was reset in 2002.

The article he edited was

Death Breathing and it was later converted to a redirect. He told The Signpost "Getting the billionth edit wasn't actually my goal, though it's certainly a nice lagniappe
– all I had been trying to do was to push us closer to the goal itself, so that we could hopefully clear it before the twentieth anniversary of Wikipedia on January 15. It was quite a shock to discover that I'd hit the magic number. The article itself holds no particular interest for me – my musical interests lie more in the direction of classical. It was just one of a large handful of articles to which I was adding authority control templates. Nothing special."

He explained the personal significance of the achievement with a story. "Not long before my mother left the Soviet Union, her employer, Leningrad State University, got a copier for use in the office. It was a revolutionary tool; previously, office staff had relied on mimeographing and typing to perform many of the tasks the copier would take over. The powers-that-be knew that the new machine could be dangerous in the wrong hands, and restricted access to it. Only the departmental secretary could let people in to use it. And they had to be witnessed while they were making their copies. My mother never used this miracle machine, but when she heard about it her first thought was: 'This is it. This marks the beginning of the end of Soviet power.'"

"Just like the paper copies in Leningrad, each of the billion edits on Wikipedia, no matter how insignificant, is another step in ensuring that knowledge, and learning, and freedom, can be made available to all. That's the force of this moment in Wikipedia's history. The cumulative effect of those billion edits shows me just how far we have managed to come, collectively, in changing the world. 'As ordinary as it all appears,' to borrow a line from Jhumpa Lahiri, 'there are times when it is beyond my imagination.' "

Foundation removes community elections from bylaws, calls for feedback on future process for "community-and-affiliate-selected" trustees

The Board of Trustees has approved a bylaw change to increase the number of trustees from a maximum of 10 up to a maximum of 16 with 8 seats to be "community-and-affiliate-selected". The minimum number of seats is 9. As a step to ensure that the balance will not favor board-appointed seats, the Board will not make an appointment if that appointment will result in a minority of "community-and-affiliate-selected" seats.

In the announcement, the board's chair remarks that "We have not used the term 'community-sourced' that we initially proposed and instead now call these 'community-and-affiliate-selected' seats." According to a three-way diff prepared by a community member, the change otherwise largely implements the revision that the board had published for feedback in October (see our detailed earlier report). In particular, it still removes the requirement that a certain number of trustees have to be "selected from candidates approved through community voting," replacing it with the wording "sourced from candidates vetted through a Community and/or Affiliate nomination process" (also combining the board seats selected by editor community with those selected by affiliates, i.e. chapters and other recognized organizations within the Wikimedia movement). With this change, the board aims for "a process that enables diverse and equitable representation from communities across the Wikimedia movement." It is also likely to help the board avoid situations like the community re-electing a community-selected trustee despite the rest of the board having dismissed him not long before.

The removal of "voting" had been controversial, with founder trustee Jimmy Wales objecting that "I will personally only support a final [bylaws] revision which explicitly includes community voting". However, the December 9 resolution approving the change was unanimous. On the other hand, the announcement notes that "the status of Jimmy Wales as Community Founder Trustee remains unchanged" – after the October consultation, the board had taken up a proposal to remove his position (see earlier coverage: "WMF Board considering the removal of Jimmy Wales' trustee position amid controversy over future of community elections").

How are the community seats to be selected? We don't know yet! You can comment on an appropriate method at Call for feedback: Community Board seats from February 1 to March 14.

Wikimania 2021 going virtual

Last year's Wikimania scheduled for Bangkok was originally delayed to 2021 because of the pandemic. This week the in-person 2021 Wikimania was cancelled in favor of a virtual event. Janeen Uzzell announced on the Wikipedia-l mailing list that "it seems unlikely that we will all be able to travel freely and convene together by August 2021". Planning for the virtual Wikimania will begin soon and comments are welcomed at Talk:Wikimania 2021 on Meta.

Brief notes

RFAs since 2007



Reader comments

2021-01-31

Wiki reporting on the United States insurrection

U.S. Capitol at 2:22 p.m., 6 January 2021 according to the photo's metadata

Two weeks before Donald Trump's presidency ended, one of its defining moments appeared in near real time on the pages of Wikipedia. Our article on the

2021 storming of the United States Capitol
was created as rioters were overwhelming Capitol security, shots were reported on Wikipedia 30 minutes after they were fired, and the inevitable conclusion, the affirmation of Joe Biden's election as president of the United States, was recorded on our pages an hour after it was recorded in Congress.

At 1:10 p.m.

Eastern Standard Time
on January 6, Donald Trump ended his speech in which he repeated the false claims that he had won the presidential election, and instructed his supporters to march to the Capitol. "After this, we’re going to walk down and I’ll be there with you," he promised, although he never would join them. Among the protesters were people who took to heart Trump's statement that "If you don't fight ... you're not going to have a country anymore". At 1:26 p.m., Capitol police ordered the evacuation of two buildings in the Capitol complex as the gathering crowd continued to swell, and rioters began to overwhelm those guarding the plaza.

Meanwhile,

2021 storming of the United States Capitol. It began as a one-sentence stub: "On January 6, 2021, thousands of Donald Trump supporters gathered in Washington, D.C.
to reject results of the November 2020 presidential election."

By 2:30 p.m., when I became the second editor of the article, it was three sentences long, with three references, and three "see also" entries. The last two sentences were "At least 10 people have been arrested. Select buildings in the Capitol Hill complex were evacuated."

When I joined Another Believer editing the page, neither of us knew that about fifteen minutes earlier, rioters had broken windows in the Capitol building, climbed through, and unlocked doors to allow a surge of people to breach the Capitol. The first editor to report this did so at 2:33 p.m., and as rioters began to flood into the Capitol, edits began to flood in to the article as Wikipedians attempted to document the breaking news event in close to real time. Before too long, a new edit to the page was being saved every ten seconds. At 2:49 p.m., a Wikidata item was created to record structured data about the event, and by 3 p.m. editors on the Arabic and Basque Wikipedias had created pages of their own.

Shortly before 3 p.m., reports began to trickle in that shots had been fired inside the Capitol building. The first reports came from unreliable sources: Twitter posts, many of which also showed photographs of Capitol police gathered with guns drawn at the doors to the U.S. House of Representatives. By 3 p.m., ABC News confirmed, and it was added to the article less than fifteen minutes later. The first page move happened around the same time, when an editor observed that the page title, "January 2021 Donald Trump rally", did not disambiguate the event from other Donald Trump rallies that month. The page remained at its new title, "January 2021 United States Capitol protests", until a requested move discussion with more than 200 participants over 12 hours closed in the early hours of January 7 with consensus to move the page to "2021 storming of the United States Capitol".

The page continued to change at breakneck speed throughout the afternoon. By 4 p.m. a total of 257 edits had been made, 206 in the previous hour. Discussions unfurled on the talk page as hundreds of editors worked to craft the page. Should "parties to the civil conflict" be represented in the infobox, or was that too militaristic? Should various leaders' comments tweets about the event be included in a "reactions" section on the page, and if so should

flag icons
be used to denote their nationality? Should the people at the Capitol be called a mob, rioters, protesters, insurrectionists, or something else, and should that name differ based on whether they entered the building or stayed outside it?

The first image to be added to the article

At around 5:40 p.m., law enforcement announced they had cleared the Senate building. By this time, the page was 1,800 words long. People had begun looking for freely licensed images and videos with which to illustrate the page, and the first image appeared in the article at 6:36 p.m. By the end of the day, over 1,000 edits had been made to create a page that was more than 4,000 words long.

As of January 31, 1,084 editors have collaborated to create a page that has been viewed over 2.7 million times. The article is more than 12,000 words long and cites 475 unique references. The talk page is up to twelve pages of archived discussion, and a new move discussion continues as editors deliberate over whether the page ought to be retitled "Insurrection at the United States Capitol". Articles about auxiliary topics too long to fit in the main page have been created to describe the

aftermath of the event. The article exists on 55 language versions of Wikipedia, and there are entries about the topic at French and Russian Wikinews and English and Italian Wikiquote. Hundreds of images and videos have been uploaded to Category:2021 storming of the United States Capitol
and its subcategories on Wikimedia Commons.

Breaking news editing on Wikipedia often reminds me of the myth about bumblebees, which holds that bumblebees are able to fly despite being scientifically unable to do so. That thousands of editors can work together, communicating only through edit summaries and talk page messages, to accurately and comprehensively document breaking news as it unfolds seems to be something that could never work in theory. But in practice, it is nothing short of remarkable that we are able to sift through inaccurate and sometimes contradictory news reports to separate the facts from the inevitable inaccuracies and hyperbole of breaking news, make our best guesses at what will have lasting notability, and, eventually, revisit the articles once time has passed to ensure they are complete and balanced.



Reader comments

2021-01-31

From Anarchy to Wikiality, Glaring Bias to Good Cop: Press Coverage of Wikipedia's First Two Decades

Media coverage of Wikipedia has radically shifted over the past two decades: once cast as an intellectual frivolity, it is now lauded as the "last bastion of shared reality" online. To increase diversity and digital literacy, journalists and the Wikipedia community should work together to advance a new "wiki journalism."
Omer Benjakob is a journalist and researcher focused on Wikipedia and disinformation online. He is the tech and cyber reporter and editor for Haaretz and his work has appeared in Wired UK as well as academic publications.
Stephen Harrison is an attorney and writer whose writings have appeared in The New York Times, The Washington Post, Wired, and The Atlantic. He writes the Source Notes column for Slate magazine about Wikipedia and the information ecosystem.
This article has been updated from a version published on November 16, 2020 by Wikipedia @ 20 and is licensed CC BY 4.0.

Introduction

Wikipedia and its complex array of policies and intricate editorial processes are frequently misunderstood by the media. However, these very same processes – which seem convoluted if not impenetrable to outsiders not versed in the nuanced minutiae of Wikipedic culture – are many times the key to its encyclopedic success.

For example, Wikipedia's main article for Wikipedia, one of the most important (as well as most edited) articles on the project, explains that the ability to lock pages and prevent anonymous public editing on the encyclopedia that anyone can edit was the key to Wikipedia's success in weeding out disinformation on the coronavirus. Citing a respected media source, Wikipedia's Wikipedia article explains: "A 2021 article in the Columbia Journalism Review identified Wikipedia's page protection policies as '[p]erhaps the most important' means at Wikipedia's disposal to 'regulate its market of ideas'".[1] The article in the Columbia Journalism Review was written by us, in wake of research we conducted for the book Wikipedia @ 20 published by the MIT Press and edited by Joseph Reagle and Jackie Koerner this year. The mutually affirming relationship between Wikipedia and the media, as exemplified by our CJR story and the role it plays on the Wikipedia article, is only the latest in a long line spanning 20 years. Increasingly, journalists covering Wikipedia like us (Omer Benjakob for Haaretz and Stephen Harrison for Slate, among others) have found our articles about Wikipedia being quoted as sources on Wikipedia's articles. However, this circular relationship between sources and the sourcers, the cited and the citers, is a hallmark of Wikipedia and its community's somewhat volatile relationship with popular media over the years.

The following is an attempt to map out the different stages of those ties over the years as well as to glean from them real conclusions to help improve them. In the end we lay out what we think are small yet key steps that can be taken by journalists, Wikipedians and the Wikimedia Foundation to help improve the public's understanding of the project. While our research for MIT Press was intended for the academic community and the WMF, and while the CJR article was focused on journalists and how they can improve their coverage of Wikipedia, this text for The Signpost is aimed at the Wikipedian community.

Diverse and divergent as human society itself, one can never truly generalize about those involved in editing Wikipedia. However, to improve the public's ability not just to understand Wikipedia's process but also participate in it and provide critical oversight of the world's leading source of knowledge also requires openness to change, forgiveness for miscommunications and most importantly: good-faith. Good-faith not just towards your fellow Wikipedians, but also those not yet involved in it, or perhaps making their first foray into Wikipedia's processes, either as journalists – or as would-be first-time editors (who are sometimes responding to what they view as an encyclopedic injustice revealed by a journalist).

We as reporters trying to explain Wikipedia outwardly have many times been met with apprehension if not suspicion by members of the community, reluctant to air internal grievances publicly, and perhaps even risk facing accusations of off-wiki canvassing. Many times, we as journalists have been told to "fix" Wikipedia instead of write about it. This position is of course understandable. However, it misses both the wider role Wikipedia and journalism play in society. Wikipedia is a volunteer-community-run project, but it does not belong exclusively to the community.

Though everyone can participate, without proper media oversight there is no realistic way to expect people to find their place within the sprawling Wikipedic world of projects, task forces, associations and even faux-cabals. The following is an attempt to stress the mutually affirming ties Wikipedia and its community have had with the media, with the explicit attempt of helping to forge a new path for them, one in which journalists accurately portray the community, but also one in which the community works with journalists to help them navigate the halls of online encyclopedic power.

Four phases

"Jimmy Wales has been shot dead, according to Wikipedia, the online, up-to-the-minute encyclopedia." That was the opening line of a blatantly false 2005 news report by the online magazine The Register.[2] Rather than being an early example of what we may today call "fake news", the report by the tech site was a consciously snarky yet prescient criticism of Wikipedia and its reliability as a source for media. Wales was still alive, of course, despite what it had briefly stated on his Wikipedia entry, but by attributing his death to English Wikipedia, The Register sought to call out a perceived flaw in Wikipedia: On Wikipedia, truth was fluid and facts were exposed to anonymous vandals who could take advantage of its anyone-can-edit model to spread disinformation.

Over the past twenty years, English Wikipedia has frequently been the subject of media coverage, from in-depth exposés to colorful features and critical Op-eds. Is Wikipedia "impolite" as The New York Times claimed, or rather a "ray of light" as The Guardian suggested?[3] Both of us are journalists who have regularly covered Wikipedia in recent years, and before that we were frequent consumers of knowledge on the site (like many of our journalist colleagues). Press coverage of Wikipedia during the past 20 years has undergone a dramatic shift, and we believe it's important to highlight how the media's understanding of Wikipedia has shifted along with the public's understanding. Initially cast as the symbol of intellectual frivolity in the digital age, Wikipedia is now being lauded as the "last bastion of shared reality" in Trump's America.[4] Coverage, we claim, has evolved from bewilderment at the project, to concern and hostility at its model, to acceptance of its merits and disappointment at its shortcomings, and finally to calls to hold it socially accountable and reform it like any other institution.

We argue that press coverage of Wikipedia can be roughly divided into four periods. We have named each period after a major theme: "Authorial Anarchy" (2001–2004/5); "Wikiality" (2005–2008); "Bias" (2011–2017); and "Good Cop" (2018–present). We note upfront that these categories are not rigid and that themes and trends from one period can and often do carry over into others. But the overall progression reveals how the dynamic relationship between Wikipedia and the press has changed since its inception, and might provide further insight into how the press and Wikipedia will continue to interact with each other in the internet's knowledge ecosystem.

In short, we argue for what we term "wiki journalism" and the need for media to play a larger role in improving the general public's "Wikipedia literacy". With the help of the Wikimedia Foundation and the Wikipedia community, the press, we claim, can play a more substantial role in explaining Wikipedia to the public and serving as a civilian watchdog for the online encyclopedia. Encouraging critical readership of Wikipedia and helping to increase diversity among its editorship will ensure greater public oversight over the digital age's preeminent source of knowledge.

Authorial anarchy (2001–2004/5)

When Wikipedia was launched in 2001, mainstream media as well as more technology minded outlets treated it as something between a fluke and quirky outlier. With quotes from co-founders Jimmy Wales and Larry Sanger, early coverage tended to focus on what seemed like Wikipedia's most novel aspects: how it is written by anyone, edited collaboratively, is free to access and, in the case of tech media, extends the culture of open software development to the realm of encyclopedias.

"Anyone who visits the site is encouraged to participate," The New York Times wrote in its first piece on Wikipedia, titled "Fact-Driven? Collegial? This Site Wants You." Reports like these laid out the basic tenets of English Wikipedia, focusing on how collaborative technology and the volunteer community regulated what was termed "authorial anarchy."[5] Many of these reports included a colorful lede ("What does Nicole Kidman have in common with Kurt Godel?" Both have Wikipedia articles) showcasing the quirky diversity of content on the new site, where "[y]ou don't even have to give your real name" to contribute.[6]

Despite Wales' lofty claims that Wikipedia was creating a world in which everyone could have "free access to the sum of all human knowledge", throughout the early 2000s, mainstream media remained skeptical towards Wikipedia.[7] Reports from 2002–2003 mostly documented with some surprise its rapid growth in scale and scope, as well as its expansion into other languages. MIT Technology Review ran a report called "Free the Encyclopedias!", which described Wikipedia as "intellectual anarchy extruded into encyclopedia form" and "a free-wheeling Internet-based encyclopedia whose founders hope will revolutionize the stodgy world of encyclopedias"[6] – then still dominated by the Enlightenment-era Britannica and its more digital savvy competitor Encarta.

Repeated comparison to Encarta and Britannica is perhaps the most prominent characteristic of early media coverage, one that will disappear in later stages as Wikipedia cements its status as a legitimate encyclopedia. MIT Technology Review, for example, unironically claimed that Wikipedia "will probably never dethrone Britannica, whose 232-year reputation is based upon hiring world-renowned experts and exhaustively reviewing their articles with a staff of more than a hundred editors".[6] The demise of the status of experts would later become a hallmark of coverage of Wikipedia (discussed in the next section), but its seeds can be found from the onset: For example, in its first exposé on Wikipedia in 2004, The Washington Post reported that Britannica's vaunted staff was now down to a mere 20 editors. Only a year prior, Wikipedia editors noted that the prestigious paper "brushed off" Wikipedia almost entirely and instead focused on CD-ROM encyclopedias[8] – all the rage since Encarta launched a decade earlier, mounting what seemed at the time to be the bigger threat towards Britannica. Within a year, however, the newspaper's take on Wikipedia changed dramatically, and it was now concerned by the long term effect of Wikipedia's success, suggesting "the Internet's free dissemination of knowledge will eventually decrease the economic value of information".[8]

At the end of 2005, this tension between the English encyclopedia of the Enlightenment and that of the digital age would reach its zenith in a now infamous Nature news study that compared Wikipedia and Britannica. Published in December of 2005, Nature's "Internet Encyclopaedias Go Head to Head" found Wikipedia to be as accurate as its Enlightenment-era competitor, bringing experts to compare randomly selected articles on scientific topics.[9] News that Wikipedia successfully passed scientific scrutiny – that its ever-changing content was deemed to be as reliable as the static entries of a vaunted print-era encyclopedia like Britannica – made headlines around the world.[10] The Nature study was the final stage in a process that peaked in 2005 and cemented Wikipedia's shift from a web novelty whose value was to be treated skeptically at best to a cultural force to be reckoned with.

In March of 2005, Wikipedia had crossed the half a million article mark and some intellectuals began to discuss the "the wikification of knowledge".[11]

Tellingly, 2005 was also the year that the Wikipedia community first began recording its coverage in the media in an organized fashion. Initially focused on instances of "Wiki love" from the press, in 2005 the community created categories and project pages like "America's Top Newspapers Use Wikipedia" for its early press clippings.[12] The Signpost, the online newspaper for the English language Wikipedia, was also founded in 2005 to report on events related to Wikipedia.[13] Over time the community grew increasingly conscious of its public role and by 2006 an organized index of all media references to Wikipedia was set up – first with a list for every year and then, as coverage swelled, one for every month as well.[14] Categories were also created for times when Wikipedia was cited as a source of information by mainstream media[15] – a rare reversal of roles that highlighted the mutually affirming relationship between Wikipedia and the media that would develop over later periods.

Indeed, 2005 was to be a key year for Wikipedia: it saw its biggest vindication – the Nature report – alongside its biggest vilification – the so-called Seigenthaler affair. By 2005, Wikipedia was no longer quirky. Now it was to be viewed within a new framework which contrasted its popularity with its accuracy and debated the risks it posed. The New York Times, for example, claimed that the Seigenthaler "case triggered extensive debate on the Internet over the value and reliability of Wikipedia, and more broadly, over the nature of online information".[16] In the next phase, Wikipedia's effect on the popular understanding of truth that would be the overriding theme.

Wikiality (2005–2008)

Stephen Colbert launched his satirical news program The Colbert Report with a segment dedicated to what would be dubbed 2005's word of the year: truthiness.[17] "We're not talking about truth, we're talking about something that seems like truth – the truth we want to exist", Colbert explained.[18] He even urged viewers to take the truth into their own hands and "save" the declining populations of elephants in Africa by changing their numbers on Wikipedia, causing its server to crash. The wider point resonated.[19] "It's on Wikipedia, so it must be true", The Washington Post wrote that year.[20] Wikipedia was no longer taken to be just another website, it was now a powerhouse undermining intellectual institutions and capable of changing our very perception of reality.

Colbert followed up his infamous segment with another potent neologism: wikiality. "Wikiality", he charged, was the reality created by Wikipedia's model, in which "truth" was based on the will of the majority and not on facts. This was a theme that had a deep political resonance in post-9/11 America, buoyed by the presidency of George W. Bush and the rise to prominence of Fox News – and Wikipedia was increasingly cast as providing its underlying intellectual conditions. "Who is Britannica to tell me that George Washington had slaves? If I want to say he didn't, that's my right," Colbert charged. "Thanks to Wikipedia, it's also a fact. [We're] bringing democracy to knowledge."[21]

During 2006–2009, the dominance of Wikipedia's encyclopedic model was solidified. In 2008, The New York Times published a "eulogy" for print encyclopedia and flagged the need to understand the "epistemology of Wikipedia" and the "wikitruth" it bred.[22] Wikipedia's underlying philosophy – its model's effects on the very nature of facticity – was now deserving of more serious and critical examination. MIT Technology Review ran a piece on "Wikipedia and the Meaning of Truth", asking "why the online encyclopedia's epistemology should worry those who care about traditional notions of accuracy".[23]

Concerns that Wikipedia's epistemic model was replacing expertise loomed large: In 2006, The New York Times debated the merits of "the nitpicking of the masses vs. the authority of the experts", and The Independent asked: "Do we need a more reliable online encyclopedia than Wikipedia?"[24] In a report that profiled Wikipedians, The New Yorker wondered: "Can Wikipedia conquer expertise?"; and Larry Sanger, who had left the project by then, lamented "the fate of expertise after Wikipedia".[25] Though largely negative, these in-depth reports also permitted a more detailed treatment of Wikipedia's theory of knowledge. Articles like Marshal Poe's "The Hive", published in The Atlantic's September 2006 edition, laid out for intellectual readers Wikipedia's history and philosophy like never before.

Epistemological and social fears of Wikipedia were also fueled by Wikipedia's biggest public media storm to date – the so-called Essjay scandal of 2007.[26] It even spurred calls to reform Wikipedia.[27] The fact that Ryan Jordan held an official status within Wikipedia's community seemed to echo an increasingly accepted political truism: facts were being manipulated by those with power.

Knowledge was increasingly being politicized and the entirety of

banned from editing Wikipedia anonymously during 2006 after politicians' articles were whitewashed in what The Washington Post called "wikipolitics".[28] During this period Wikipedia also first faced allegations of having a liberal bias – for example by "evangelical Christians" who opened a conservative wiki of their own.[29]
Reports like these helped grant social currency to the claim that knowledge was political like never before.

The politicization of knowledge, alongside a proliferation of alternative wikis – exacerbated in part by Wales' for-profit website Wikia, launched in 2006 – all served to highlight the wikiality of America's political and media landscape.[30] It was at this time that the first cases of "citogenesis" – circular and false reporting originating from Wikipedia – appeared. These showed how dependent classic media was on Wikipedia – and therefore how politically vulnerable and unreliable it was by proxy. These included reports that cited the unfounded claim regarding Hillary Clinton being the valedictorian of her class at Wellesley College, an error born from false information introduced to her Wikipedia article.[31] The edit wars on Bush's Wikipedia page highlighted the online encyclopedia's role in what The New York Times termed the "separate realities" within America.[32]

By 2007, Wikipedia was among the top ten most popular websites in the world. And though it was a non-profit, it maintained partnerships with corporate juggernauts like Google, whose donations and usage of Wikipedia helped it walk among giants, giving it a privileged position on the search engine results and sparking concerns of a "googlepedia" by internet thinkers.[33]

Wikipedia was now a primary source of knowledge for the information age, and its internal workings mattered to the general public.[34] Coverage shifted in accordance. Reports began to focus on the internal intellectual battles raging within the community of editors: For example, The Guardian wrote about the ideological battle between "deletionists" and "inclusionists".[35] For the first time, coverage of Wikipedia was no longer monolithic and the community was permitted diverging opinions by the press. Wikipedia was less a unified publisher and more a vital discursive arena. Policy changes were debated in the media, and concerns over Wikipedia's "declining user base" were also covered – mostly by Noam Cohen, who covered the encyclopedia for the New York Times.[36] Wikipedia was now a beat. Its worldview was now fully embedded within our social and political reality. The question was what was it telling us, who was writing it, and who was being excluded.

Bias (2011–2017)

In February 2011, The New York Times ran a series of articles on the question "Where Are the Women of Wikipedia?" in its opinion pages. These 2011 articles have very different headlines than the paper's coverage of Wikipedia in the prior decade. Reporting between roughly the years 2006 to 2009 focused on Wikipedia's reliability, with headlines like "Growing Wikipedia Refines its ‘Anyone Can Edit' Possibility" (2006) and "Without a Source, Wikipedia Can't Handle the Truth" (2008).[37]

By 2011, however, the press coverage had zeroed in on the site's gender imbalance. Headlines were much more openly critical of the community itself than in the past, with The New York Times' series calling out "trolls and other nuisances" and Wikipedia's "antisocial factor," as well as "nerd avoidance".[38] Press coverage had shifted from the epistemological merits of Wikipedia to legitimate concerns about bias in its contributor base.

The 2011 series about gender on Wikipedia followed a 2010 survey conducted by United Nations University and UNU-MERIT that indicated only 12.64 percent of Wikipedia contributors were female among the respondents.[39] Although the results of that study were later challenged,[40] much like with Nature's study, the fact that its results were contested by Britannca made little difference. The fact that the UN study received an entire series of articles indicates how the results struck a cultural nerve. What did it say about Wikipedia–and internet knowledge generally – that a disproportionate number of the contributors were men?

One could argue that this shift – from grappling with underpinnings of Wikipedia's model of knowledge production to a critique of the actual forces and output of the wiki way of doing things – symbolized an implicit acceptance of Wikipedia's status in the digital age as the preeminent source of knowledge. Media coverage during this period no longer treated Wikipedia as an outlier, a fluke, or as an epistemological disaster to be entirely rejected. Rather, the press focused on negotiating with Wikipedia as an existing phenomena, addressing concerns shared by some in the community – especially women, predating the GamerGate debate of 2014.

Press coverage of Wikipedia throughout the period of 2011 to roughly 2017 largely focused on the online encyclopedia's structural bias. This coverage also differed markedly from previous years in its detailed treatment of Wikipedia's internal editorial and community dynamics. The press coverage highlighted not only the gender gap in percentage of female contributors, but in the content of biographical articles, and the efforts by some activists to change the status quo. Publications ranging from The Austin Chronicle to The New Yorker covered feminist edit-a-thons, annual events to increase and improve Wikipedia's content for female, queer, and women's subject, linking contemporary identity politics with the online project's goal of organizing access to the sum of human knowledge.[41] In addition to gender, the press covered other types of bias such geographical blind spots and the site's exclusion of oral history and other knowledge that did not meet the Western notions of verifiable sources.[42]

During this period, prestigious publications also began profiling individual Wikipedia contributors, giving faces and names to the forces behind our knowledge. "Wikipedians" were increasingly cast as activists and recognized outside the community. The Washington Post, for example, covered Dr. Adrianne Wadewitz's death in 2014, noting that Wadewitz was a "Wikipedian" who had "empower[ed] everyday Internet users to be critical of how information is produced on the Internet and move beyond being critical to making it better".[43] The transition from covering Wikipedia's accuracy to covering Wikipedians themselves perhaps reflects an increased concern with awareness about the human motivations of the people contributing knowledge online. Many times this took on a humorous tone, like the case of the "ultimate WikiGnome" Bryan Henderson whose main contribution to Wikipedia was deleting the term "comprised of" from over 40,000 articles.[44] Journalists (including the authors of this paper) have continued this trend of profiling Wikipedians themselves.

A 2014 YouGov study found that around two thirds of British people trust the authors of Wikipedia pages to tell the truth, a significantly higher percentage than those who trusted journalists.[45] At the same time, journalists were increasingly open to recognizing how crucial Wikipedia had become to their profession: With the most dramatic decline in newsroom staffs since the Great Recession, Wikipedia was now used by journalists for conducting initial research[46] –another example of the mutually affirming relationship between the media and Wikipedia.

As more journalists used and wrote about Wikipedia, the tone of their writing changed. In one of his reports for The New York Times, Noam Cohen quoted a French reporter as saying, "Making fun of Wikipedia is so 2007".[47] When Cohen first began covering Wikipedia, most people saw Wikipedia as a hobby for nerds–but that characterization had by now become passé. The more pressing concern, according to Cohen, was "[s]eeing Wikipedia as The Man".[47]

Overall, press coverage of Wikipedia during this period oscillates between fear about the site's long-term existential prospects[48] and concern that the site is continuing the masculinist and Eurocentric biases of historical encyclopedias. The latter is significant, as it shows how Wikipedia's pretenses of upending the classic print-model of encyclopedias have been accepted by the wider public, which, in turn, is now concerned or even disappointed that despite its promise of liberating the world's knowledge from the shackles of centralization and expertise, it has in fact recreated most of the biases of yesteryear.

Good Cop (2018–present)

In April 2018, Cohen wrote an article for The Washington Post titled "Conspiracy Videos? Fake News? Enter Wikipedia, the 'Good Cop' of the Internet."[49] For more than a decade, Cohen had written about Wikipedia in the popular press, but his "Good Cop" piece was perhaps his most complimentary and it signaled a wider change in perception regarding Wikipedia. He declared that "fundamentally … the project gets the big questions right." This would become the main theme of coverage and by the time Wikipedia marked its 20th anniversary, most main stream media sources were gushing with what the community in 2005 called "wiki love" for the project.

Cohen's "Good Cop" marks the latest shift in coverage of Wikipedia, one that embarks from the issue of truthiness and reexamines its merits in the wake of the "post-truth" politics and "fake news" – 2016 and 2017's respective words of the year.

The Wall Street Journal credited Wikipedia's top arbitration body, ArbCom, with "keep[ing] the peace at [the] internet encyclopedia."[50] Other favorable headlines from 2018 and 2019 included: "There's a Lot Wikipedia Can Teach Us About Fighting Disinformation" and "In a Hysterical World, Wikipedia is a Ray of Light – and That's the Truth."[51] Wikipedia was described by The Atlantic as "the last bastion of shared reality" online, and for its 18th birthday, it was lauded by The Washington Post as "the Internet's good grown up."[52]

But this was only the beginning. What caused press coverage of Wikipedia to pivot from criticizing the encyclopedia as the Man to casting Wikipedia as the web's good cop? Two main external factors seem to have played a key role: U.S. President Donald Trump and the coronavirus pandemic. Since Trump's election in 2016, the mainstream press expressed concerns about whether traditional notions of truth and reality-based argument. This resulted in large part from what seemed to be an official assault by the administration in the White House – both against the media itself and facts, and its offering alternative outlets for communication and "alternative facts" in their stead. The "truthiness" culture of intellectual promiscuity represented by the Presidency of George W. Bush had deteriorated into the so-called "post-truth" culture of the Trump White House. Wikipedia's procedural answers for the question of what is a fact, initially hailed as flawed due to their inherent beholdance to existing sources, could now be taken in a different light.[53]

Wikipedia had also gotten better, but its model had remained the same. What had changed was the internet and our understanding of it. Wikipedia's emphasis on neutral point of view, and the community's goal to maintain an objective description of reality, represents an increasingly striking contrast to politicians around the world whose rhetoric is not reality-based.[54] Moreover, the Wikipedia community's commitment to sourcing claims (exemplified by Wikipedia's community ban on the Daily Mail in 2017 and of Breitbart in 2018) highlighted how Wikipedia's model was seemingly more successful than the traditional media in fighting its own flaws and the rise of "fake news."[55]

2018 also saw Wikipedia lock horns with some of those considered supportive of Trump and the "post-truth" discourse, including Breitbart and even Russian media. The so-called "Philip Cross affair"[56] saw a British editor face accusation that he was in fact a front for the U.K.'s Ministry of Defense or even the American C.I.A., claims that were parroted out by Sputnik and RT. Breitbart all but declaring war on the online encyclopedia (running no less than 10 negative reports about it in as many months, including headlines like "Wikipedia Editors Paid to Protect Political, Tech, and Media Figures" and "Wikipedia Editors Post Fake News on Summary of Mueller Probe").[57] 2018 also saw the clearest example of Russian intervention in Wikipedia, with Russian agent Maria Butina being outed by the community for trying to scrub her own Wikipedia page.[58]

The shift toward more positive press treatment of Wikipedia also overlaps with a general trend toward negative coverage of for-profit technology sites. In recent years, Facebook, Google, Twitter, and YouTube have been chastised in the press for privacy violations, election hacking, and platforming hateful content. But Wikipedia has largely dodged these criticisms. Complimentary journalists have noted the site's rare position as a nonprofit in the most visited websites in the world – the only site in the global top ten that is not monetized with advertising or by collecting and selling personal information of users. Journalists have also praised Wikipedia's operating model. As Brian Feldman pointed out in a New York Magazine piece titled, "Why Wikipedia Works", the site's norms of review and monitoring by a community of editors, and deleting false information and inflammatory material, seems vastly superior to the way social media platforms like Twitter fail to moderate similarly problematic content.[59]

With the 2020 election, this process reached its zenith. While Twitter and Facebook scrambled to prevent the New York Post report about Hunter Biden from spreading through its platform, sparking yet again a debate on the limits of free speech online, the dynamic on Wikipedia was very different: Instead of censoring the link or trying to prevent its content from being disseminated, Wikipedia's editors contextualized its publication as part of a wider "conspiracy theory" relating to Biden being pushed out by Trump's proxies. By election day, Reuters, Wired, Vox and others were all praising Wikipedia for being more prepared for election disinformation than social media.

It's important to note that even during this period of relatively favorable press coverage of Wikipedia, newspapers still published highly critical articles. But the focus has been on reforming Wikipedia's governance policies rather than rejecting its underlying model of crowdsourced knowledge.[60] For example, Wikipedia received significant media attention in 2018 when Donna Strickland won a Nobel Prize in physics and, at the time of her award, did not have a Wikipedia page; an earlier entry had been deleted by an editor who found that Strickland lacked sufficient notability, despite the fact her two male co-laureates had pages for the same academic research that earned the three the prestigious award. But note how press coverage of Strickland did not dispute Wikipedia's underlying premise of community-led knowledge production. Rather press coverage was continuing the structural critique from the previous phase.

Furthermore, by this era the Wikimedia Foundation had increasingly begun speaking publicly about matters of concern to the Wikipedia community. When it came to the Strickland incident, the Wikimedia Foundation was not overly apologetic in its public statements, with Executive Director Katherine Maher writing an op-ed for the Los Angeles Times titled "Wikipedia Mirrors the World's Gender Biases, it Doesn't Cause Them".[61] Maher challenged journalists to write more stories about notable women so that volunteer Wikipedians have sufficient material to source in their attempt to fix the bias. Maher's comments, in other words, advocate further awareness of the symbiotic relationship between the media and Wikipedia.

The Strickland incident is in some ways an outlier during a time of relatively favorable press coverage of Wikipedia. How long will this honeymoon period last? One indication that the pendulum will swing back in a more-critical direction is the coverage of large technology companies using Wikipedia. The press widely covered YouTube's 2018 announcement that it was relying on Wikipedia to counteract videos promoting conspiracy theories when there had been no prior notice to the Wikimedia Foundation regarding YouTube's plans. Journalists also wrote – at times critically – about Facebook's plan to give background information from Wikipedia about publications to combat "fake news", Google's use of Wikipedia content for its knowledge panels, and how smart assistants like Siri and Alexa pull information from the site.

Prominent tech critics have questioned whether it is truly appropriate to leverage Wikipedia as the "good cop" since the site is maintained by unpaid volunteers, and tech companies are using it for commercial purposes. But from a news perspective, it might not matter so much whether it's fair or prudent for technology companies to leverage Wikipedia in this way–the appearance of partnership is enough to spur a news story. The more it seems as if Wikipedia has become aligned with Big Tech, the more likely the encyclopedia will receive similarly adverse coverage.

Nonetheless, coronavirus proved a pivotal moment, giving Wikipedia and its model what seems to be its biggest vindication since Nature 2005. As a key node in the online information ecosphere, and with concerns regarding Covid-19 disinformation causing the WHO to label the pandemic an "infodemic" in February 2020, vandalism on coronavirus articles was not something WP and perhaps the world could afford. Luckily, the WikiProject Medicine was prepared. "Our editing community often concentrates on breaking news events such that content rapidly develops. The recent outbreak of novel coronavirus has been no exception," Doc James told Wired magazine, in what would be the first of many stories praising Wikipedia's response to the virus due to its especially rigid sourcing policy.

The project and Doc James would turn into the face of Wikipedia's response to the pandemic – so much so that the community would even open a special article on "Wikipedia's response to the Covid-19" pandemic – a rare sign of WP's community efforts meeting their own notability guidelines. Wikipedia, as the article about its response to the pandemic both said and showed, was notably reliable due to its emphasis on neutral point of view and verification of sources. Wikipedia's Covid-19 task force had relied on the top tier of legacy media – building a list of recommended sources from popular and scientific media – and they vindicated it for their commitment to trusted institutions of authority. By the start of 2021, as the pandemic marked its one year anniversary and Wikipedia its 20th, publications across the world were praising it.

Conclusion

Over the span of nearly two decades, Wikipedia went from being heralded as the original fake news, a symbol of all that was wrong with the internet, to being the "grown up" of the web and the best medicine against the scourge of disinformation. This process was predicated on Wikipedia's epistemic model gaining social acceptance as well as the erosion of status of mainstream media and traditional knowledge sources. Comparisons to older encyclopedias have all but disappeared. More common are appeals like Maher's request following the Strickland affair that journalists aid Wikipedia in the attempt to reform by publishing more articles about women. This dynamic highlights how Wikipedia is now a fixture within our media landscape, increasingly both the source of coverage and the story itself.

Understanding the mutually affirming and dynamic between media and Wikipedia opens up a rare opportunity to engage directly with some of the issues underscoring information as well as disinformation – from critical reading of different sources, to basic epistemological debates, issues that were once considered too academic for mainstream media are now finding their place in the public discourse through coverage of Wikipedia. For example, reports about Strickland's lack of a Wikipedia article helped make accessible the feminist theory regarding knowledge being "gendered". The idea that history is his-story was highlighted in debates about Wikipedia's gender bias, with the dire lack of articles about women scientists being easily explained by the lack of historical sources regarding women. Meanwhile, reports about Wikipedia being blocked in countries such as China and Turkey have allowed for a discussion of the politics of knowledge online as well as a debate regarding the differences between Wikipedias in different languages and local biases. Detailed and critical reports like these are part of a new sub-genre of journalism that has emerged in the past years, what we term "wiki journalism": Coverage of Wikipedia as a social and political arena in its own right.[62]

Nonetheless, much more can be done – by journalists, the Wikimedia Foundation and even the Wikipedia community of volunteers. Though Wikipedia's technology purportedly offers fully transparency, public understanding of Wikipedia's processes, bureaucracy, and internal jargon is still a massive obstacle for would-be editors and journalists alike. Despite its open format, the majority of Wikipedia is edited by a fraction of its overall editors, indicating the rise of an encyclopedic elite not too dissimilar in characteristics than that of media and academia. To increase diversity in Wikipedia and serve the public interest requires journalists to go beyond "gotcha" headlines. Much of the popular coverage of Wikipedia is still lacking and is either reductive or superficial, treating Wikipedia as a unified voice and amplifying minor errors and vandalism. Many times, reports like these needlessly politicize Wikipedia. For example, after a vandal wrote that the Republican Party of California believed in "Nazism" and the error was aggregated by Alexa and Google, reports attributed blame to Wikipedia.[63]

These issues are discussed at length in our text in the Columbia Journalism Review, which urges journalists to help increase Wikipedia literacy, dedicating more coverage to the project's inner workings and policies.[1] But the media is not alone and the WMF as well as Wikipedians can also help. In recent years, the Wikimedia Foundation has taken the helpful step of hiring communications specialists and other employees to help members of the press connect with sources both at the Foundation and the larger Wikimedia movement. Yet although the Wikimedia Foundation has made press contacts much more accessible, there is still work to be done to enhance communication between Wikipedia and the media. Creating a special status for wiki journalists, for example, recognizing their users and granting them read-only status for deleted articles and censored edits – a right currently reserved for official administrators – could help reporters better understand the full context of edit wars and other content disputes.

The community must too be more open to working with media and take a much less aggressive approach to external coverage of their debates. Many times, editors are reluctant to speak to reporters and are antagonistic towards unversed users who have come to mend an error or bias they have read about in the media. In some cases, editors who speak with the press are stigmatized. Much like the interviews with editors of WikiProject Medicine's Covid-19 task force showed, working with the media can actually help highlight the important work taking place on the encyclopedia. Wikipedia editors must accept that they and their community play a much wider social role than they may perceive – a social role that places the output of their volunteer activity center stage online and also makes them part of the public debate. To help bridge the gap between public discourse and the Wikipedic one, editors need to go beyond the "just fix it yourself" mentality and help increase public oversight of Wikipedia.

As Wikipedia is fully transparent, the demand for further public oversight may seem misplaced or even anachronistic. However in much the same way we need the media to help oversee public committee hearings in town halls or national legislatures plenums, so too does the public need the media's help in participating in Wikipedia's transparency. Much like we need a strong active media to help encourage and facilitate civilian oversight of political processes, so too do we need robust media coverage to help encourage civic involvement in encyclopedic processes.

Wikipedians may not perceive themselves to be gatekeepers in the same way lawmakers or congressional aides are, but for those viewing Wikipedia from the outside, many times they do actually appear to play just such a role. Lack of public engagement in Wikipedia cannot be blamed solely on the public or its purported apathy. Wikipedians must not just allow the media to highlight problems within their community, but proactively flag issues, helping reporters sift through countless debates and find the truly important stories, instead of limiting themselves to internal forums and demanding journalists and the public fix Wikipedia themselves.

Together, journalists, the Wikimedia Foundation and the community, can help increase critical digital literacy through deeply reported coverage of Wikipedia. High-quality wiki journalism would not treat Wikipedia as a monolithic agent that speaks in one voice, but rather would seek to understand the roots of its biases and shortcomings. This will serve to highlight the politics of knowledge production instead of politicizing knowledge itself.

References

  1. ^ a b Stephen Harrison and Omer Benjakob, "Wikipedia is twenty. It's time to start covering it better", Columbia Journalism Review, January 14, 2021. https://www.cjr.org/opinion/wikipedia-is-twenty-its-time-to-start-covering-it-better.php
  2. ^ Andrew Orlowski, "Wikipedia Founder 'Shot by Friend of Siegenthaler,'" Register, December 17, 2005, https://www.theregister.co.uk/2005/12/17/jimmy_wales_shot_dead_says_wikipedia/
  3. ^ Marshall Poe, "The Hive," Atlantic 298, no. 2 (2006): 86, https://www.theatlantic.com/magazine/archive/2006/09/the-hive/305118/; Noam Cohen, "Conspiracy Videos? Fake News? Enter Wikipedia, the 'Good Cop' of the Internet," Washington Post, April 6, 2018, https://www.washingtonpost.com/outlook/conspiracy-videos-fake-news-enter-wikipedia-the-good-cop-of-the-internet/2018/04/06/ad1f018a-3835-11e8-8fd2-49fe3c675a89_story.html; Noam Cohen, "Defending Wikipedia's Impolite Side," The New York Times, August 20, 2007, https://www.nytimes.com/2007/08/20/technology/20link.html; John Naughton, "In a Hysterical World, Wikipedia Is a Ray of Light – and That's the Truth," The Guardian, September 2, 2018, https://www.theguardian.com/commentisfree/2018/sep/02/in-hysterical-world-wikipedia-ray-of-light-truth.
  4. ^ Alexis C. Madrigal, "Wikipedia, the Last Bastion of Shared Reality," Atlantic, August 7, 2018, https://www.theatlantic.com/technology/archive/2018/08/jeongpedia/566897/.
  5. ^ Peter Meyers, "Fact-Driven? Collegial? This Site Wants You", The New York Times, September 20, 2001, https://www.nytimes.com/2001/09/20/technology/fact-driven-collegial-this-site-wants-you.html
  6. ^ a b c Judy Heim, "Free the Encyclopedias!" MIT Technology Review, September 4, 2001, https://www.technologyreview.com/s/401190/free-the-encyclopedias/
  7. ^ Rob Miller, "Wikipedia Founder Jimmy Wales Responds," Slashdot, July 28, 2004, https://slashdot.org/story/04/07/28/1351230/wikipedia-founder-jimmy-wales-responds. Citing Wikipedia, The New York Times Magazine chose "Populist editing" as one of 2001's big ideas: Steven Johnson, "The Year in Ideas: A to Z; Populist Editing," New York Times Magazine, December 9, 2001, https://www.nytimes.com/2001/12/09/magazine/the-year-in-ideas-a-to-z-populist-editing.html.
  8. ^ a b Leslie Walker, "Spreading Knowledge: The Wiki Way," The Washington Post, September 9, 2004, http://www.washingtonpost.com/wp-dyn/articles/A5430-2004Sep8.html; Wikipedia contributors, "Wikipedia: Press coverage 2003," Wikipedia, accessed August 29, 2019, https://en.wikipedia.org/w/index.php?title=Wikipedia:Press_coverage_2003&oldid=760053436.
  9. ^ Jim Giles, "Internet Encyclopaedias Go Head to Head," Nature, December 14, 2005, https://www.nature.com/articles/438900a
  10. ^ Dan Goodin, "Science Journal: Wikipedia Pretty Accurate," Associated Press, December 14, 2005. The AP report was reprinted in over 112 media outlets, including Al Jazeera and The China Post. The BBC also reported about Nature's findings, writing: "Wikipedia Survives Research Test" (BBC News Online, December 15, 2005), as did Australia's The Age (Stephen Cauchi, December 15, 2005), to name a few.
  11. ^ John C. Dvorak, "The Wikification of Knowledge," PC Magazine, November 7, 2005, https://www.pcmag.com/article2/0,2817,1835857,00.asp; see also Jaron Lanier, "Digital Maoism: The Hazards of the New Online Collectivism," Edge, May 29, 2006, https://www.edge.org/conversation/jaron_lanier-digital-maoism-the-hazards-of-the-new-online-collectivism.
  12. ^ Wikipedia, s.v. "Wikipedia: Wikilove from the Press," last modified January 14, 2017, Special:Diff/760052982; Wikipedia, s.v. "Wikipedia: America's Top Newspapers Use Wikipedia," accessed August 29, 2019, Special:Diff/760053107.
  13. ^ Wikipedia, s.v. "The Signpost," Wikipedia, accessed August 29, 2019, https://en.wikipedia.org/w/index.php?title=The_Signpost&oldid=913126984.
  14. ^ Wikipedia, s.v. "Wikipedia: Press coverage," accessed August 29, 2019, https://en.wikipedia.org/w/index.php?title=Wikipedia:Press_coverage&diff=877404926&oldid=33446632.
  15. ^ 18. Wikipedia, s.v. "Wikipedia: Wikipedia as a Press Source," accessed August 29, 2019, https://en.wikipedia.org/w/index.php?title=Wikipedia:Wikipedia_as_a_press_source&oldid=601134577.
  16. ^ Katherine Q. Seelye, "Snared in the Web of a Wikipedia Liar," The New York Times, December 4, 2005, https://www.nytimes.com/2005/12/04/weekinreview/snared-in-the-web-of-a-wikipedia-liar.html; Jason Fry, "Wikipedia's Woes," Wall Street Journal, December 19, 2005, https://www.wsj.com/articles/SB113450010488821460-LrYQrY_jrOtOK3IiwMcglEq6aiE_20061223. It was also on the front page of Asharq Al-Awsat's December 9, 2005 print edition and was even reported in Serbian: S. R., "Неслана шала са Википедије," Politika, December 16, 2005.
  17. ^ Allan Metcalf, "Truthiness Voted 2005 Word of the Year by American Dialect Society," American Dialect Society, January 6, 2005, 1–7.
  18. ^ Matthew F. Pierlott, "Truth, Truthiness, and Bullshit for the American Voter," in Stephen Colbert and Philosophy: I Am Philosophy (and So Can You!), ed. A. A. Schiller (Peru, IL: Open Court Publishing, 2009), 78.
  19. ^ The stunt earned Colbert (or at least a user associated with him) a lifetime ban from editing Wikipedia, but others followed suit: the Washington Post 's satirical columnist Gene Weingarten, for example, introduced errors into his own page to see how long they would stand online – a genre in its own right – see, for example, Gene Weingarten, "Wiki Watchee," Washington Post, March 11, 2007, http://www.washingtonpost.com/wp-dyn/content/article/2007/03/06/AR2007030601573.html.
  20. ^ Frank Ahrens, "It's on Wikipedia, So It Must Be True," Washington Post, August 6, 2006, https://www.washingtonpost.com/archive/business/2006/08/06/its-on-wikipedia-so-it-must-be-true/c3668ff7-0b66-4968-b669-56c982c2c3fd/.
  21. ^ David Detmer, "Philosophy in the Age of Truthiness," in Stephen Colbert and Philosophy: I Am Philosophy (and So Can You!), ed. Aaron Allen Schiller, (Peru, IL: Open Court Publishing, 2009); The Colbert Report, July 31, 2016.
  22. ^ Noam Cohen, "Start Writing the Eulogies for Print Encyclopedias," New York Times, March 16, 2008, https://www.nytimes.com/2008/03/16/weekinreview/16ncohen.html.
  23. ^ Simson L. Garfinkel, "Wikipedia and the Meaning of Truth," MIT Technology Review, 111, no. 6 (2008): 84–86.
  24. ^ George Johnson, "The Nitpicking of the Masses vs. the Authority of the Experts," The New York Times, January 3, 2006, https://www.nytimes.com/2006/01/03/science/the-nitpicking-of-the-masses-vs-the-authority-of-the-experts.html; Paul Vallely, "The Big Question: Do We Need a More Reliable Online Encyclopedia than Wikipedia?" Independent, October 18, 2006.
  25. ^ Lawrence M. Sanger, "The Fate of Expertise After Wikipedia," Episteme 6, no. 1 (2009): 52–73.
  26. ^ Catherine Elsworth, "Wikipedia Professor Is 24-Year-Old College Dropout," 'Telegraph', March 7, 2007.
  27. ^ Brian Bergstein, "After Flap Over Phony Professor, Wikipedia Wants Writers to Share Real Names," Associated Press, March 7, 2007; Noam Cohen, "After False Claim, Wikipedia to Check Whether a User Has Advanced Degrees," New York Times, March 12, 2007; Noam Cohen, "Wikipedia Tries Approval System to Reduce Vandalism on Pages," New York Times, July 17, 2008.
  28. ^ Yuki Noguchi, "On Capitol Hill, Playing WikiPolitics," The Washington Post, February 4, 2006, www.washingtonpost.com/wp-dyn/content/article/2006/02/03/AR2006020302610.htm; Nate Anderson, "Congressional Staffers Edit Boss's Bio on Wikipedia," Ars Technica, January 30, 2006 https://arstechnica.com/uncategorized/2006/01/6079-2/
  29. ^ Bobbie Johnson, "Conservapedia – The US Religious Right's Answer to Wikipedia," Guardian, March 2, 2007.
  30. ^ Jacqueline Hicks Grazette, "Wikiality in My Classroom," Washington Post, March 23, 2007, http://www.washingtonpost.com/wp-dyn/content/article/2007/03/23/AR2007032301614.html.
  31. ^ Wikipedia, s.v. "Reliability of Wikipedia," accessed August 29, 2019, https://en.wikipedia.org/w/index.php?title=Reliability_of_Wikipedia&oldid=907354472.
  32. ^ John Schwarz, "When No Fact Goes Unchecked," New York Times, October 31,
  33. ^ Steve Fuller, Post-Truth: Knowledge as a Power Game (London, UK: Anthem Press, 2018), 125; John C. Dvorak, "Googlepedia: The End is Near," PC Magazine, February 17, 2005.
  34. ^ Katie Hafner, "Growing Wikipedia Refines Its 'Anyone Can Edit' Policy," New York Times, June 17, 2006, https://www.nytimes.com/2006/06/17/technology/17wiki.html.
  35. ^ Bobbie Johnson, "Deletionists vs. Inclusionists. Teaching People about the Wikipedia World," Guardian, August 12, 2009.
  36. ^ Noam Cohen, "Wikipedia May Restrict Public's Ability to Change Entries," New York Times, January 23, 2009.
  37. ^ Katie Hafner, "Growing Wikipedia Refines Its 'Anyone Can Edit' Policy"; Idea of the Day Series, "The Epistemology of Wikipedia," New York Times, October 23, 2008.
  38. ^ Terri Oda, "Trolls and Other Nuisances," New York Times, February 4, 2011; Anna North, "The Antisocial Factor," New York Times, February 2, 2011; Henry Etzkowitz and Maria Ranga, "Wikipedia: Nerd Avoidance," New York Times, February 4, 2011.
  39. ^ Ruediger Glott, Phillip Schmidt, and Rishab Ghoseh, "Wikipedia Survey – Overview of Results," United Nations University, March 15, 2010.
  40. ^ Benjamin Mako Hill and Aaron Shaw, "The Wikipedia Gender Gap Revisited: Characterizing Survey Response Bias with Propensity Score Estimation," Plos One, June 26, 2013, https://journals.plos.org/plosone/article?id=10.1371/journal.pone.006578.
  41. ^ Sarah Marloff, "Feminist Edit-a-Thon Makes Wikipedia More Diverse," Austin Chronicle, March 10, 2017, https://www.austinchronicle.com/daily/news/2017-03-10/feminist-edit-a-thon-makes-wikipedia-more-diverse/.
  42. ^ Seth Stephens-Davidowitz, "The Geography of Fame," New York Times, March 22, 2014, https://www.nytimes.com/2014/03/23/opinion/sunday/the-geography-of-fame.html.
  43. ^ Elaine Woo, "Adrianne Wadewitz, Wikipedia Contributor, Dies at 37," The Washington Post, April 25, 2014.
  44. ^ Andrew McMillen, "Meet the Ultimate Wikignome," Wired, February 3, 2015, https://www.wired.com/2015/02/meet-the-ultimate-wikignome/.
  45. ^ William Jordan, "British People Trust Wikipedia More Than the News," YouGov, August 9, 2014, https://yougov.co.uk/topics/politics/articles-reports/2014/08/09/more-british-people-trust-wikipedia-trust-news.
  46. ^ Holly Epstein Ojalvo, "How Do You Use Wikipedia? Media People Talking About How They Use It," New York Times, November 9, 2010, https://learning.blogs.nytimes.com/2010/11/09/how-do-you-use-wikipedia/.
  47. ^ a b Noam Cohen, "When Knowledge Isn't Written, Does It Still Count?" The New York Times, August 7, 2011.
  48. ^ Andrew Lih, "Wikipedia Just Turned 15 Years Old. Will it Survive 15 More?" The Washington Post, January 15, 2016, https://www.washingtonpost.com/news/the-intersect/wp/2016/01/15/wikipedia-just-turned-15-years-old-will-it-survive-15-more.
  49. ^ Noam Cohen, "Conspiracy Videos? Fake News? Enter Wikipedia, the 'Good Cop' of the Internet," The Washington Post, April 6, 2018, https://www.washingtonpost.com/outlook/conspiracy-videos-fake-news-enter-wikipedia-the-good-cop-of-the-internet/2018/04/06/ad1f018a-3835-11e8-8fd2-49fe3c675a89_story.html.
  50. ^ Corine Ramey, "The 15 People Who Keep Wikipedia's Editors from Killing Each Other," Wall Street Journal, May 7, 2018, https://www.wsj.com/articles/when-wikipedias-bickering-editors-go-to-war-its-supreme-court-steps-in-1525708429.
  51. ^ Omer Benjakob, "Verified: Wikipedia Is Our Unlikely Champion in the War of Fake News," Wired UK, July 2019, 22–23.
  52. ^ Madrigal, "Wikipedia, the Last Bastion of Shared Reality"; Stephen Harrison, "Happy 18th Birthday, Wikipedia. Let's Celebrate the Internet's Good Grown-up," Washington Post, January 14, 2019, https://www.washingtonpost.com/opinions/happy-18th-birthday-wikipedia-lets-celebrate-the-internets-good-grown-up/2019/01/14/e4d854cc-1837-11e9-9ebf-c5fed1b7a081_story.html.
  53. ^ Omer Benjakob, "Is the Sky Blue? How Wikipedia Is Fighting for Facts by Redefining the Truth," Haaretz, December 15, 2017, https://www.haaretz.com/science-and-health/.premium.MAGAZINE-how-wikipedia-is-fighting-for-facts-by-redefining-the-truth-1.5628749.
  54. ^ Madrigal, "Wikipedia, the Last Bastion of Shared Reality."
  55. ^ Stephen Amstrong, "Inside Wikipedia's Volunteer-run Battle Against Fake News," Wired, August 21, 2018, https://www.wired.co.uk/article/fake-news-wikipedia-arbitration-committee.
  56. ^ Benjakob, Omer (May 17, 2018). "The Witch Hunt Against a 'Pro-Israel' Wikipedia Editor". 'Haaretz.
  57. ^ Omer Benjakob, "Breitbart Declares War on Wikipedia as Encyclopedia Gets Drafted into Facebook's 'Fake News' Battle," Haaretz, April 24, 2018, https://www.haaretz.com/us-news/.premium-breitbart-declares-war-on-wikipedia-in-facebook-s-fight-against-fake-news-1.5991915.
  58. ^ Lachlan Markay and Dean Sterling Jones, "Who Whitewashed the Wiki of Alleged Russian Spy Maria Butina?" Daily Beast, July 24, 2018, https://www.thedailybeast.com/who-whitewashed-the-wiki-of-alleged-russian-spy-maria-butina/.
  59. ^ Brian Feldman, "Why Wikipedia Works," New York Magazine, March 16, 2018, http://nymag.com/intelligencer/2018/03/why-wikipedia-works.html.
  60. ^ Martin Dittus and Mark Graham, "To Reduce Inequality, Wikipedia Should Consider Paying Editors," Wired UK, September 11, 2018, https://www.wired.co.uk/article/wikipedia-inequality-pay-editors.
  61. ^ Katherine Maher, "Wikipedia Mirrors the World's Gender Biases, It Doesn't Cause Them," Los Angeles Times, October 18, 2018, https://www.latimes.com/opinion/op-ed/la-oe-maher-wikipedia-gender-bias-20181018-story.html.
  62. ^ Brian Feldman, "Why Wikipedia Works," New York Magazine, March 16, 2018, http://nymag.com/intelligencer/2018/03/why-wikipedia-works.html.
  63. ^ "Google Blames Wikipedia for Linking California GOP to Nazism," Associated Press, June 1, 2018.



Reader comments

2021-01-31

Wikipedia's war against scientific disinformation

Mary Mark Ockerbloom was a Wikipedian-in-residence at the Science History Institute from 2013 to 2020.

The playbook for undermining scientific expertise was created in the 1920s. Perhaps surprisingly, its creators were scientists. Wikipedia is one battleground in the hundred years' war against scientific disinformation. The tactics of scientific disinformation that were developed then are still used today. We need to be aware of these tactics so that we can counter them; we also need more scientifically-savvy allies to help us get it right.

The practice of scientific disinformation

Tetraethyllead (TEL) at the point of sale
Charles Kettering, on a Time cover, 1933

On December 3, 1921,

Charles Kettering, scientists at General Motors, discovered that adding tetraethyllead (TEL) to gasoline improved a car's performance.[1] As patent holders of the new technology who held senior positions in the companies producing and marketing TEL, Kettering and Midgely had everything to gain by promoting TEL's use rather than developing other non-patentable alternatives.[1][2]

TEL contains lead: an odorless, colorless, tasteless, poisonous element. Once present in a person, lead does not degrade, it accumulates.[3][4] Kettering and Midgley were warned of the dangers of lead and TEL by other scientists. Erik Krause of the Institute of Technology, Potsdam, Germany had studied TEL extensively, and described it as "a creeping and malicious poison".[2] Internal confidential reports document that the companies involved knew TEL was "extremely hazardous",[1] while their safety precautions were "grossly inadequate".[1]

"Loony gas", New York Journal, October 31, 1924

By 1923, cases of violent madness and death were being reported among workers at TEL plants. At least 17 persons died, and hundreds more suffered neurological damage.[5][6] Postmortems confirmed the cause as tetraethyllead.[7] Coworkers referred to the fumes they breathed as "loony gas" and to a building they worked in as the "House of Butterflies", because workers had hallucinations of insects.[6]

Tetraethyllead

The newspapers, the public, and the government began to take notice.[5] Midgley himself was treated for lead poisoning. Despite this, he protested that TEL was perfectly safe, even washing his hands in it in front of reporters on October 30, 1924,[1][3] the same day the New York City Board of Health banned the sale of TEL-enhanced gasoline.[1][3]

As concerns about TEL grew, Kettering took protective measures for the company, but not its workers. He hired Robert A. Kehoe, a 1920 medical school graduate, as an in-house "medical expert" whose job was to prove that leaded gasoline did not harm humans. An early study is illustrative of the methodological problems in his work. In that study he reported that workers handling TEL had levels of exposure no higher than a control group, which was composed of workers at the same plant. He suggested their levels of lead were "normal", equating normal and harmless.[8]

Alice Hamilton was among those who criticized Kehoe's research. A pioneer in industrial toxicology and occupational health, Hamilton was America's leading authority on lead poisoning, with decades of experience in public health research and policy-making.[9] On May 20, 1925, Hamilton and other public health advocates from Harvard, Yale and Columbia faced off against Kettering, at a conference called by the Surgeon General of the Public Health Service to consider the use of lead in gasoline. Would lead in gasoline be released into the air when the fuel was burned, putting the general public at risk? Hamilton warned that lead posed environmental as well as occupational dangers. "You may control conditions within a factory, but how are you going to control the whole country?"[10] Kettering argued that TEL was the only way to improve gasoline, and asserted that no one had proved that leaded gasoline was harmful.[1][10]

The conference ended with the formation of a committee to further investigate the possible effects of leaded gasoline. The committee could have taken responsibility for further independent research, but public health advocates lacked funding, and the government wasn't willing to provide it. In a classic case of setting the fox to watch the henhouse, the commission chose to rely on industry to monitor itself, ignoring the inherent conflict of interest. Not surprisingly, Robert Kehoe reported that the industry-funded research showed "no evidence of immediate danger to the public health."[10][11]

Hamilton reportedly told Kettering to his face that he was "nothing but a murderer".[2][1] Kettering formed the Kettering Foundation for public policy-related research. Kettering and Midgley developed Freon, which eventually became another environmentally disastrous product.[2] Kehoe became the gasoline industry's chief spokesperson, largely controlling the next fifty years of scientific and public narrative around leaded gasoline.[11][8]

Risk, responsibility and the undermining of science

Public health advocates and corporate representatives took different approaches to risk and responsibility in the debate over leaded gasoline. Public health followed a precautionary principle, arguing that unless scientists could demonstrate that something was safe, it should not be used. The Kehoe paradigm laid the burden of proof on the challenger. Rather than demonstrating that their product was safe, they demanded that critics prove it was harmful. But Kehoe's rule is logically flawed – absence of evidence of risk does not imply evidence of the absence of risk. Regardless, the Kehoe paradigm became extremely influential in the United States.[11][8][12]

This not only set scientist against scientist, it made the undermining of science tactically useful. Experimental research is a method of testing ideas and assessing the likelihood that they are correct based on observable data. By its very nature, the results of a scientific study do not provide a 100% yes-or-no answer. Disinformation exploits this lack of absolute certainty by implying that uncertainty means doubt. If evidence is presented to challenge a position, it is suggested that the proof is not sufficiently compelling, that doubt still remains, that more research must be done, and that no responsibility need be taken in the meantime. By repeatedly raising the issue of doubt, companies and scientists use "cascading uncertainty" to manipulate public opinion and protect their own interests.[13][14][15][12]

In the 1960s geochemist

Clair Patterson developed sophisticated monitoring equipment and methods to measure the history of the earth's chemical composition. Initially uninterested in lead, his research provided compelling evidence of its extraordinary increase in the planet's recent history. Patterson broke the industry's leaded gasoline narrative. The industry spent almost 25 years trying to discredit him and his work through professional, personal, and public attacks. Nonetheless Patterson did what Kettering and Kehoe had challenged public health officials to do in 1925 – demonstrate the impact of TEL from gasoline.[11][8][12]

By then, lead contamination from gasoline (and from paint) was found worldwide, not just in North America. The World Health Organization (WHO) considers lead to be one of the top ten chemicals posing a major public health risk, with immense personal, social and economic costs. One of the "key facts" they state is "There is no level of exposure to lead that is known to be without harmful effects."[4]

Kettering's playbook has been replicated repeatedly. Companies and their researchers have argued that

fossil fuels and climate change
(to name only a few) aren't really dangers; that concerns about public safety and environmental damage have not been sufficiently proven and so do not require action; and that raising a shadow of doubt is enough to challenge widespread scientific consensus. I strongly recommend reading Merchants of Doubt (Oreskes & Conway, 2010)[13] and Doubt Is Their Product (2008)[14] and The Triumph of Doubt (2020) by David Michaels.[15]

At its most extreme, we face an attitude that assumes scientific issues are simply matters of opinion and belief, independent of underlying scientific evidence and informed consensus. The dangers of this are apparent: We are in the midst of a pandemic that some refuse to believe exists. Anti-mask and anti-vaccination propaganda are recent areas of scientific disinformation, putting us all at risk.[16]

Defeating scientific disinformation on Wikipedia

Companies and individuals who try to whitewash their home pages or promote pseudoscience cures are fairly obvious examples of conflict of interest and promotional editing. I suspect most Wikipedians both recognize and know how to deal with such situations. But there are patterns of disinformation, manipulation, and the undermining of science that are harder to spot and address. We need to watch for patterns of bias that go far beyond articles for a specific company or product.

Tactics in the disinformation playbook include:

  • Ignore warnings and deny reports of possible safety issues,
  • Accuse critics and the media of sensationalism,
  • Fund "experts" to conduct research that can only support your position,
  • Control the narrative and reframe issues to avoid consideration of risk,
  • Use any level of doubt to claim there is not enough evidence of risk,
  • Ignore scientific consensus,
  • Personally attack those who challenge you,
  • Avoid acknowledgement of responsibility and delay taking actions that could address issues.

Do funders and researchers have a vested interest in a particular outcome? Too often, companies are funding the research that claims their products are safe or effective. Scientific articles will generally indicate who supported the research. Check out the funders, and watch out for industry organizations, think tanks and researchers with a history of anti-regulatory bias. Look for independent evidence-based sources from credible groups like Consumer Reports.

In striving for balance, we need to recognize the importance of counter-narratives around public health. Actively look for scientific work that raises issues of public health and follows a precautionary principle. These are important concerns, and addressing these concerns is part of presenting a balanced picture. Be wary of the burden of proof. Are different expectations being applied to the research of proponents and critics of an idea? Keep in mind that industry pours huge amounts of money into its research while public health receives far less.

In a war of supposedly scientific claims and counterclaims, are proponents of a particular position using the technique of raising doubt? Is research critical of a position repeatedly minimized or dismissed on the grounds that it fails to meet some increasingly strict or absolute standard of proof? Are personal attacks and appeals to public opinion being used to displace scientific evidence? These tactics can be particularly insidious in Wikipedia articles, because we are encouraged to present all sides of an issue. As editors we need to remember that writing in a fair and balanced way doesn't mean that all ideas have to be given equal weight. Present them in proportion to their importance.

Beware of cherry picking: it is encouraged and amplified by social media, and Wikipedia articles can be particularly susceptible to the selective presentation of information. Look for the weight of scientific consensus on an issue. If the vast majority of scientists worldwide accept that climate change is real (as they do), you can be definite about it.

Wikipedia needs more scientific expertise

The tactics of disinformation complicate the difficulties of writing about science. Let's be clear – writing about science is hard! Writing about science for Wikipedia's readers – much of the world's population – is an even more challenging task. First you must wrap your own head around an area of expertise that may be highly specialized, and then you must communicate your understanding to readers who may lack a scientific background or do not share your frame of reference. As the Wikipedian-in-residence at the Science History Institute in Philadelphia for seven years, I rarely read a science-related article on Wikipedia that did not need significant improvement. Working on biographies of scientists and other pages with historical content, I asked myself "Where is the science?"

When it comes to science, Wikipedia needs all the help it can get. There are scientists who have manipulated and undermined scientific information, as I describe above, but there are many more scientists whose credentials, expertise and research are rock-solid. We need to find ways to engage with those scientists and leverage their knowledge. Doing that was the most challenging part of my job as a Wikipedian-in-residence.

We need more editors with scientific expertise. We need editors with an awareness of how scientific information can be manipulated. We need science communicators who can evaluate scientific materials and write about science in a way that is comprehensible to nonscientists. We need as many people as possible keeping a careful eye on scientific information on Wikipedia to prevent the kinds of manipulation we know occurs. I often worry that a collection of volunteer editors hasn't a hope of keeping up. We need allies.

In the ongoing war against disinformation, Wikipedia needs scientific expertise – and science needs Wikipedians to get things right.

References

  1. ^ .
  2. ^ .
  3. ^ a b c Kitman, Jamie Lincoln (March 20, 2000). "The Secret History of Lead". The Nation.
  4. ^ a b "Lead poisoning and health". World Health Organization. 23 August 2019.
  5. ^ a b Kovarik, Bill. "The Ethyl conflict & the media: Paper to the AEJMC, April 1994". Prof. Kovarik Utinam patribus nostrius digni simus. Retrieved 28 January 2021.
  6. ^ a b Blum, Deborah (2011). The Poisoner's Handbook: Murder and the Birth of Forensic Medicine in Jazz Age New York. New York: Penguin.
  7. .
  8. ^ .
  9. .
  10. ^ .
  11. ^ .
  12. ^ a b c Percival, Robert V. (2006). "Who's Afraid of the Precautionary Principle?". Pace Environmental Law Review. 23 (1).
  13. ^ .
  14. ^ .
  15. ^ .
  16. .



Reader comments

2021-01-31

The world's press says "Happy Birthday!" with a few twists

Media throughout the world have written congratulatory articles about Wikipedia on the encyclopedia's 20th birthday. The Wikimedia Foundation has done a good job letting the press know that we're big, popular, appear in about 300 languages, and are frequently edited, mostly by volunteers. See, for example this list. But what else is new? A couple of dozen newspapers just focus on these numbers and add some seemingly random or local-interest facts. The more interesting stories take different angles, often telling as much about the author and their relationship with the encyclopedia, as about Wikipedia itself. Fair enough! No journalist can be expected to summarize Wikipedia in a single article.

United States Capitol protests/storming/riots/insurrection

With a twist

Where we go from here?

  • Wikipedia is twenty. It’s time to start covering it better Stephen Harrison and Omer Benjakob, in the Columbia Journalism Review address their fellow journalists using many of the same examples as they use to address academics in their chapter in Wikipedia @ 20 and to address Wikipedia editors in this issue's In focus. They conclude that "journalism should help readers better understand Wikipedia’s policies and inner workings." Since "Wikipedia’s diverse and often divergent community is the secret to its success," journalists should report on the community with more nuance than writing as if all editors have a single point of view or engage in a constant free-for-all.
  • Wikipedia's Biggest Challenge Awaits in 2021 in Wired by the dean of Wiki-journalists Noam Cohen, starts with a retelling of Wikipedia's many achievements but then moves to the Universal code of conduct and why it is sorely needed. The changes proposed by the WMF and many others to diversify the editing community raise existential questions for the online encyclopedia. He uses the examples of the articles Mythology of Benjamin Banneker and Der Giftpilz, a Nazi children's book. The mythology article is not really about Banneker, an accomplished African-American, but a strong critique of stories made up about him. Why don't we have similar articles about the mythology of George Washington or Albert Einstein? The Nazi children's book article included a link to the book's contents and might promote the book in other ways. Both articles will make editing Wikipedia a harrowing experience for some editors. "This is the gnawing challenge for Wikipedia. After a period of wild, unrestrained growth, it needs some civilizing laws. The equivalent of a fair housing act and safety inspections to ensure it won’t exclude certain groups from its pages and allow hateful material to grow and fester."

In brief

Hunters and busybodies-Infographic (Melissa Pappas (UPenn))

Notes

  1. ^ The page was being discussed for deletion on ja.wiki when this issue was being prepared, and was deleted on January 27.



Do you want to contribute to "In the media" by writing a story or even just an "in brief" item? Edit next week's edition in the Newsroom or leave a tip on the suggestions page.



Reader comments

2021-01-31

The people who built Wikipedia, technically

Servers in the Wikimedia Foundation's "eqiad" datacenter in Ashburn, Virginia.
Legoktm is a site reliability engineer for the Wikimedia Foundation. He wrote this in his volunteer capacity.

Wikipedia has seen incredible success thanks to the efforts of thousands of dedicated contributors and the technical side is no exception. From backend and infrastructure developers to translators to bot and script writers, Wikipedia grew from a hobby project sitting on a single server to a robust platform operating out of datacenters on multiple continents.

Erik Möller, former Deputy Director of the Wikimedia Foundation and an early developer, noted in an email, "What always amazed me working in this environment is how many brilliant people found a niche in which they individually made incredibly impactful contributions, and that seems as true today as it ever was."

Starting with trust

At the very beginning, getting server access was "pretty loosey-goosey",

Brion Vibber
, an early developer and the first employee of the WMF, said in an email. "If you showed up and put in good work helping out, you might well hear 'yes' to getting some fairly direct access to things because that was the only way we were going to get anyone to do it!"

Assume good faith applied here as well. Möller said, "As I recall, in the very early years, these decisions were made based largely on trust by folks like [Jimmy Wales] or the people he had already delegated access to (e.g., [Vibber]), with a high assumption of good faith that the people who showed up to help had no darker motives."

The source code was maintained and developed in a CVS repository on Sourceforge.net. Gabriel Wicke, a developer and later Principal Software Engineer at the WMF, said in an email, "Getting revision control access (CVS at the time) basically was about winning trust with whoever set up accounts, which I strongly suspect was [Vibber]."

Tim Starling, an early developer and current Principal Software Architect at the WMF, said in an email he got CVS access from Lee Daniel Crocker as soon as he said he was interested. "There was no pre-commit review, but the code on the server (just one or two servers back then) was not automatically updated, so commits were theoretically reviewed before they went live," he said.

Getting root (also known as superuser access) was a bit harder.

"I remember there being a ridiculously long and painful period between getting shell access and getting root, like six months," Starling said. "I had read/write access to the database, I could edit the code, I could view the server access logs, but for some reason root was a big deal."

Domas Mituzas, a former system administrator and WMF board member, said in an email, "It took me a bus trip to Berlin and sleeping on a German Wikipedian's couch and meeting everyone (we all were meeting each other for the first time!) that eased everyone into the idea of giving me root."

Scaling up

By 2003, significant changes needed to be made on the software side to accommodate the quickly increasing traffic. At the time, Wikipedia was running on only two servers and was rendering every page view from scratch, Wicke said.

"A then-fancy 64-bit Opteron DB server upgrade helped briefly, until it started crashing," he said. "The site was often down, and it was clear that any further growth would quickly consume whatever hardware had just been added."

A diagram from Gabriel Wicke's Squid caching proposal demonstrating how the servers could be laid out.

After posting a proposal to add caching using Squid, and receiving feedback from Wales, Vibber and Jens Frank, Wicke "...went ahead and prototyped a basic Squid integration with active cache purging, initially on [his] own servers," which would be serving the main site by February 5, 2004.

"There were issues of course, like missing purges for images or transclusions," he said. "Those were fairly quickly resolved or worked around, and there was a lot of tolerance given the preceding phase of poor site availability."

A few weeks later on February 25, Wikipedia was featured in German news program Tagesthemen (watch on YouTube). Watching the live traffic stats, Wicke said they got "...all excited on IRC when the site briefly went from ~25 [requests per second] to around 1,500 without falling over."

More servers

Wikipedia originally ran on servers managed by Bomis, a dot-com startup. Starling said, "Before February 2004, Jason Richey, a Bomis employee, managed the hardware and would occasionally log in and restart things or otherwise try to fix downtime."

Sometimes this involved literally going the extra mile, as he lived in Los Angeles, while the servers were in San Diego. "I remember [Richey] having to drive 4 hours to San Diego to fix downtime caused by a really simple problem, like a broken hard disk," Starling said.

Some tasks required his intervention that today seem unthinkable. "My favorite memory from the very early Bomis days is that if you wanted to upload an image, you emailed a guy named Jason who would helpfully place it on the server for you," Möller said.

The servers in Tampa, Florida on the first day they were installed in 2004.

In 2004, Wikipedia moved to a datacenter in Tampa, Florida, to be closer to Wales's new home. "I believe Wales helped to rack the first batch of servers in Tampa," Starling said.

A year later, the Board named Mituzas the Hardware Officer, putting him in charge of, as he describes it, placing servers in a shopping cart and then asking Wales to pay for them.

"Instead of ordering servers one by one I tried a more exponential approach (buy 20, then 40, then 80, ...) – and each time we'd land those batches the site would get much snappier and within a few weeks we'd have more users to fill all the capacity," Mituzas said. "We bought cheap servers that needed hands in datacenters to do anything with them, but we had the capacity to survive the growth."

When others in the WMF's leadership wanted to use some funding to pay other bills, he said he pointed out that if the site wasn't up, there wouldn't be any other bills to pay.

"The challenging part of the role was being the first one to grab [Wales] when he got online – so paying attention to IRC notifications was key, otherwise we would not get our servers," Mituzas said.

Growing pains

Between rapid growth in traffic and not having enough technical resources while constantly implementing new features, Wikipedia was down rather frequently.

Mituzas recalled one outage from when nearly all of the developers met up for the first time in Berlin. "Kate, who wasn't at the meeting deployed Lucene-based search that nobody knew about and thus we were trying to understand why is Java running on our servers and why is it taking everything down."

Other times, developers created contingency plans in response to real world events; in one case it was because of Hurricane Charley.

"I didn't really know much about hurricanes or what to expect, but the local media was talking up the threat," Starling said (he lived in Australia at the time). "There was a risk that power and/or network access would be temporarily lost. We were making off-site backups of our data as if there was a chance of Tampa being flattened like a modern-day Galveston, which was maybe a bit of an overreaction, although I guess it's nice that we were making off-site backups of private data for the first time ever."

In 2005, The Signpost reported how tripped circuit breakers took the site down. It took developers a full day to restore editing after the primary database became corrupt.

"[That] outage caused Wales to quip that downtime is our most profitable product," Starling said.

Changes were often queued up for months at a time before being deployed to wikis in a single release. But eventually the breaking point was hit, Roan Kattouw, a current Principal Software Engineer at the WMF, explained in an email.

"In early 2011, we scheduled a 6-hour window to attempt to deploy the 1.17 upgrade and fix any resulting issues, and decided that if we couldn't fix things within those 6 hours, we would roll back to 1.16," Kattouw said. "This time window was from the morning to the early afternoon for me in the Netherlands, from the late afternoon into the evening for [Starling] in Australia, and from the late evening into the night for our US-based colleagues. The first two times we tried it, a lot went wrong, the site went down at times and we had major issues that we couldn't fix quickly enough, so we rolled back."

It took about three tries for them to get it working "successfully," he said.

"For some definition of success, that is: the site was up and was stable, without any critical issues. After the window ended, I spent the rest of the day fixing various outstanding issues while the others slept, and I passed the baton to Tim when he woke up," Kattouw said. "One of the issues I remember us encountering was all redirects being broken on French-language wikis. Today, that kind of issue would be considered a major problem and a train blocker, but that day it was so far down the priority list that we left it broken for about 12 hours."

Soon after developers began working on "heterogeneous deployment", allowing for progressive deployments.

"This way we can deploy a new version to only a few small wikis first, and work out the kinks before deploying it to larger wikis," Kattouw said. "We were able to accelerate this over time, and nowadays the deployment train runs every week, with major wikis getting new changes only two days after the first test wikis get them."

Expanding functionality

Wikipedia's Main Page as it appeared on 20 December 2001, powered by UseModWiki.

Wikipedia originally ran on the

watchlists. It would be officially named "MediaWiki" when it was rewritten by Lee Daniel Crocker
.

Other features taken for granted today like an autogenerated table of contents and section editing were controversial when initially introduced.

"As I recall the table of contents feature was a bit more contentious (no pun intended), mostly because of the automagic behavior (hence flags like __NOTOC__ were created to control it)," Möller said. "With section editing, the first visual design was a bit cluttered (and of course there were still kinks to iron out, e.g., in the interaction with protected pages), but I think most people could fairly quickly see the appeal."

In other cases, editors forced developers to add features to the software. Carl Fürstenberg, a Wikipedia administrator, created {{qif}}, which allowed for conditional logic in templates.

"At one point I realized that the parameter expansion logic could be 'misused' to create a way to inject boolean logic into the at time limited template syntax, which could be helpful for creating more generic templates which didn't have to call chains of helper templates to create the same as it was before," he said in an email. "Thus I created Qif, Switch, and different boolean templates."

The developers weren't pleased. Starling wrote to the wikitech-l mailing list that he "...caved in and [had] written a few reasonably efficient parser functions...[that] should replace most uses of {{qif}}, and improve the efficiency of similar templates."

Fürstenberg said he never expected {{qif}} to ever be used so widely. "I think I first realized it had become widely used when it had to be protected as any edit to it halted Wikipedia for a while," he said.

In his 2006 mailing list post, Starling blamed the 2003 introduction of templates and the MediaWiki namespace and said he didn't understand "what a Pandora's box" he opened. But that functionality was key to enabling one of MediaWiki's greatest strengths: localization, allowing users to use the software in their preferred language.

Niklas Laxström, the founder of translatewiki.net and a WMF Staff Software Engineer, said in an email he originally submitted translations via Bugzilla, and then worked up his courage to ask Vibber to deploy them for him, sometimes breaking the Finnish Wikipedia because he forgot a semicolon.

"It was no wonder then, that many opted to do translations in the Wikipedias itself using Special:AllMessages. There was no risk of syntax errors and changes were live immediately, as opposed to potentially taking many months as deployments were few and far between," Laxström said. "By the way, this is a unique feature; I have not seen other websites which allows translation and customization of the whole interface by the users of the site using the site itself."

Scratching his own itch, Laxström started modifying Special:AllMessages to make translation easier, but didn't feel those changes were acceptable to go back into MediaWiki, so he hosted them on his own wiki. Today, nearly all localization of Wikipedia's interface is done via translatewiki.net, rather than on individual wikis.

He credits Raimond Spekking with having managed MediaWiki's localization process for over a decade now.

"[Spekking] checks the changes to mark the translations outdated where necessary, he renames messages and performs other maintenance activities. He exports translation updates multiple times per week." Laxström said. "He does this so well that it can feel like magic."

Divesting power

Early versions of Wikipedia's software gave immense power to developers. Only developers could block users, promote new administrators, rename users, and so on.

Seeing this as a problem, in 2004 Starling wrote an email to the Wikipedia-l mailing list titled "Developers should mind their own business", proposing that certain user rights be split into a separate group.

Wikipedia should not be a technocracy, ruled by those with knowledge of computer systems. Wikipedia should be a democracy. Those in power should be accountable to the community at large, and ideally selected from and by the community at large.

Today, Starling describes that shift in power as having been a big deal at the time. "I was very conscious of the fact that I was designing a social system," he said. "As you can guess from that email, I was uncomfortable with the fact that the power to do so had somehow fallen to me, but I wanted to get it right."

He credits Sunir Shah, the founder of MeatballWiki, for discussing "...that change with me at length, as well as other changes at the interface of social policy and technical design."

It's unclear how much lasting impact this change had, given the WMF's rise to the top of the Wikimedia power structure, in large part because it controls the majority of developers and servers. In 2014, Möller instituted "superprotect", which allowed the WMF to protect a page from even administrators editing it.

"[Möller's] idea was that it would be used in cases of conflict between the Foundation and the community, as a softer alternative to de-sysopping," Starling said. "When that conflict came, [Möller] asked me to make the necessary group rights changes. I said that I was uncomfortable putting my name to that on the wiki, so he found someone else to press the button."

Starling has a simple conclusion as to why the WMF has risen to the top of the power structure: Wikipedia lacks leadership.

"I would like to see an elected editorial board with the mandate and courage to make major policy changes," he said. "Without such a body, WMF necessarily fills in the power vacuum, although it is too timid to do so effectively, especially on any question relating to the content."

"The most controversial move in Wikipedia history"

Derek Ramsey originally wanted to create an article about a town he knew, but couldn't come up with anything more than a couple of sentences.

"So I came up with a solution: find a public domain mass data set that allowed a (somewhat) stub useful article to be created," he said in an email. "I've always had an interest in mass data processing, so this was something I knew I could do. I imported a number of census data tables into a MySQL database tables running on my own Linux computer. I correlated the data with other geographic data sources."

After cleaning up and performing other validation steps, Ramsey said he generated more than 3,000 text files for articles about United States counties and began adding them to Wikipedia by hand.

"This was extremely tedious and slow, but effective," he said. "However, there were 33,832 cities, and that would have taken an order of magnitude longer to complete."

He first wrote a Java program to read each article and make HTTP requests to post it to Wikipedia, later coding "... in features like error checking, error correction, throttling, pauses for human verifications, and other features."

Increasing the number of articles in Wikipedia by 40%, Andrew Lih later called it "the most controversial move in Wikipedia history" in his 2009 book The Wikipedia Revolution.

"I was

ignored all rules
. You could still do that back then," he said. "After all, if I could edit articles manually, what difference did it make if I did the same thing automatically? It saved me time, but the end result was identical."

Out of all the controversy around mass article creation came two key things that Wikipedia still uses today.

First, Ramsey created {{cite web}}, which most references on Wikipedia today use (and featured in xkcd).

"I wanted a generic way for Wikipedians to cite their sources easily, since prior to this the only citations made were manual and inconsistent," he said. "This was necessary because it was before the devs created native reference support in the Wikimedia software."

Second, Ramsey worked with other Wikipedians to develop an early bot policy. The initial version contained the contradictory statement, "In general bots are generally frowned [upon]."

"I thought the concern was overblown, but the consensus was demanding that something be done to address the perceived issues," Ramsey said. "It was my desire to get ahead of the issue before draconian measures shut it all down, so I created bot policy as a sort of compromise. I figured it was better to do that than to have all bots banned wholesale."

Soon after users started running bots and scripts under administrator accounts, dubbed "

.

"There was a lot of hysteria surrounding adminbots on the English Wikipedia but a few people quietly ran them, as far back as like 2005," Max McBride, a bot operator who previously ran adminbots, said in an email. "Some of these scripts were admittedly kind of terrifying and there weren't as many tools to mass undo unintentional damage."

McBride described people's attitudes on adminbots as "beyond reason" and suggested they were based on some sort of jealousy. "Like a random script gets admin rights and an admin gets two admin accounts, but not lots of regular users," he said. "I think that bred and fed some opposition."

Strong security

Unlike many other prominent websites, Wikipedia hasn't suffered from an embarrassing, public security incident losing its users' private data. Some of this is due to a lack of collecting private data in the first place, but since the beginning there has been a strong culture of focusing on security.

"In the early days, the worst case scenario was irreversible destruction of large amounts of user work, since we didn't have the resources to make frequent backups," Starling said. "I spent a lot of time doing security reviews, and informed by that work, I wrote policies and improved our APIs and conventions."

He also credited Vibber with making key policy decisions for other MediaWiki installations (disabling uploads by default and having a non-web writable source tree), that ensured MediaWiki didn't become a "...constant source of botnet nodes like some other PHP web applications."

But for readers and editors to visit the site securely required using a special secure.wikimedia.org gateway until native HTTPS support was rolled out in 2011 as an opt-in option.

An NSA slide explicitly listing Wikipedia as being visited over insecure HTTP protocol.

Then in 2013, whistleblower Edward Snowden revealed that the NSA was targeting Wikipedia users visiting the site over the default, unencrypted HTTP protocol.

great firewall of China
was well known for this, and they were sharing this tech with other governments), but the Snowden leaks showed that the government was explicitly targeting Wikipedia users," he said.

Kattouw worked on the MediaWiki-side of the HTTPS change, allowing for

protocol-relative URLs
to be used. "I think the [Site Reliability Engineering] people who worked on the HTTPS migration deserve more credit," he said. "That was a much more difficult migration than many people thought."

The politics involved in making the switch weren't the regular WMF vs. community ones, it was actual global politics.

"For example, Russian Wikipedia asked us to implement HTTPS only (for all users, not just signed-in users) as soon as possible, as they wanted to head off Russian legislation that would have enabled per-page censorship, and it would have forced the government to choose between blocking all of Wikipedia, which was politically difficult, or dropping their aim at per-page censorship," Lane said. "This is why Russian Wikipedia got support before any other wiki (and more extensive support, at that). Chinese Wikipedia, on the other hand, asked us to delay rollout, as the Chinese government was already doing per-page censorship, and had previously blocked all of Wikipedia a number of times."

There's one large exception to this focus on security: the ability for users to create custom scripts and styles and share them with other users, on the wiki. In web development, this is typically known as a cross-site scripting vulnerability, but for Wikipedia it was a feature.

Fürstenberg created one of the most popular user scripts, Twinkle. He said it started as a helper for himself to "...conduct anti-vandalism and maintenance easier, from the point of reverting quickly, to the tedious task of filing reports to different sections. It pretty much boiled over from there."

Looking back, Vibber thinks the idea of user scripts is great, but implemented incorrectly. He said there are two primary problems:

  1. Running someone else's malicious code can lead to your account being taken over
  2. Script code access internal data and methods that aren't going to stay stable, potentially breaking over time.

"Both can be solved by using a sandboxed environment (probably a suitable iframe)," Vibber said. "I think there's a lot of cool stuff that can be built on top of this method, with full-on APIs for accessing an editor state as a plugin, or whatever."

Missed opportunities

Mockup of what Athena might have looked like from Brandon Harris's Wikimania presentation.

At Wikimania 2012 and then in a Signpost op-ed, then-WMF Senior Designer Brandon Harris presented the "Athena Project", outlining a vision for what Wikipedia should look like in 2015.

Suffice to say, that vision was never fully implemented, and Harris said in an email he could write a book as to what went wrong. "I'd say the primary reason was the fact that Foundation had a stellar lack of focus and a muddled leadership direction which allowed for lower-level political infighting to thrive," he said.

The reaction to Harris's proposal was generally mixed to negative, but that's what Harris was hoping for. "A thing a lot of people – even professional designers – don't understand about the design process is that only 10% of it is actually 'designing,'" he said. "Most of it is marketing: You have to understand the market you're designing for to know what to design and you have to convince folk that your design solves the problem. You may have to sell the idea that the problem even exists!"

Part of the purpose of his proposal was an exercise in re-examining the entire interface, something he said neither the WMF nor community do enough of. "Look at what happens, every day! Nothing has changed since 2015," Harris said. "The Foundation still doesn't know how to sell its ideas and it keeps trying to fix the same problems with the same tepid changes to the toolchain. The community still doesn't know how to govern itself and still keeps using the same broken processes to inadequately deal with the same issues."

McBride wrote an op-ed response to Harris, titled

FlaggedRevs, which was supposed to solve the BLP problem
.

"It wasn't a proposed solution as much as it was the only 'solution,'" he said. "And a lot of people had pinned their hopes on it being successful, but I was more interested in it failing fast so we could move on and try other solutions."

After various trials, and years of RfCs, Flagged Revisions (now rebranded on the English Wikipedia as "Pending Changes") is barely used on BLP pages, unmaintained and no longer enabled on new wikis. (It's worth noting that some communities like the German Wikipedia view it as a success.)

"The BLP problem is definitely not fixed," McBride said. "And there's an enormous gap between current tech and what could be implemented to alleviate the problem."

In his op-ed, McBride questioned whether upcoming projects like VisualEditor would end up with a similar fate as FlaggedRevs. As it turned out, the rollout of VisualEditor and Media Viewer a few months later were both extremely controversial among Wikipedians (not to mention the separate but related issue of superprotect), something that Möller acknowledges in hindsight.

"In both cases, a more gradual rollout (probably adding at least 1–2 years to the release timeline for VE, and 6–12 months for MV) could have prevented a lot of pain and frustration," Möller said. "I take on my share of responsibility for that."

Both Möller and McBride independently brought up the same saying: "change moves at the speed of trust" (Möller credited Lydia Pintscher for teaching it to him).

"For that principle to work in practice, an organization has to be prepared to let go of overly rigid timelines and commitments, because its commitments must always first and foremost be to the people whose trust it seeks to earn and keep," Möller said. "That doesn't mean it's impossible to make radical, transformative changes, but it can certainly feel that way."

McBride put it more bluntly. "Wikimedians don't like shitty software, they quickly embrace good software (think @

mass messages
or...)," he said. "A lot of software is bad and is imposed on the communities without consultation or input. Of course people will dislike that and reject it."

Harris doesn't disagree. "...I think the primary reason is that editors are rightly concerned about impacts to their workflows, and the Foundation has been historically terrible about thinking about this and accounting for it," he said. "This is why I designed the New Pages Feed to work independently of the existing workflows and scripts that people had developed themselves."

Recognition

Brion Vibber celebrating his day in 2013.

Early on, Wales recognized key development milestones by giving developers their own holidays:

Brion Vibber Day
(June 1).

"It isn't really clear who gets the credit now – whenever you step away, very few people remember what you did," Mituzas said. "Being recognized and rewarded by the community was definitely part of the motivation to keep on working."

Mituzas himself is remembered on the

blame wheel
, where he's responsible for 25% of Wikipedia's problems. "Sometimes it feels that the blame wheel is the only part that is left of any fame I had," he said.

Harris is likely the best known Wikimedia developer, having appeared on fundraising banners in 2011.

"We had three 'storytellers' who interviewed a lot of us about why we were working there and they liked what I had to say. They took photos," he said. "Later one of them ended up being used as a test and performed fairly well. This became popular and weird because the internet is weird."

Unsurprisingly, there's a direct parallel to how credit operates on Wikipedia itself.

"It's always fascinated me how much wiki editing has mirrored open source software contributions," McBride said. "In both, a lot of people making small suggestions and improvements are the ones who push the project forward."

Some of those names can be found on Special:Version or in the credits. Others might be in mailing list archives, forgotten bugs and long lost IRC logs, but their contributions nonetheless built Wikipedia into what it is today.



Reader comments

2021-01-31

Celebrating 20 years

Most Wikipedians do not use video or audio very often or very well in articles, relying almost exclusively on text and photos. Perhaps the COVID pandemic has improved editors' skills in this area with all the Zoom calls we've had to make. The pandemic has also increased our need for multimedia as in-person conferences, such as Wikimania 2020, have been cancelled. This week Wikimania 2021 was turned into a virtual event (see this message from Janeen Uzzell on the Wikipedia-l mailing list).

The video and audio resources created to celebrate Wikipedia's 20th birthday demonstrate the flexibility of these media. The videos include a slick two-minute celebratory video, an hour-long video press conference from the WMF, and even longer deep dives produced by the community.

Videos

Happy 20th birthday, Wikipedia! from the Wikimedia Foundation.
  • Happy 20th birthday, Wikipedia! celebrates 20 years of the Wikipedia movement. Also available on YouTube, Vimeo, and the Wikimedia Foundation. (time 2:17)
  • Wikipedia 20 Press Call The press talks to Jimmy Wales, Katherine Maher, Dr. Netha Hussain, and Wikipedia editor Risker. Moderated by Anusha Alikhan. Includes transcript. (time 1:15:00)
Wikimedia New York City celebrates 20 years of Wikipedia.

Audio

  • Economist asks Jimmy Wales "Is social media still a force for good?" Wales covers several variants of this question, but part of his answer is that Wikipedia serves our readers, not our advertisers. Everybody knows what an encyclopedia is and the community has stayed faithful to the project's original goals. (time 27:15)
  • One Page At A Time, Jess Wade Is Changing Wikipedia: Jess Wade on NPR about how she has changed Wikipedia by writing articles on scientists from under-represented groups such as Sarah Gilbert, Kizzmekia Corbett, and Gladys West. (time 13:29)
  • The World According to Wikipedia, the new Irish podcast has now completed 5 episodes. In this GLAM/Birthday episode they interview a digital docent based in Washington, D.C. who works with the Smithsonian Institution. (time 26:56)
  • "Wikipedia At 20: The Promises And Pitfalls Of The 'Free Encyclopedia' ":
    Connecticut Public Radio
    interviews Wikipedians Molly White, Sherry Antoine, and Ariella Rotramel to hear an insiders' view of the Wikipedia editing process, including how Wikipedians have developed tools to fight misinformation, and how they are working to address "persistent gender, racial, and geographic disparities". (time 49:00)




Reader comments

2021-01-31

Wikipedia celebrates 20 years of free, trusted information for the world

20.wikipedia.org launches to commemorate Wikipedia's birthday, look back at the past 20 years, and meet the humans behind the platform.
This article was originally published by the Wikimedia Foundation on January 14, 2021 to mark Wikipedia's 20th birthday, at Wikimedia Foundation News. It is licensed CC-BY SA 3.0.

Wikipedia, the world's largest free online encyclopedia, turned 20 years old on 15 January. This birthday commemorates two decades of global efforts to support free knowledge, open collaboration, and trust on the internet. In a time when disinformation and polarization challenge our trust in information and institutions, Wikipedia is more relevant than ever. Wikipedia celebrates its past and looks ahead to how it will meet the challenges of tomorrow to grow into a more resilient, equitable knowledge resource.

"Wikipedia has evolved from a seemingly impossible idea into a sprawling testament to humanity—a place where we can collaborate, share, and learn about anything and everything," said Katherine Maher, CEO of the Wikimedia Foundation. "It is a global effort of volunteers who elevate knowledge and cooperation over conflict and self-interest. We are committed to preserving the integrity and value of information at a time when the world needs it most."

Today, Wikipedia's more than 55 million articles can be accessed in over 300 languages, for free, and without advertisements, all created by volunteers. During the COVID-19 pandemic, Wikipedia saw record-breaking increases in daily traffic with a growing community of medical professionals contributing knowledge about COVID-19 to the platform. Studies have shown that Wikipedia is one of the most-viewed sources for health information, and its role in providing trusted access to information about COVID-19 in the pandemic led to a milestone collaboration with the World Health Organization in October 2020.

"In a world where information is increasingly commoditized, Wikipedia's model has always been based on the belief that knowledge belongs to all humans," said Wikipedia founder Jimmy Wales. "Wikipedia forces us to step outside of our echo chambers and contend with what a shared understanding of the world could really look like. After all, a Wikipedia article is the same no matter who or where you are in the world, and if something isn't right in the article, you can change it."

An Art+Feminism edit-a-thon at the Royal Armoury in Stockholm, 2017

Wikipedia aspires to reflect the sum of all human knowledge. As it celebrates 20 years and looks to the future, the people who contribute to Wikipedia also acknowledge that there are deep knowledge gaps that must be addressed. Initiatives such as WikiProject Women in Red, WikiGap, and AfroCROWD, among many others, aim to ensure that Wikipedia more fully represents the diversity of the world. The Wikimedia movement's 2030 strategic direction also reflects this commitment. In addition, the Foundation recently developed a $4.5 million Equity Fund that will offer grants to advance more equitable, inclusive representation in Wikimedia projects, including Wikipedia.

The Wikimedia Foundation and Wikipedia's volunteer communities are actively evolving to face the challenges of our time, including by:

  • Investing in the overall health and diversity of volunteer communities to be more inclusive and globally representative
  • Developing tools and programs to expand the quantity and diversity of Wikipedia content
  • Modernizing Wikipedia's technology to meet the needs of the future
  • Advocating for policies that preserve user privacy, free expression, and the free exchange of knowledge online
  • Enhancing volunteers' ability to combat disinformation with better resources and insights

The Wikimedia Foundation held a virtual event, hosted by Katherine Maher, Wikimedia Foundation CEO, and Wikipedia founder Jimmy Wales, to showcase the contributions of Wikipedia's global volunteer communities. The event kicked-off a year-long celebration themed "20 Years Human," a nod to the humans who make Wikipedia possible—including regular contributors, donors, and readers. Viewers can tune in via the livestream link.

To learn more about Wikipedia's growth over the last two decades visit 20.wikipedia.org.

To attend Wikipedia virtual celebrations hosted by volunteers around the world, please visit the events page.

Share your favorite Wikipedia story on social media using #Wikipedia20, and tag @Wikipedia in your posts.

To make a donation and continue to sustain Wikipedia's future, visit donate.wikimedia.org.

Wikipedia facts and figures

  • Wikipedia has more than 55 million articles across over 300 languages.
  • Wikipedia is edited by more than 280,000 volunteer editors every month around the world.
  • Wikipedia is edited 350 times per minute and read more than 8000 times a second.
  • Wikipedia is accessed by 1.5 billion unique devices every month and read more than 15 billion times every month.
  • Roughly 89% of articles on Wikipedia are in languages other than English.
  • Most vandalism (edits that do not meet Wikipedia's reliability and neutrality standards) is addressed within five minutes on Wikipedia.
  • Wikipedia is supported by nearly 7 million donors, with the average donation being about $15 USD.

About Wikipedia

Wikipedia is the world's free knowledge resource. It is a collaborative creation that has been added to and edited by millions of people from around the globe since it was created in 2001: everyone can edit it, at any time. Wikipedia is offered in more than 300 languages containing a total of more than 55 million articles. It is the largest, collaborative collection of free knowledge in human history, and today its content is contributed and edited by a community of more than 280,000 volunteer editors each month.

About the Wikimedia Foundation

The Wikimedia Foundation is the nonprofit organization that operates Wikipedia and the other Wikimedia free knowledge projects. Our vision is a world in which every single human can freely share in the sum of all knowledge. We believe that everyone has the potential to contribute something to our shared knowledge, and that everyone should be able to access that knowledge freely. We host Wikipedia and the Wikimedia projects, build software experiences for reading, contributing, and sharing Wikimedia content, support the volunteer communities and partners who make Wikimedia possible, and advocate for policies that enable Wikimedia and free knowledge to thrive.

The Wikimedia Foundation is a charitable, not-for-profit organization that relies on donations. We receive donations from millions of individuals around the world, with an average donation of about $15. We also receive donations through institutional grants and gifts. The Wikimedia Foundation is a United States 501(c)(3) tax-exempt organization with offices in San Francisco, California, USA.




Reader comments

2021-01-31

Students still have a better opinion of Wikipedia than teachers do


A monthly overview of recent academic research about Wikipedia and other Wikimedia projects, also published as the Wikimedia Research Newsletter.


Wikipedia as OER: the “Learning with Wikipedia” project

Reviewed by Matthew Sumpter

An article

Open Education (OE). The goal was to determine the effectiveness of learning subject-specific content with communal forms of assessment and to stimulate digital competencies involved in information literacy, while also creating open content
and encouraging the spirit of OE.

The project was divided into planning and teaching phases. The planning phase required that instructors were trained on Wikipedia philosophies and rules, and activities were designed for each course, while also setting up test environments for students to familiarize themselves with. Several potential student activities were designed, including: translation of an English Wikipedia article to Italian, elaboration on course content, validation of Wikipedia content when compared to other sources, as well as new article creation regarding course content. Students attended a workshop that introduced them to the goals and expectations of the project before beginning, and then were continuously assisted throughout with a focus on developing skills related to OE. Specifically, these included skills in "finding, selecting and evaluating information, digital citizenship actions, applying guidelines for online etiquette, creating and developing digital content, becoming familiar with copyright issues and the use of Creative Commons licenses, [and] protecting personal data". In this way, the project served the joint tasks of teaching course content and digital OE skills.

The article analyzes the outcomes of this project using a questionnaire provided to participants to answer two research questions: 1) How do teachers and students perceive Wikipedia? and 2) What digital competences do teachers and students believe have been improved by creating a Wikipedia article in the project? The authors found that teachers and students have contrasting opinions regarding Wikipedia - teachers exhibited a poor opinion while students on average had a good opinion. The authors point to this finding as an important indication that teachers must be involved in the planning and implementation of educational Wikipedia projects, in order to prevent negative bias towards the medium. In regards to the digital competences, students expressed they had learned Wikipedia's rules, how to browse, search, create, and manipulate digital content. They also indicated that "evaluating data, information and digital content", an essential goal, was not sufficiently covered. This was again in contrast to the teachers' answers, of which 100% indicated that this task was well stimulated. The authors suspect this is because many bibliographic sources used for writing the articles were provided to the students by the instructors. This becomes one of the authors' main conclusions when considering future work - namely that students should be provided more freedom in finding primary sources. Additionally, they cite the need for carefully managed learning strategies to ensure the work is educational while following Wikipedia's rules, so that student articles do not risk deletion.


Briefly


Other recent publications

Other recent publications that could not be covered in time for this issue include the items listed below. Contributions, whether reviewing or summarizing newly published research, are always welcome.

Compiled by Tilman Bayer


Wikipedia is "largely successful" at removing harmful content

From the abstract:[2]

"... we aim to assess the degree to which English-language Wikipedia is successful in addressing harmful speech with a particular focus on the removal of deleterious content. We have conducted qualitative interviews with Wikipedians and carried out a text analysis using machine learning classifiers trained to identify several variations of problematic speech. Overall, we conclude that Wikipedia is largely successful at identifying and quickly removing a vast majority of harmful content despite the large scale of the project. The evidence suggests that efforts to remove malicious content are faster and more effective on Wikipedia articles compared to removal efforts on article talk and user talk pages."

See also earlier coverage of the researchers' presentation at Wikimania 2019, and the research project's page on Meta-wiki


Wiki Loves Africa Logo

Recruiting previous participants of the "Wiki Loves Africa" photo contest

From the abstract of a preprint titled "Broadening African Self-Representation on Wikipedia: A Field Experiment":[3]

"Wiki Loves Africa (WLA) is a Wikipedia community project that hosts an annual photo competition focused on increasing African contributions to imagery that represents African people, places, and culture. In a field experiment with [5,905 previous] participants, we randomly assigned past contributors to receive a recruitment message and observed their contributions to the 2020 competition. On average, receiving a message caused a 1.2 percentage point increase in competition entrants (p=0.002). Among those who contributed to the contest in 2019, there was an increase of 2.7 percentage points compared to the control group (p=0.029)."


"WiTPy: A Toolkit to Parse and Analyse Wikipedia Talk Pages"

From the abstract:[4]

"In this article, we propose an opensource toolkit to extract, parse, and analyze the Wikipedia talk pages. [...] User-friendly and high-level analysis methods are created on the top of NoSQL database, which can be used to understand the collaboration dynamics on article talk pages."


"Modelling User Behavior Dynamics with Embeddings"

From the abstract:[5]

"... we present a user behavior model built using behavior embeddings to compare behaviors and their change over time. To this end, we first define the formal model and train the model using both action (e.g., copy/paste) embeddings and user interaction feature (e.g., length of the copied text) embeddings. Having obtained vector representations of user behaviors, we then define three measurements to model behavior dynamics over time, namely: behavior position, displacement, and velocity. To evaluate the proposed methodology, we use three real world datasets [ ... including] (iii) thousands of editors completing unstructured editing tasks on Wikidata. Through these datasets, we show that the proposed methodology can: (i) surface behavioral differences among users; (ii) recognize relative behavioral changes; and (iii) discover directional deviations of user behaviors."


"Deriving Geolocations in Wikipedia"

From the abstract:[6]

"We study the problem of deriving geolocations for Wikipedia pages. To this end, we introduce a general four-step process to location derivation, and consider different instantiations of this process, leveraging both textual and categorical data. [...] our system can be used to augment the geographic information of Wikipedia, and to enable more effective geographic information retrieval."


"Testing the validity of Wikipedia categories for subject matter labelling of open-domain corpus data"

From the abstract:[7]

"In this article, a hierarchical taxonomy of three-level depth is extracted from the Wikipedia category system. The resulting taxonomy is explored as a lightweight alternative to expert-created knowledge organisation systems (e.g. library classification systems) for the manual labelling of open-domain text corpora. Combining quantitative and qualitative data from a crowd-based text labelling study, the validity of the taxonomy is tested and the results quantified in terms of interrater agreement. While the usefulness of the Wikipedia category system for automatic document indexing is documented in the pertinent literature, our results suggest that at least the taxonomy we derived from it is not a valid instrument for manual subject matter labelling of open-domain text corpora."


"Wikipedia, The Free Online Medical Encyclopedia Anyone Can Plagiarize: Time to Address Wiki-Plagiarism"

From the abstract and paper:[8]

"... plagiarism of Wikipedia in peer-reviewed publications has received little attention. Here, I present five cases of PubMed-indexed articles containing Wiki-plagiarism, i.e. copying of Wikipedia content into medical publications without proper citation of the source. [...] ... I subsequently contacted the authors of the three other Wiki-plagiarizing papers, as well as the publishers and Editors of the journals involved, to ask for an explanation, correction or retraction. None of them replied, despite the fact that these journals are members of the Committee on Publication Ethics (COPE). Of note, the article on exome sequencing was edited by the same author as the 2010 published paper. ... "


"Hierarchical Trivia Fact Extraction from Wikipedia Articles"

From the abstract:[9]

"In this paper, we propose a new unsupervised algorithm that automatically mines trivia facts for a given entity. Unlike previous studies, the proposed algorithm targets at a single Wikipedia article and leverages its hierarchical structure via top-down processing. [...] Experimental results demonstrate that the proposed algorithm is over 100 times faster than the existing method which considers Wikipedia categories. Human evaluation demonstrates that the proposed algorithm can mine better trivia facts regardless of the target entity domain and outperforms the existing methods."


"Multilingual Contextual Affective Analysis of LGBT People Portrayals in Wikipedia"

From the abstract:[10]

"Prior work has examined descriptions of people in English using contextual affective analysis, a natural language processing (NLP) technique that seeks to analyze how people are portrayed along dimensions of power, agency, and sentiment. Our work presents an extension of this methodology to multilingual settings [...] We then demonstrate the usefulness of our method by analyzing Wikipedia biography pages of members of the LGBT community across three languages: English, Russian, and Spanish. Our results show systematic differences in how the LGBT community is portrayed across languages, surfacing cultural differences in narratives and signs of social biases. Practically, this model can be used to surface Wikipedia articles for further manual analysis---articles that might contain content gaps or an imbalanced representation of particular social groups."

(The underlying article dataset has also been used by another researcher to create this interactive visualization.)


References

  1. ISSN 1971-8829
    .
  2. ^ Clark, Justin; Faris, Robert; Gasser, Urs; Holland, Adam; Ross, Hilary; Tilton, Casey (2019-11-01). "Content and Conduct: How English Wikipedia Moderates Harmful Speech".
    SSRN 3489176
    .
  3. ^ J. Nathan Matias; Florence Devouard; Julia Kamin; Max Klein; Eric Pennington (2020-08-06). "Broadening African Self-Representation on Wikipedia: A Field Experiment". OSF.io. {{cite web}}: Unknown parameter |lay-url= ignored (help)
  4. ^ Verma, Amit Arjun; Iyengar, S.R.S; Gandhi, Nitin (2020-08-01). "WiTPy: A Toolkit to Parse and Analyse Wikipedia Talk Pages". Proceedings of the ACM/IEEE Joint Conference on Digital Libraries in 2020. JCDL '20. New York, NY, USA: Association for Computing Machinery. pp. 535–536.
  5. .
  6. ^ Kwon, Jingun; Kamigaito, Hidetaka; Song, Young-In; Okumura, Manabu (December 2020). "Hierarchical Trivia Fact Extraction from Wikipedia Articles". Proceedings of the 28th International Conference on Computational Linguistics. COLING 2020. Barcelona, Spain (Online): International Committee on Computational Linguistics. pp. 4825–4834.
  7. ^ Park, Chan Young; Yan, Xinru; Field, Anjalie; Tsvetkov, Yulia (2020-10-21). "Multilingual Contextual Affective Analysis of LGBT People Portrayals in Wikipedia". ].



Reader comments

2021-01-31

Dr. Seuss's Guide to Wikipedia

One hundred years ago, Theodor Seuss Geisel (better known as "Dr. Seuss") arguably began his writing career by authoring a book review of a railroad timetable and contributing to the Dartmouth College humor magazine Jack-O-Lantern.[1] Since then millions of children – young and old – have delighted in the wit and wisdom of his iconic books with their hallucinatory illustrations. In celebration of this centennial, The Signpost has blatantly hijacked some of Seuss's most familiar works (and a few less familiar ones) in service of the perhaps questionable goal of indoctrinating a new generation of editors in the wacky ways of Wikipedia.
fished
.
"It's a pretty good zoo",
said young Gerald McGrew,
"And the fellow who runs it
seems proud of it, too."
– Dr. Seuss, If I Ran the Zoo
Dr. Seuss's ABCs of Editing[2]

All around the world
they're editing.
And so they

don't need me.
Let others edit Switzerland
or Memphis, Tennessee.
Let them edit in Alaska
and in China.
I don't care.
Let others edit Italy.
Let others edit Spain.
Let them edit Massachusetts,
Connecticut and Maine.
Someone else can edit London
and Paris and Berlin.
Let them edit all they want to.
But not me.
I'm sleeping in.[9]

The Bitter Battleground Book

My wonderful weapon, the

Rollbacker-Snatchem,
will roll back those edits as quick as we catch 'em.[11]

You have brains in your head.
You have feet in your shoes.
You can

redirect articles
any places you choose.[12]

Ei! What a committee! Fit for its purpose!
Our clerks they love clerking. They say, "Work us! Please work us!
We'll clerk and we'll clerk until we're lightheaded
(But you'll only see half 'cause discussion's unthreaded!)" [13]

The Zax's Conduct Guidance

I Had Trouble Reporting to WP:ANEW

I learned there are

I Had Trouble Reporting to AIV, Too
)

I'll just have to

fix it. Because, after all
An error's an error, no matter how small.[25]

Oh, the Links You Can Link!

Link

safe linking whenever you can! [27]

Editors know, up top you see great sights,
but down here at the bottom, we too should have rights.[36]

Wardrobe Advice from
The Cat in the Hat

Hop on POP3 *

FATHER MOTHER SISTER BROTHER
That wasn't me,

POP3
.


A Little Dab'll Do Ya!
Too Many Dabs[46]

Too Many Debs



Aides-mémoire

Thanks Doc, we couldn't have done it without you!
Bonus video
video icon
I do not like that single-payer,
I do not want it for my Bayer.
I will not use
Obamacare
,
I will not use it anywhere.
I will not use it for a sprain,
I will not use it in the brain.
I will not use it for my gout,
I will not use it on my snout.
I will not use it for my head,
I will not use it when I've bled.
I will not use it here or there,
I will not use it anywhere! [47]
Bonus audio
Ön Beyond Zebra!

     "I'm telling you this
           'cause you're one of my friends.
     My alphabet starts
           where your alphabet ends! ...
     So, on beyond Z!
           It's high time you were shown
     That you really don't know
           All there is to be known."
— Dr. Seuss, On Beyond Zebra!
  1. ^ Lathem, Edward Connery, ed. (2004) [1976]. "The Beginnings of Dr. Seuss: An Informal Reminiscence". Dartmouth College. Retrieved 2021-01-22.
  2. ^ Dr. Seuss's ABC
  3. ^ One Fish, Two Fish, Red Fish, Blue Fish
  4. ^ I Can Lick 30 Tigers Today! and Other Stories
  5. ^ I Can Read with My Eyes Shut!
  6. ^ Come over to My House
  7. ^ Hooray for Diffendoofer Day!
  8. ^ The Bippolo Seed and Other Lost Stories
  9. ^ I Am Not Going to Get Up Today!: "All around the world / they're getting up. / And that's okay with me. / Let the kids get up in Switzerland / ... or Memphis, Tennessee. / Let the kids get up in Alaska / ... and in China. / I don't care. / Let the kids get up in Italy. / Let the kids get up in Spain. / Let them get up in Massachusetts / and Connecticut and Maine. / Let the kids get up in London / and in Paris and Berlin. / Let them get up all they want to. / But not me. / I'm sleeping in."
  10. ^ Marvin K. Mooney Will You Please Go Now!: "Just go. Go. GO! / I don't care how. / You can go by foot. / You can go by cow."
  11. ^ The Butter Battle Book: ""My wonderful weapon, the Jigger-Rock Snatchem, / will fling 'em right back just as quick as we catch 'em."
  12. ^ Oh, the Places You'll Go!: "You have brains in your head. / You have feet in your shoes. / You can steer yourself any direction you choose."
  13. ^ If I Ran the Circus: "Ei! Ei! What a circus! My Circus McGurkus! / My workers love work. They say, "Work us! Please work us! / We'll work and we'll work up so many surprises / You'd never see half if you had forty eyeses!"
  14. ^ You're Only Old Once!
  15. ^ Oh Say Can You Say?
  16. ^ Great Day for Up!
  17. ^ Did I Ever Tell You How Lucky You Are?
  18. ^ How the Grinch Stole Christmas!
  19. ^ I Had Trouble in Getting to Solla Sollew: "And I learned there are troubles / Of more than one kind. / Some come from ahead / And some come from behind. / But I've bought a big bat. / I'm all ready, you see. / Now my troubles are going / To have troubles with me!"
  20. ^ And to Think That I Saw It on Mulberry Street
  21. ^ The Foot Book
  22. ^ Green Eggs and Ham
  23. ^ Horton Hatches the Egg
  24. ^ My Book about ME
  25. ^ Horton Hears a Who!: "I'll just have to save him. Because, after all, / A person's a person, no matter how small."
  26. ^ Horton Hatches the Egg: "I meant what I said, and I said what I meant. / An elephant's faithful, one hundred per cent!"
  27. ^ Oh, the Thinks You Can Think!: "Think left and think right / and think low and think high. / Oh, the thinks you can think up if only you try!"
  28. ^ Daisy-Head Mayzie
  29. ^ Horton and the Kwuggerbug and More Lost Stories
  30. ^ What Pet Should I Get?
  31. ^ Thidwick the Big-Hearted Moose
  32. ^ The King's Stilts
  33. ^ Bartholomew and the Oobleck
  34. ^ The Shape of Me and Other Stuff
  35. ^ If I Ran the Zoo: "In the Far Western part / Of south-east North Dakota / Lives a very fine animal / Called the Iota. / But I'll capture one / Who is even much finer / In the north-eastern west part / Of South Carolina."
  36. ^ Yertle the Turtle and Other Stories: "I know, up on top you are seeing great sights, / But down here at the bottom, we, too, should have rights."
  37. ^ Horton Hears a Who!: "Are you sure every Who down in Who-ville is working? / Quick! Look through your town! Is there anyone shirking?"
  38. ^ Oh, the Places You'll Go!: "I'm sorry to say so / but, sadly, it's true / that Bang-ups / and Hang-ups / can happen to you."
  39. ^ There's a Wocket in My Pocket
  40. ^ Fox in Socks
  41. ^ The 5,000 Fingers of Dr. T.
  42. ^ Dr. Seuss's Sleep Book
  43. ^ The 500 Hats of Bartholomew Cubbins
  44. ^ The Cat in the Hat
  45. ^ Hop on Pop: "FATHER / MOTHER / SISTER / BROTHER / That one is my other brother. / My brothers read a little bit. / Little words like / If and it."
  46. Too Many Daves
  47. ^ Green Eggs and Ham: "I will not eat them in the rain. / I will not eat them on a train. / Not in the dark! Not in a tree! / Not in a car! You let me be! / I do not like them in a box. / I do not like them with a fox. / I will not eat them in a house. / I do not like them with a mouse. / I do not like them here or there. / I do not like them ANYWHERE! / I do not like green eggs and ham! / I do not like them, / Sam-I-am."