Manuscript with arrow icon Book and magnifying glass icon Cross-check icon Process checklist icon Reputation ribbon icon Graduation cap icon Question speech bubble icon Headset call icon Mobile phone call icon Login arrow icon B+ Paper Icon Feedback Speech Bubble Icon Similarity Check Icon Professional Development Icon Admin Training Icon Instructor Training Icon Student Training Icon Integrations Icon System Status Icon System Requirements Icon Menu Icon Checkmark Icon Download Icon Rubric Icon Prompt Icon QuickMark Set Icon Lesson Plan Icon Success Story Icon Infographic Icon White Paper Icon White Paper Icon Press Release Icon News Story Icon Event Icon Webcast Icon Video Icon Envelope Icon Plaque Icon Lightbulb Icon Training Icon Search Icon Turnitin Logo (Text and Icon) Icon Facebook Icon Twitter Icon LinkedIn Icon Google Plus Icon Lightbulb Icon Binoculars Icon Drama Masks Icon Magnifying Glass Icon Signal Check Indicator Bars Red Flag Icon Analysis and Organization Icon
Contact Sales

When you’re one of the largest content-based websites on the internet, how do you ensure the accuracy of information as well as prevent potential instances of plagiarism? How do you preserve an open editing model with 350 edits coming in per minute?

This was Wikipedia’s challenge.

Wikipedia is admittedly unique, with an open editing model and status as one of the largest websites on the internet, with an average of 15 billion page views per month. But it is a website with content that impacts companies, educational institutions, and publishers alike—because Wikipedia is all three of these things for its users.

English Wikipedia contains more than 5 million articles and is edited constantly by thousands of volunteers. Historically, volunteers used another API to review articles for potential issues of plagiarism, but when that API was pulled in 2014, English Wikipedia volunteers decided to take inventory of both short-term and long-term needs in searching for a new solution.

Volunteer editors were handling instances of potential plagiarism issues individually but suspected more were going undetected. Wikipedia’s volunteers also wanted to detect instances of plagiarism as quickly as possible, given that Wikipedia is so extensively mirrored on the internet. These were immediate needs that needed to be addressed.

Long-term, they wanted a more automated solution to free up time so volunteers could work on other projects.

Turnitin and Wikipedia volunteers worked together to find such a solution. Volunteer editors routinely review, edit, and detect inappropriate content in over 2,000 edits per month using a plagiarism detection bot called CopyPatrol powered by iThenticate and shepherded by Eran Rosenthal and User:Diannaa. This has helped to ensure the reusability and quality of Wikipedia content.

It is always gratifying to speak to customers and to learn who they are, and how they use our products. And it is all the more edifying when we have such an interesting use case as Wikipedia’s and their position at the crossroads of information and research.

To learn more about our work with Wikipedia, read the case study here.