Content similarity detection

This approach aims to recognize changes in the unique writing style of an author as an indicator for potential plagiarism.

Similarities and writing style features are computed with the help of predefined document models and might represent false positives.

[citation needed] Fingerprinting is currently the most widely applied approach to content similarity detection.

Minutiae matching with those of other documents indicate shared text segments and suggest potential plagiarism if they exceed a chosen similarity threshold.

Nonetheless, substring matching remains computationally expensive, which makes it a non-viable solution for checking large collections of documents.

[20][21][22] Bag of words analysis represents the adoption of vector space retrieval, a traditional IR concept, to the domain of content similarity detection.

[27][29][30][31] Stylometry subsumes statistical methods for quantifying an author's unique writing style[32][33] and is mainly used for authorship attribution or intrinsic plagiarism detection.

[35] More recent approaches to assess content similarity using neural networks have achieved significantly greater accuracy, but come at great computational cost.

Comparative evaluations of content similarity detection systems[6][39][40][41][42][43] indicate that their performance depends on the type of plagiarism present (see figure).

In particular, substring matching procedures achieve good performance for copy and paste plagiarism, since they commonly use lossless document models, such as suffix trees.

The performance of systems using fingerprinting or bag of words analysis in detecting copies depends on the information loss incurred by the document model used.

By applying flexible chunking and selection strategies, they are better capable of detecting moderate forms of disguised plagiarism when compared to substring matching procedures.

Stylometric comparisons are likely to fail in cases where segments are strongly paraphrased to the point where they more closely resemble the personal writing style of the plagiarist or if a text was compiled by multiple authors.

The results of the International Competitions on Plagiarism Detection held in 2009, 2010 and 2011,[6][42][43] as well as experiments performed by Stein,[34] indicate that stylometric analysis seems to work reliably only for document lengths of several thousand or tens of thousands of words, which limits the applicability of the method to computer-assisted plagiarism detection settings.

[45] The design of content similarity detection software for use with text documents is characterized by a number of factors:[46] Most large-scale plagiarism detection systems use large, internal databases (in addition to other resources) that grow with each additional document submitted for analysis.

[citation needed] Plagiarism in computer source code is also frequent, and requires different tools than those used for text comparisons in document.

It does not pick up poorly paraphrased work, for example, or the practice of plagiarizing by use of sufficient word substitutions to elude detection software, which is known as rogeting.

[58] Another complication with TMS is its tendency to flag much more content than necessary, including legitimate citations and paraphrasing, making it difficult to find real cases of plagiarism.

[57] This issue arises because TMS algorithms mainly look at surface-level text similarities without considering the context of the writing.

Classification of computer-assisted plagiarism detection methods [ 16 ]
Detection performance of computer-assisted plagiarism detection approaches depending on the type of plagiarism being present