A quantitative comparison of program plagiarism detection tools

Daniël Heres, Jurriaan Hage

    Research output: Chapter in Book/Report/Conference proceedingConference contributionAcademicpeer-review

    Abstract

    In this work we compare a total of 9 different tools for the detection of source code plagiarism. We evaluated the plagiarism or copy detection tools CPD, JPlag, Sherlock, Marble, Moss, Plaggie and SIM and two baselines, one based on the Unix tool diff and one based on the difflib module from the Python Standard Library. We provide visualizations of the output of these tools and compare the performance of each tool when running it on different tasks by comparing both the F-measures and the area under the precision-recall-curve (AUC-PR). We compare the performance using these metrics on each task and identify the best performing tools.

    Original languageEnglish
    Title of host publicationCSERC '17: Proceedings of the 6th Computer Science Education Research Conference
    EditorsVreda Pieterse, Marko van Eekelen, Michalis Giannakos
    PublisherAssociation for Computing Machinery
    Pages73-82
    Number of pages10
    ISBN (Electronic)9781450363389
    DOIs
    Publication statusPublished - 14 Nov 2017
    Event6th Computer Science Education Research Conference, CSERC 2017 - Helsinki, Finland
    Duration: 13 Nov 201714 Nov 2017

    Publication series

    NameACM International Conference Proceeding Series

    Conference

    Conference6th Computer Science Education Research Conference, CSERC 2017
    Country/TerritoryFinland
    CityHelsinki
    Period13/11/1714/11/17

    Bibliographical note

    Publisher Copyright:
    © 2017 Association for Computing Machinery.

    Keywords

    • Empirical study
    • Program plagiarism detection
    • Quantitative comparison
    • Tools

    Fingerprint

    Dive into the research topics of 'A quantitative comparison of program plagiarism detection tools'. Together they form a unique fingerprint.

    Cite this