Seeing with Machines: Decipherability and Obfuscation in Adversarial Images

Rosemary Lee

    Research output: Conference Article in Proceeding or Book/Report chapterArticle in proceedingsResearchpeer-review

    Abstract

    Adversarial images, inputs designed to produce errors in ma-chine learning systems, are a common way for researchers to test the ability of algorithms to perform tasks such as image classification. "Fooling images" are a common kind of adversari-al image, causing miscategorisation errors which can then be used to diagnose problems within an image classification algo-rithm. Situations where human and computer categorise an image differently, which arise from adversarial images, reveal discrepancies between human image interpretation and that of computers. In this paper, aspects of state of the art machine learning research and relevant artistic projects touching on adversarial image approaches will be contextualised in reference to current theories. Harun Farocki's concept of the operative image will be used as a model for understanding the coded and procedural nature of automated image interpretation. Through comparison of current adversarial image methodolo-gies, this paper will consider what this kind of image production reveals about the differences between human and computer visual interpretation.
    Original languageEnglish
    Title of host publicationProceedings of the 24th International Symposium on Electronic Art : ISEA2018
    EditorsRufus Adebayo, Ismail Farouk, Steve Jones, Maleshoane Rapeane- Mathonsi
    Number of pages4
    Place of PublicationDurban, South Africa
    PublisherDurban University of technology (DUT)
    Publication date24 Jun 2018
    Pages321-324
    ISBN (Electronic)978-0-620-80332-8
    Publication statusPublished - 24 Jun 2018

    Keywords

    • Adversarial Images
    • Image Classification
    • Machine Learning
    • Human-Computer Interaction
    • Operative Image

    Fingerprint

    Dive into the research topics of 'Seeing with Machines: Decipherability and Obfuscation in Adversarial Images'. Together they form a unique fingerprint.

    Cite this