Carnegie Mellon University
Browse
Sharif_cmu_0041E_10480.pdf (2.27 MB)

Practical Inference-Time Attacks Against Machine-Learning Systems and a Defense Against Them

Download (2.27 MB)
thesis
posted on 2020-01-22, 19:28 authored by Mahmood SharifMahmood Sharif
Prior work has shown that machine-learning algorithms are vulnerable to evasion by socalled adversarial examples. Nonetheless, the majority of the work on evasion attacks
has mainly explored Lp-bounded perturbations that lead to misclassification. From a computer-security perspective, such attacks have limited practical implications. To fill
the gap, we propose evasion attacks that satisfy multiple objectives, and show that these attacks pose a practical threat to computer systems. In particular, we demonstrate how to produce adversarial examples against state-of-the-art face-recognition and malwaredetection systems that simultaneously satisfy multiple objectives (e.g., smoothness and robustness against changes in imaging conditions) to mislead the systems in practical settings. Against face recognition, we develop a systematic method to automatically generate attacks, which are realized through printing a pair of eyeglass frames. When worn by attackers, the eyeglasses allow them mislead face-recognition algorithms to evade recognition or impersonate other individuals. Against malware detection, we develop an attack that guides binary-diversification tools via optimization to transform binaries in a functionality preserving manner and mislead detection. The attacks that we initially demonstrate achieve the desired objectives via ad hoc optimizations. We extend these attacks via a general framework to train a generator neural network to emit adversarial examples satisfying desired objectives. We demonstrate
the ability of the proposed framework to accommodate a wide range of objectives, including imprecise ones difficult to model, in two application domains. Specifically,
we demonstrate how to produce adversarial eyeglass frames to mislead face recognition with better robustness, inconspicuousness, and scalability than previous approaches, as well as a new attack to fool a handwritten-digit classifier. Finally, to protect computer-systems from adversarial examples, we propose n-ML—a novel defense that is inspired by n-version programming. n-ML trains an ensemble of n classifiers, and classifies inputs by a vote. Unlike prior approaches, however, the classifiers are trained to classify adversarial examples differently than each other,
rendering it very difficult for an adversarial example to obtain enough votes to be misclassified. In several application domains (including face and street-sign recognition),
we show that n-ML roughly retains the benign classification accuracies of state-of-theart models, while simultaneously defending against adversarial examples (produced
by our framework, or Lp-based attacks) with better resilience than the best defenses known to date and, in most cases, with lower inference-time overhead.

History

Date

2019-11-17

Degree Type

  • Dissertation

Department

  • Electrical and Computer Engineering

Degree Name

  • Doctor of Philosophy (PhD)

Advisor(s)

Lujo Bauer Nicolas Christin

Usage metrics

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC