Adversarial examples are test images which have been perturbed slightly to cause misclassification. As these adversarial examples are usually unproblematic for us humans, but are able to easily fool deep neural networks, their discovery has sparked quite some interest in the deep learning and privacy/security communities. In this article, I want to provide a rough overview of the topic including a brief survey of relevant literature and some ideas on future research directions.
Last week, I attended my very first CVPR in Salt Lake City, where I also presented my work on weakly-supervised 3D shape completion. In the course of the week, I attended several tutorials as well as all oral and poster sessions. In this article, I want to share my notes and some general comments.