pith. machine review for the scientific record. sign in

arxiv: 2605.01098 · v1 · submitted 2026-05-01 · 💻 cs.LG · cs.CV

Recognition: unknown

Almost for Free: Crafting Adversarial Examples with Convolutional Image Filters

Authors on Pith no claims yet

Pith reviewed 2026-05-09 18:56 UTC · model grok-4.3

classification 💻 cs.LG cs.CV
keywords adversarial examplesconvolutional filtersedge detectiontransferable attacksneural network vulnerabilityefficient attacksmachine learning security
0
0 comments X

The pith

Optimized 3x3 convolutional filters based on edge detection can generate transferable adversarial examples using five orders of magnitude fewer parameters than generative models.

A machine-rendered reading of the paper's core claim, the machinery that carries it, and where it could break.

The paper establishes that adversarial examples can be crafted by optimizing small convolutional filters drawn from classic edge detection methods rather than relying on gradients or large generative models. These filters produce untargeted attacks that transfer across different neural networks and require only a single pass over each input image. Experiments demonstrate success rates between 30 and 80 percent on various networks while drastically reducing the number of parameters needed. A sympathetic reader would care because this reveals a low-cost mechanism for exposing model weaknesses that links basic image processing to machine learning vulnerabilities.

Core claim

Adversarial image filters are created by taking classic edge detection algorithms and optimizing their parameters specifically to deceive learning models. The resulting untargeted attacks transfer across models and require only a single pass over the input. With 3x3 filters, success rates reach 30 to 80 percent on different neural networks, and the approach reduces the number of parameters by five orders of magnitude compared to methods that use generative models.

What carries the argument

Adversarial image filters: small 3x3 convolutional kernels derived from edge detection algorithms and optimized to maximize model misclassification in one forward pass.

If this is right

  • Adversarial example generation becomes possible with only a single forward pass and no need for gradient computation or model queries during the attack.
  • Attacks can be performed with parameters reduced by five orders of magnitude relative to generative-model approaches, enabling very efficient execution.
  • Learned filters exhibit structures common to classic image filters, providing a direct link between standard image processing and model vulnerabilities.
  • High transferability between models allows attacks optimized on one network to succeed on others without retraining.

Where Pith is reading between the lines

These are editorial extensions of the paper, not claims the author makes directly.

  • This method could be tested on other data types such as audio or text to check whether similar low-parameter filter-based attacks exist outside images.
  • The filters might serve as a lightweight benchmark for measuring robustness across a wide range of models without requiring full retraining for each test.
  • If the approach scales, it could motivate new preprocessing defenses that detect or neutralize perturbations resembling optimized edge filters.

Load-bearing premise

Filters optimized on a limited set of models and data will remain effective and transferable to unseen architectures and datasets without additional per-model tuning or retraining.

What would settle it

Applying the optimized filters to a new neural network architecture or a different dataset and measuring attack success rates that drop to the level of random guessing would falsify the transferability claim.

Figures

Figures reproduced from arXiv: 2605.01098 by Alexander Warnecke, Konrad Rieck.

Figure 1
Figure 1. Figure 1: Example of adversarial image filters. An input image (left) is perturbed using a convolutional filter (middle) to generate an adversarial example (right). In this paper, we deviate from this paradigm and propose adversarial image filters as a simple strategy for crafting adversarial examples. These filters deceive learning mod￾els through a single convolution and hence do not require the gradient informati… view at source ↗
Figure 2
Figure 2. Figure 2: Input images (top row) with LRP explanations (middle row) and edges (bottom row) detected by a Sobel filter [PITH_FULL_IMAGE:figures/full_fig_p003_2.png] view at source ↗
Figure 3
Figure 3. Figure 3: Examples for spatial filters that can be convolved with an image to transform it in various ways. we can initialize the filters with random values and apply optimization schemes like SGD to solve the problem above. An example for a resulting filter is shown in [PITH_FULL_IMAGE:figures/full_fig_p005_3.png] view at source ↗
Figure 4
Figure 4. Figure 4: Success rate of the edge filter attack for Sobel- and Laplacian-of-Gaussian (LoG) filter (left) and PSNR between origi￾nal image and the outcome of the attack (right). The dashed line indicate an attack where amplified Gaussian noise is added to the image such that the PSNR (left) or success rate (right) is equal to the corresponding Sobel filter attack outcome. 5.1. Edge Filter Attack [PITH_FULL_IMAGE:fi… view at source ↗
Figure 5
Figure 5. Figure 5: Outcome of the Sobel attack for different magnitudes µ. The classification result is based on a VGG-13 model and is shown under the images. The Gaussian noise is computed such that the success rate is equal to the Sobel attack. is clearly more prone to our attack, whereas the ResNet and Inception models behave very similar. To get a feeling for the strength of the perturbations, we present a trajectory of … view at source ↗
Figure 6
Figure 6. Figure 6: Success rate of the adaptive filter attack for different numbers of training examples and filter sizes. The PSNR is fixed to 20 for all models. 56 % with only 100 training examples. With 1 000 training examples, the success rate of the ResNet- and Inception model are increased by a factor of two and three respec￾tively. Allowing larger filters also increases the success rate strongly, especially when few t… view at source ↗
Figure 7
Figure 7. Figure 7: Resulting convolution filters of the adversarial filter attack after optimization with 2 000 training examples. VGG filters for channel one and three: The left and right neighbor pixels are subtracted from the current pixel with an (approximately) equal factor, resulting in an output of zero if the pixel values are constant. Secondly, the filters are different from averaging- or Gaussian filters utilized f… view at source ↗
Figure 8
Figure 8. Figure 8: Perturbations after convolving inputs with the adversarial filters optimized on different models. Filter training model PSNR=20 PSNR=30 VGG ResNet Inception VGG ResNet Inception VGG 100% 102% 96% 100% 98% 102% ResNet 93% 100% 82% 50% 100% 51% Inception 86% 69% 100% 91% 75% 100% [PITH_FULL_IMAGE:figures/full_fig_p009_8.png] view at source ↗
Figure 9
Figure 9. Figure 9: Adversarial examples resulting from the adversarial filter attack on the ResNet. Classification is given below the images. proach that allows creating highly imperceptible and highly effective adversarial examples with a fraction of parameters compared to the state of the art. A natural extension to our approach is thus the usage of sequences of convolutions, maybe equipped with non-linear transformation f… view at source ↗
Figure 10
Figure 10. Figure 10: Adversarial filters of size 3 × 3 and a PSNR value of 20 for all models.       −0.97 0.34 0.63 −0.05 −0.23 0.53 −0.95 0.03 0.5 −0.09 0.71 0.0 −0.45 0.12 0.56 −0.38 0.6 0.36 −1.0 0.33 −0.04 −0.11 0.38 0.27 −0.8       Channel 1       0.04 0.29 −0.37 −0.43 0.24 0.27 −0.6 0.23 0.34 −0.26 −0.46 0.34 1.0 0.3 −0.44 −0.34 0.45 0.24 −0.68 0.23 0.26 −0.58 −0.35 0.43 0.02       Channel 2   … view at source ↗
Figure 11
Figure 11. Figure 11: Adversarial filters of size 5 × 5 for a PSNR value of 20 for the ResNet architecture [PITH_FULL_IMAGE:figures/full_fig_p013_11.png] view at source ↗
read the original abstract

Adversarial examples in machine learning are typically generated using gradients, obtained either directly through access to the model or approximated via queries to it. In this paper, we propose a much simpler approach to craft adversarial examples, drawing inspiration from insights of explainable machine learning. In particular, we design \emph{adversarial image filters} that are based on classic edge detection algorithms but optimized to deceive learning models. The resulting untargeted attacks are transferable and require only a single pass over the input. Empirically, we find that 3x3 filters already enable success rates between 30% and 80% on different neural networks. Compared to related approaches using generative models for crafting adversarial examples, we reduce the number of parameters by five orders of magnitude, resulting in a very efficient attack. When investigating the parameters of the learned filters, we observe interesting properties such as a high transferability between models and structures common to classic image filters. Our results provide further insights into the vulnerability of neural networks and their fragility to malicious noise.

Editorial analysis

A structured set of objections, weighed in public.

Desk editor's note, referee report, simulated authors' rebuttal, and a circularity audit. Tearing a paper down is the easy half of reading it; the pith above is the substance, this is the friction.

Referee Report

3 major / 2 minor

Summary. The paper proposes adversarial image filters: 3x3 convolutional kernels inspired by edge-detection algorithms but with weights optimized to maximize model misclassification. These filters generate untargeted adversarial examples via a single forward pass through the filter, claiming 30-80% success rates across neural networks, high transferability, and a five-order-of-magnitude reduction in parameters relative to generative-model attacks.

Significance. If the empirical claims are substantiated with proper optimization details, baselines, and cross-architecture transfer results, the work would demonstrate that extremely low-parameter structured perturbations can reliably fool deep networks. This would strengthen understanding of model fragility to simple, interpretable noise and offer a computationally cheap attack primitive that could inform both attack and defense research.

major comments (3)
  1. [§4] §4 (Experiments): the abstract states success rates of 30-80% but provides no information on the optimization procedure (loss function, surrogate models/datasets used to learn the filter weights, number of optimization steps, or whether a single filter set is used for all target models). Without these details the central efficiency and transferability claims cannot be evaluated.
  2. [§4.3] §4.3 (Transferability): the claim that the learned 3x3 filters exhibit 'high transferability between models' and require 'no per-model tuning' is load-bearing for the practical utility argument. The manuscript must report success rates when filters optimized on one architecture family (e.g., CNNs) are applied to held-out families (e.g., Vision Transformers) and on datasets disjoint from the optimization set; current evidence appears limited to the models used for filter learning.
  3. [Table 1] Table 1 / §4.1: the five-order-of-magnitude parameter reduction is compared only to generative-model baselines. The paper should also include simple non-learned baselines (random 3x3 filters, classic Sobel/Prewitt kernels, Gaussian noise of matched magnitude) to establish that the optimization step itself is responsible for the reported success rates rather than the filter structure alone.
minor comments (2)
  1. [Abstract / §3] The abstract and introduction use 'almost for free' and 'single pass' without clarifying that filter optimization itself requires gradient access or queries to surrogate models; this should be stated explicitly in the method section.
  2. [§3] Notation for the filter coefficients and the exact convolution operation applied to produce the adversarial image should be formalized with an equation in §3.

Simulated Author's Rebuttal

3 responses · 0 unresolved

We thank the referee for the constructive feedback. The comments highlight important areas for improving the clarity, reproducibility, and strength of the empirical claims. We address each major comment below and will revise the manuscript to incorporate the requested details and comparisons.

read point-by-point responses
  1. Referee: [§4] §4 (Experiments): the abstract states success rates of 30-80% but provides no information on the optimization procedure (loss function, surrogate models/datasets used to learn the filter weights, number of optimization steps, or whether a single filter set is used for all target models). Without these details the central efficiency and transferability claims cannot be evaluated.

    Authors: We agree that these optimization details are essential for evaluating the claims. The current manuscript provides only a high-level description. In the revised version we will expand §4 with a dedicated paragraph (or subsection) that explicitly states the loss function, the surrogate models and datasets used to learn the filter weights, the number of optimization steps, and confirms that a single filter set is learned and applied to all target models without per-model retuning. revision: yes

  2. Referee: [§4.3] §4.3 (Transferability): the claim that the learned 3x3 filters exhibit 'high transferability between models' and require 'no per-model tuning' is load-bearing for the practical utility argument. The manuscript must report success rates when filters optimized on one architecture family (e.g., CNNs) are applied to held-out families (e.g., Vision Transformers) and on datasets disjoint from the optimization set; current evidence appears limited to the models used for filter learning.

    Authors: We acknowledge that transfer results are currently limited to the CNN architectures used during filter optimization. The manuscript already shows that the same 3x3 filter works across those models with no per-model tuning. To strengthen the claim we will add, in the revised §4.3, success rates obtained by applying the CNN-optimized filters to Vision Transformers and on a dataset partition held out from the optimization set. revision: yes

  3. Referee: [Table 1] Table 1 / §4.1: the five-order-of-magnitude parameter reduction is compared only to generative-model baselines. The paper should also include simple non-learned baselines (random 3x3 filters, classic Sobel/Prewitt kernels, Gaussian noise of matched magnitude) to establish that the optimization step itself is responsible for the reported success rates rather than the filter structure alone.

    Authors: We agree that non-learned baselines are needed to isolate the benefit of optimization. In the revised manuscript we will augment Table 1 and the §4.1 discussion with results for random 3x3 filters, the classic Sobel and Prewitt kernels, and Gaussian noise of matched magnitude. These additions will demonstrate that the learned filters outperform the unoptimized alternatives. revision: yes

Circularity Check

0 steps flagged

No circularity in derivation chain

full rationale

The paper presents an empirical attack construction: 3x3 convolutional filters are explicitly optimized against model outputs to produce adversarial perturbations, with success rates measured directly on held-out models and datasets. No equations or first-principles derivations are offered that reduce to the optimization inputs by construction; the reported 30-80% success rates and five-order parameter reduction are comparative experimental outcomes, not fitted predictions renamed as results. No self-citation chains, uniqueness theorems, or ansatzes are invoked to justify the core method. The approach is self-contained against external benchmarks via direct measurement.

Axiom & Free-Parameter Ledger

1 free parameters · 1 axioms · 0 invented entities

The central claim rests on the empirical observation that nine-parameter filters can be optimized to produce transferable adversarial effects; the only free parameters are the filter coefficients themselves.

free parameters (1)
  • 3x3 filter coefficients
    Nine scalar values per filter that are optimized to maximize model deception on chosen networks.
axioms (1)
  • domain assumption Neural networks remain vulnerable to structured, low-parameter perturbations derived from image-processing primitives
    Invoked when the authors assume that edge-detection-inspired kernels can be tuned to produce effective attacks.

pith-pipeline@v0.9.0 · 5477 in / 1199 out tokens · 60664 ms · 2026-05-09T18:56:27.113426+00:00 · methodology

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.

Reference graph

Works this paper leans on

88 extracted references · 9 canonical work pages

  1. [1]

    ImageNet Classification with Deep Convolutional Neural Networks , author =

  2. [2]

    Very Deep Convolutional Networks for Large-Scale Image Recognition , booktitle = iclr, year =

    Karen Simonyan and Andrew Zisserman , @editor =. Very Deep Convolutional Networks for Large-Scale Image Recognition , booktitle = iclr, year =

  3. [3]

    , author =

    Towards Evaluating the Robustness of Neural Networks. , author =

  4. [4]

    Goodfellow and Rob Fergus , title =

    Christian Szegedy and Wojciech Zaremba and Ilya Sutskever and Joan Bruna and Dumitru Erhan and Ian J. Goodfellow and Rob Fergus , title =. 2014 , @url =

  5. [5]

    Goodfellow and Jonathon Shlens and Christian Szegedy , title =

    Ian J. Goodfellow and Jonathon Shlens and Christian Szegedy , title =. 2015 , @url =

  6. [6]

    2018 , @url =

    Andrew Ilyas and Logan Engstrom and Anish Athalye and Jessy Lin , title =. 2018 , @url =

  7. [7]

    Gardner and Yurong You and Andrew Gordon Wilson and Kilian Q

    Chuan Guo and Jacob R. Gardner and Yurong You and Andrew Gordon Wilson and Kilian Q. Weinberger , @editor =. Simple Black-box Adversarial Attacks , booktitle = icml, @series =. 2019 , @url =

  8. [8]

    2017 , pages =

    ZOO: Zeroth Order Optimization Based Black-box Attacks to Deep Neural Networks without Training Substitute Models , author=. 2017 , pages =

  9. [9]

    2016 , pages=

    Universal Adversarial Perturbations , author=. 2016 , pages=

  10. [10]

    2021 , pages =

    Learning Transferable Adversarial Perturbations , author=. 2021 , pages =

  11. [11]

    Xiao, Chaowei and Li, Bo and Zhu, Jun-Yan and He, Warren and Liu, Mingyan and Song, Dawn , title =. Proc. of the International Joint Conference on Artificial Intelligence , pages =. 2018 , @isbn =

  12. [12]

    Marco Ancona and Enea Ceolini and Cengiz

  13. [13]

    2018 , pages=

    Sanity Checks for Saliency Maps , author=. 2018 , pages=

  14. [14]

    Kaiming He and Xiangyu Zhang and Shaoqing Ren and Jian Sun , booktitle = cvpr, pages =

  15. [15]

    2016 , volume=

    Szegedy, Christian and Vanhoucke, Vincent and Ioffe, Sergey and Shlens, Jon and Wojna, Zbigniew , booktitle=cvpr, title=. 2016 , volume=

  16. [16]

    , title=

    Sun, Rui and Lei, Tao and Chen, Qi and Wang, Zexuan and Du, Xiaogang and Zhao, Weiqiang and Nandi, Asoke K. , title=. Frontiers in Signal Processing , volume=

  17. [17]

    2009 IEEE conference on computer vision and pattern recognition , pages=

    Imagenet: A large-scale hierarchical image database , author=. 2009 IEEE conference on computer vision and pattern recognition , pages=. 2009 , organization=

  18. [18]

    2019 , pages =

    Do ImageNet Classifiers Generalize to ImageNet? , author=. 2019 , pages =

  19. [19]

    2017 , pages=

    Generative Adversarial Perturbations , author=. 2017 , pages=

  20. [20]

    Learning to Attack: Adversarial Transformation Networks , volume=

    Baluja, Shumeet and Fischer, Ian , pages =. Learning to Attack: Adversarial Transformation Networks , volume=. doi:10.1609/aaai.v32i1.11672 , @number=

  21. [21]

    On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , journal =

    Sebastian Bach and Alexander Binder and Grégoire Montavon and Frederick Klauschen and Klaus-Robert M. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , journal =

  22. [22]

    Niels J. S. M. Visualization of neural networks using saliency maps , booktitle =. 1995 , @url =. doi:10.1109/ICNN.1995.488997 , timestamp =

  23. [23]

    Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps , Author =

  24. [24]

    Axiomatic Attribution for Deep Networks , author =. Proc. of International Conference on Machine Learning (ICML) , pages =

  25. [25]

    Prewitt, J. M. S. Object enhancement and extraction. Picture Processing and. Psychopictorics. 1970

  26. [26]

    Scharr, Hanno , year =

  27. [27]

    Image Segmentation using Extended Edge Operator for Mammographic Images , volume =

    Kekre, Hemant and Gharge, Saylee , year =. Image Segmentation using Extended Edge Operator for Mammographic Images , volume =

  28. [28]

    A 3×3 isotropic gradient operator for image processing , journal =

    Sobel, Irwin and Feldman, Gary , year =. A 3×3 isotropic gradient operator for image processing , journal =

  29. [30]

    2016 , pages =

    Practical Black-Box Attacks against Machine Learning , author=. 2016 , pages =

  30. [31]

    Berkay Celik and Ananthram Swami , title =

    Nicolas Papernot and Patrick McDaniel and Somesh Jha and Matt Fredrikson and Z. Berkay Celik and Ananthram Swami , title =

  31. [32]

    Moosavi-Dezfooli, Seyed-Mohsen and Fawzi, Alhussein and Frossard, Pascal , title =

  32. [33]

    Andrew Ilyas and Logan Engstrom and Aleksander Madry , title =

  33. [34]

    2019 , pages=

    HopSkipJumpAttack: A Query-Efficient Decision-Based Attack , author=. 2019 , pages=

  34. [35]

    International Conference on Learning Representations (ICLR) , @publisher =

    Wieland Brendel and Jonas Rauber and Matthias Bethge , title =. International Conference on Learning Representations (ICLR) , @publisher =. 2018 , biburl =

  35. [36]

    Generative Adversarial Nets , year =

    Goodfellow, Ian and Pouget-Abadie, Jean and Mirza, Mehdi and Xu, Bing and Warde-Farley, David and Ozair, Sherjil and Courville, Aaron and Bengio, Yoshua , booktitle = nips, pages=. Generative Adversarial Nets , year =

  36. [37]

    2018 , @url =

    Aleksander Madry and Aleksandar Makelov and Ludwig Schmidt and Dimitris Tsipras and Adrian Vladu , title =. 2018 , @url =

  37. [38]

    2019 , @url=

    Adversarial Training for Free! , author=. 2019 , @url=

  38. [39]

    The Space of Transferable Adversarial Examples , journal =

    Florian Tram. The Space of Transferable Adversarial Examples , journal =

  39. [40]

    , booktitle=

    Agarwal, Akshay and Vatsa, Mayank and Singh, Richa and Ratha, Nalini K. , booktitle=. Noise is Inside Me! Generating Adversarial Perturbations with Noise Derived from Natural Filters , year=

  40. [41]

    Crafting Adversarial Perturbations via Transformed Image Component Swapping , year=

    Agarwal, Akshay and Ratha, Nalini and Vatsa, Mayank and Singh, Richa , journal=. Crafting Adversarial Perturbations via Transformed Image Component Swapping , year=

  41. [42]

    Edgefool: an Adversarial Image Enhancement Filter , year=

    Shamsabadi, Ali Shahin and Oh, Changjae and Cavallaro, Andrea , booktitle=. Edgefool: an Adversarial Image Enhancement Filter , year=

  42. [44]

    2023 , eprint=

    Don't Look into the Sun: Adversarial Solarization Attacks on Image Classifiers , author=. 2023 , eprint=

  43. [46]

    Zeiler and Rob Fergus , title =

    Matthew D. Zeiler and Rob Fergus , title =. Proceedings of the European Conference on Computer Vision (ECCV) , volume =. 2014 , url =

  44. [47]

    Proceedings of the International Conference on Learning Representations 2018 , year=

    ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness , author=. Proceedings of the International Conference on Learning Representations 2018 , year=

  45. [48]

    Watch out! Motion is Blurring the Vision of Your Deep Neural Networks , url =

    Guo, Qing and Juefei-Xu, Felix and Xie, Xiaofei and Ma, Lei and Wang, Jian and Yu, Bing and Feng, Wei and Liu, Yang , booktitle =. Watch out! Motion is Blurring the Vision of Your Deep Neural Networks , url =

  46. [49]

    J., Hardt, M., and Kim, B

    Adebayo, J., Gilmer, J., Muelly, M., Goodfellow, I. J., Hardt, M., and Kim, B. Sanity checks for saliency maps. In Advances in Neural Information Proccessing Systems ( NIPS ) , pp.\ 9505--9515, 2018

  47. [50]

    Agarwal, A., Vatsa, M., Singh, R., and Ratha, N. K. Noise is inside me! generating adversarial perturbations with noise derived from natural filters. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp.\ 3354--3363, 2020. doi:10.1109/CVPRW50498.2020.00395

  48. [51]

    Towards better understanding of gradient-based attribution methods for deep neural networks

    Ancona, M., Ceolini, E., \" O ztireli, C., and Gross, M. Towards better understanding of gradient-based attribution methods for deep neural networks. In Proc. of the International Conference on Learning Representations ( ICLR ) , 2018

  49. [52]

    On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation

    Bach, S., Binder, A., Montavon, G., Klauschen, F., M \"u ller, K.-R., and Samek, W. On pixel-wise explanations for non-linear classifier decisions by layer-wise relevance propagation. PLoS ONE, 10 0 (7), July 2015

  50. [53]

    and Fischer, I

    Baluja, S. and Fischer, I. Learning to attack: Adversarial transformation networks. In Proc. of the AAAI Conference on Artificial Intelligence (AAAI) , volume 32, pp.\ 2687--2695, 2018

  51. [54]

    Decision-based adversarial attacks: Reliable attacks against black-box machine learning models

    Brendel, W., Rauber, J., and Bethge, M. Decision-based adversarial attacks: Reliable attacks against black-box machine learning models. In International Conference on Learning Representations (ICLR), 2018

  52. [55]

    and Wagner, D

    Carlini, N. and Wagner, D. A. Towards evaluating the robustness of neural networks. In Proc. of the IEEE Symposium on Security and Privacy (S&P) , pp.\ 39--57, 2017

  53. [56]

    and Jordan, M

    Chen, J. and Jordan, M. I. Hopskipjumpattack: A query-efficient decision-based attack. pp.\ 1277--1294, 2019

  54. [57]

    Zoo: Zeroth order optimization based black-box attacks to deep neural networks without training substitute models

    Chen, P.-Y., Zhang, H., Sharma, Y., Yi, J., and Hsieh, C.-J. Zoo: Zeroth order optimization based black-box attacks to deep neural networks without training substitute models. Proc. of ACM Workshop on Artificial Intelligence and Security ( AISEC ) , pp.\ 15--26, 2017

  55. [58]

    Intriguing properties of adversarial examples

    Cubuk, E., Zoph, B., Schoenholz, S., and Le, Q. Intriguing properties of adversarial examples. arXiv:1711.02846, 11 2017

  56. [59]

    Alan Chan, Carson Ezell, Max Kaufmann, Kevin Wei, Lewis Hammond, Herbie Bradley, Emma Bluemke, Nitarshan Rajkumar, David Krueger, Noam Kolt, Lennart Heim, and Markus Anderljung

    Debenedetti, E., Carlini, N., and Tramer, F. Evading black-box classifiers without breaking eggs. In 2024 IEEE Conference on Secure and Trustworthy Machine Learning (SaTML), pp.\ 408--424, apr 2024. doi:10.1109/SaTML59370.2024.00027. URL https://doi.ieeecomputersociety.org/10.1109/SaTML59370.2024.00027

  57. [60]

    Imagenet: A large-scale hierarchical image database

    Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., and Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In 2009 IEEE conference on computer vision and pattern recognition, pp.\ 248--255. Ieee, 2009

  58. [61]

    doi:10.48550/ARXIV.1811.12231

    Geirhos, R., Rubisch, P., Michaelis, C., Bethge, M., Wichmann, F., and Brendel, W. Imagenet-trained cnns are biased towards texture; increasing shape bias improves accuracy and robustness. In Proceedings of the International Conference on Learning Representations 2018, volume abs/1811.12231, 2018. URL https://api.semanticscholar.org/CorpusID:54101493

  59. [62]

    Generative adversarial nets

    Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., and Bengio, Y. Generative adversarial nets. In Advances in Neural Information Proccessing Systems ( NIPS ) , pp.\ 2672--2680, 2014

  60. [63]

    J., Shlens, J., and Szegedy, C

    Goodfellow, I. J., Shlens, J., and Szegedy, C. Explaining and harnessing adversarial examples. In Proc. of the International Conference on Learning Representations ( ICLR ) , 2015

  61. [64]

    R., You, Y., Wilson, A

    Guo, C., Gardner, J. R., You, Y., Wilson, A. G., and Weinberger, K. Q. Simple black-box adversarial attacks. In Proc. of the International Conference on Machine Learning ( ICML ) , volume 97, pp.\ 2484--2493, 2019

  62. [65]

    Watch out! motion is blurring the vision of your deep neural networks

    Guo, Q., Juefei-Xu, F., Xie, X., Ma, L., Wang, J., Yu, B., Feng, W., and Liu, Y. Watch out! motion is blurring the vision of your deep neural networks. In Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M., and Lin, H. (eds.), Advances in Neural Information Processing Systems, volume 33, pp.\ 975--985. Curran Associates, Inc., 2020. URL https://proceedi...

  63. [66]

    Deep residual learning for image recognition

    He, K., Zhang, X., Ren, S., and Sun, J. Deep residual learning for image recognition. In Proc. of the International Conference on Computer Vision and Pattern Recognition (CVPR), pp.\ 770--778, 2016

  64. [67]

    Edge enhancement improves adversarial robustness in image classification

    He, L., Ai, Q., Lei, Y., Pan, L., Ren, Y., and Xu, Z. Edge enhancement improves adversarial robustness in image classification. Neurocomputing, 518: 0 122--132, 2023. ISSN 0925-2312. doi:https://doi.org/10.1016/j.neucom.2022.10.059. URL https://www.sciencedirect.com/science/article/pii/S092523122201342X

  65. [68]

    Black-box adversarial attacks with limited queries and information

    Ilyas, A., Engstrom, L., Athalye, A., and Lin, J. Black-box adversarial attacks with limited queries and information. In Proc. of the International Conference on Machine Learning ( ICML ) , pp.\ 2142--2151, 2018

  66. [69]

    Prior convictions: Black-box adversarial attacks with bandits and priors

    Ilyas, A., Engstrom, L., and Madry, A. Prior convictions: Black-box adversarial attacks with bandits and priors. In Proc. of the International Conference on Learning Representations ( ICLR ) , 2019

  67. [70]

    and Gharge, S

    Kekre, H. and Gharge, S. Image segmentation using extended edge operator for mammographic images. International Journal on Computer Science and Engineering, 2, 07 2010

  68. [71]

    Krizhevsky, A., Sutskever, I., and Hinton, G. E. Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Proccessing Systems ( NIPS ) , pp.\ 1106--1114. Curran Associates, Inc., 2012

  69. [72]

    Towards deep learning models resistant to adversarial attacks

    Madry, A., Makelov, A., Schmidt, L., Tsipras, D., and Vladu, A. Towards deep learning models resistant to adversarial attacks. In Proc. of the International Conference on Learning Representations ( ICLR ) , 2018

  70. [73]

    Universal adversarial perturbations

    Moosavi-Dezfooli, S.-M., Fawzi, A., Fawzi, O., and Frossard, P. Universal adversarial perturbations. Proc. of the International Conference on Computer Vision and Pattern Recognition (CVPR), pp.\ 86--94, 2016 a

  71. [74]

    Deepfool: A simple and accurate method to fool deep neural networks

    Moosavi-Dezfooli, S.-M., Fawzi, A., and Frossard, P. Deepfool: A simple and accurate method to fool deep neural networks. In Proc. of the International Conference on Computer Vision and Pattern Recognition (CVPR), pp.\ 2574--2582, June 2016 b

  72. [75]

    Nakka, K. K. and Salzmann, M. Learning transferable adversarial perturbations. In Advances in Neural Information Proccessing Systems ( NIPS ) , pp.\ 13950--13962, 2021

  73. [76]

    J., Jha, S., Celik, Z

    Papernot, N., Mcdaniel, P., Goodfellow, I. J., Jha, S., Celik, Z. B., and Swami, A. Practical black-box attacks against machine learning. Proc. of the ACM Asia Conference on Computer and Communications Security ( ASIA CCS ) , pp.\ 506--519, 2016 a

  74. [77]

    B., and Swami, A

    Papernot, N., McDaniel, P., Jha, S., Fredrikson, M., Celik, Z. B., and Swami, A. The limitations of deep learning in adversarial settings. In Proc. of the IEEE European Symposium on Security and Privacy ( EuroS&P ) , pp.\ 372--387, 2016 b

  75. [78]

    Poursaeed, O., Katsman, I., Gao, B., and Belongie, S. J. Generative adversarial perturbations. Proc. of the International Conference on Computer Vision and Pattern Recognition (CVPR), pp.\ 4422--4431, 2017

  76. [79]

    Prewitt, J. M. S. Object enhancement and extraction. Picture Processing and. Psychopictorics, 1970

  77. [80]

    Do imagenet classifiers generalize to imagenet? In Proc

    Recht, B., Roelofs, R., Schmidt, L., and Shankar, V. Do imagenet classifiers generalize to imagenet? In Proc. of the International Conference on Machine Learning ( ICML ) , pp.\ 5389--5400, 2019

  78. [81]

    Optimal operators in digital image processing [Elektronische Ressource] /

    Scharr, H. Optimal operators in digital image processing [Elektronische Ressource] /. PhD thesis, 09 2014

  79. [82]

    P., Studer, C., Davis, L

    Shafahi, A., Najibi, M., Ghiasi, A., Xu, Z., Dickerson, J. P., Studer, C., Davis, L. S., Taylor, G., and Goldstein, T. Adversarial training for free! In Advances in Neural Information Proccessing Systems ( NIPS ) , pp.\ 3353--3364, 2019

  80. [83]

    Vggsound: A Large-Scale Audio-Visual Dataset

    Shamsabadi, A. S., Oh, C., and Cavallaro, A. Edgefool: an adversarial image enhancement filter. In ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.\ 1898--1902, 2020. doi:10.1109/ICASSP40776.2020.9054368

Showing first 80 references.