Results 51 to 60 of about 2,268,403 (339)
Benchmarking Adversarial Robustness
Deep neural networks are vulnerable to adversarial examples, which becomes one of the most important research problems in the development of deep learning. While a lot of efforts have been made in recent years, it is of great significance to perform correct and complete evaluations of the adversarial attack and defense algorithms.
Dong, Yinpeng +6 more
openaire +2 more sources
Adversarial training, a widely used technique for fortifying the robustness of machine learning models, has seen its effectiveness further bolstered by modifying loss functions or incorporating additional terms into the training objective.
Sander Joos +4 more
doaj +1 more source
The prediction accuracy has been the long-lasting and sole standard for comparing the performance of different image classification models, including the ImageNet competition.
Su, Dong +5 more
core +1 more source
Adversarially Robust Neural Architectures
13 pages, 5 figures, 8 ...
Minjing Dong +3 more
openaire +3 more sources
Decoupled Adversarial Contrastive Learning for Self-supervised Adversarial Robustness
Accepted by ECCV 2022 oral ...
Zhang, Chaoning +6 more
openaire +2 more sources
A Comparative Study on the Performance and Security Evaluation of Spiking Neural Networks
The brain-inspired Spiking neural networks (SNN) claim to present advantages for visual classification tasks in terms of energy efficiency and inherent robustness.
Yanjie Li +3 more
doaj +1 more source
Disentangling Adversarial Robustness and Generalization
Obtaining deep networks that are robust against adversarial examples and generalize well is an open problem. A recent hypothesis even states that both robust and accurate models are impossible, i.e., adversarial robustness and generalization are ...
Hein, Matthias +2 more
core +1 more source
Adversarial Robustness: Softmax versus Openmax
Deep neural networks (DNNs) provide state-of-the-art results on various tasks and are widely used in real world applications. However, it was discovered that machine learning models, including the best performing DNNs, suffer from a fundamental problem ...
Boult, Terrance E. +2 more
core +1 more source
Analysis of classifiers' robustness to adversarial perturbations
The goal of this paper is to analyze an intriguing phenomenon recently discovered in deep networks, namely their instability to adversarial perturbations (Szegedy et. al., 2014).
Fawzi, Alhussein +2 more
core +1 more source
Understanding the Energy vs. Adversarial Robustness Trade-Off in Deep Neural Networks
Adversarial examples, which are crafted by adding small perturbations to typical inputs in order to fool the prediction of a deep neural network (DNN), pose a threat to security-critical applications, and robustness against adversarial examples is ...
Kyungmi Lee, Anantha P. Chandrakasan
doaj +1 more source

