Results 41 to 50 of about 173,113 (165)

Adversarial Examples Detection Beyond Image Space [PDF]

open access: yesICASSP 2021 - 2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2021
To appear in ICASSP ...
Chen, Kejiang   +6 more
openaire   +2 more sources

Are adversarial examples inevitable?

open access: yes, 2018
ISBN:978-1-7138-7273 ...
Shafahi, Ali   +4 more
openaire   +3 more sources

A Framework for Robust Deep Learning Models Against Adversarial Attacks Based on a Protection Layer Approach

open access: yesIEEE Access
Deep learning (DL) has demonstrated remarkable achievements in various fields. Nevertheless, DL models encounter significant challenges in detecting and defending against adversarial samples (AEs).
Mohammed Nasser Al-Andoli   +4 more
doaj   +1 more source

Improving Adversarial Robustness of CNNs via Maximum Margin

open access: yesApplied Sciences, 2022
In recent years, adversarial examples have aroused widespread research interest and raised concerns about the safety of CNNs. We study adversarial machine learning inspired by a support vector machine (SVM), where the decision boundary with maximum ...
Jiaping Wu, Zhaoqiang Xia, Xiaoyi Feng
doaj   +1 more source

Downstream-agnostic Adversarial Examples

open access: yes2023 IEEE/CVF International Conference on Computer Vision (ICCV), 2023
This paper has been accepted by the International Conference on Computer Vision (ICCV '23, October 2--6, 2023, Paris, France)
Zhou, Ziqi   +6 more
openaire   +2 more sources

Distinguishability of adversarial examples [PDF]

open access: yesProceedings of the 15th International Conference on Availability, Reliability and Security, 2020
Machine learning models can be easily fooled by adversarial examples which are generated from clean examples with small perturbations. This poses a critical challenge to machine learning security, and impedes the wide application of machine learning in many important domains such as computer vision and malware detection. From a unique angle, we propose
Yi Qin, Ryan Hunt, Chuan Yue
openaire   +1 more source

A Brute-Force Black-Box Method to Attack Machine Learning-Based Systems in Cybersecurity

open access: yesIEEE Access, 2020
Machine learning algorithms are widely utilized in cybersecurity. However, recent studies show that machine learning algorithms are vulnerable to adversarial examples.
Sicong Zhang, Xiaoyao Xie, Yang Xu
doaj   +1 more source

Robust Audio Adversarial Example for a Physical Attack

open access: yes, 2019
We propose a method to generate audio adversarial examples that can attack a state-of-the-art speech recognition model in the physical world. Previous work assumes that generated adversarial examples are directly fed to the recognition model, and is not ...
Sakuma, Jun, Yakura, Hiromu
core   +1 more source

Simple Transparent Adversarial Examples

open access: yes, 2021
There has been a rise in the use of Machine Learning as a Service (MLaaS) Vision APIs as they offer multiple services including pre-built models and algorithms, which otherwise take a huge amount of resources if built from scratch. As these APIs get deployed for high-stakes applications, it's very important that they are robust to different ...
Borkar, Jaydeep, Chen, Pin-Yu
openaire   +2 more sources

Unrestricted Adversarial Examples

open access: yes, 2018
We introduce a two-player contest for evaluating the safety and robustness of machine learning systems, with a large prize pool. Unlike most prior work in ML robustness, which studies norm-constrained adversaries, we shift our focus to unconstrained adversaries.
Brown, Tom B.   +5 more
openaire   +2 more sources

Home - About - Disclaimer - Privacy