Results 31 to 40 of about 79,418 (169)

Text Adversarial Purification as Defense against Adversarial Attacks

open access: yesProceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023
Accepted by ACL2023 main ...
Li, Linyang, Song, Demin, Qiu, Xipeng
openaire   +2 more sources

Hadamard’s Defense Against Adversarial Examples [PDF]

open access: yesIEEE Access, 2021
Adversarial images have become an increasing concern in real-world image recognition applications with deep neural networks (DNN). We observed that all the architectures in DNN use one-hot encoding after a softmax layer. The attacker can make minute modifications in the adversarial example rendering those changes imperceptible to a human observer.
Angello Hoyos, Ubaldo Ruiz, Edgar Chavez
openaire   +1 more source

Deepfake Cross-Model Defense Method Based on Generative Adversarial Network [PDF]

open access: yesJisuanji gongcheng
To reduce social risks caused by the abuse of deepfake technology, an active defense method against deep forgery based on a Generative Adversarial Network (GAN) is proposed. Adversarial samples are created by adding imperceptible perturbation to original
DAI Lei, CAO Lin, GUO Yanan, ZHANG Fan, DU Kangning
doaj   +1 more source

Adversarial Attacks and Defenses in Deep Learning

open access: yesEngineering, 2020
With the rapid developments of artificial intelligence (AI) and deep learning (DL) techniques, it is critical to ensure the security and robustness of the deployed algorithms.
Kui Ren   +3 more
doaj   +1 more source

Developing a Robust Defensive System against Adversarial Examples Using Generative Adversarial Networks

open access: yesBig Data and Cognitive Computing, 2020
In this work, we propose a novel defense system against adversarial examples leveraging the unique power of Generative Adversarial Networks (GANs) to generate new adversarial examples for model retraining. To do so, we develop an automated pipeline using
Shayan Taheri   +3 more
doaj   +1 more source

Gotta Catch 'Em All: Using Honeypots to Catch Adversarial Attacks on Neural Networks

open access: yes, 2020
Deep neural networks (DNN) are known to be vulnerable to adversarial attacks. Numerous efforts either try to patch weaknesses in trained models, or try to make it difficult or costly to compute adversarial examples that exploit them.
Li, Bo   +5 more
core   +1 more source

Leveraging linear mapping for model-agnostic adversarial defense

open access: yesFrontiers in Computer Science, 2023
In the ever-evolving landscape of deep learning, novel designs of neural network architectures have been thought to drive progress by enhancing embedded representations.
Huma Jamil   +5 more
doaj   +1 more source

Adversarial Robustness Enhancement of UAV-Oriented Automatic Image Recognition Based on Deep Ensemble Models

open access: yesRemote Sensing, 2023
Deep neural networks (DNNs) have been widely utilized in automatic visual navigation and recognition on modern unmanned aerial vehicles (UAVs), achieving state-of-the-art performances.
Zihao Lu, Hao Sun, Yanjie Xu
doaj   +1 more source

Defense against Adversarial Attacks Using High-Level Representation Guided Denoiser

open access: yes, 2018
Neural networks are vulnerable to adversarial examples, which poses a threat to their application in security sensitive systems. We propose high-level representation guided denoiser (HGD) as a defense for image classification.
Dong, Yinpeng   +5 more
core   +1 more source

Adversarial defenses via vector quantization

open access: yesNeurocomputing
This is the author-accepted version of our paper published in Neurocomputing.
Zhiyi Dong, Yongyi Mao
openaire   +2 more sources

Home - About - Disclaimer - Privacy