Results 1 to 10 of about 215,342 (268)

Avoiding catastrophic overfitting in fast adversarial training with adaptive similarity step size. [PDF]

open access: yesPLoS ONE
Adversarial training has become a primary method for enhancing the robustness of deep learning models. In recent years, fast adversarial training methods have gained widespread attention due to their lower computational cost.
Jie-Chao Zhao   +5 more
doaj   +2 more sources

Adversarial Robustness Enhancement for Deep Learning-Based Soft Sensors: An Adversarial Training Strategy Using Historical Gradients and Domain Adaptation [PDF]

open access: yesSensors
Despite their high prediction accuracy, deep learning-based soft sensor (DLSS) models face challenges related to adversarial robustness against malicious adversarial attacks, which hinder their widespread deployment and safe application.
Runyuan Guo   +3 more
doaj   +2 more sources

STS-AT: A Structured Tensor Flow Adversarial Training Framework for Robust Intrusion Detection [PDF]

open access: yesSensors
Network intrusion detection is a key technology for ensuring cybersecurity. However, current methods face two major challenges: reliance on manual feature engineering, which leads to the loss of discriminative information, and the vulnerability of deep ...
Juntong Zhu   +4 more
doaj   +2 more sources

Exploring generative adversarial networks and adversarial training

open access: yesInternational Journal of Cognitive Computing in Engineering, 2022
Recognized as a realistic image generator, Generative Adversarial Network (GAN) occupies a progressive section in deep learning. Using generative modeling, the underlying generator model learns the real target distribution and outputs fake samples from ...
Afia Sajeeda, B M Mainul Hossain, Ph.D
doaj   +2 more sources

Adversarial Training for Free!

open access: yes, 2019
Adversarial training, in which a network is trained on adversarial examples, is one of the few defenses against adversarial attacks that withstands strong attacks.
Davis, Larry S.   +8 more
core   +4 more sources

Lightweight defense mechanism against adversarial attacks via adaptive pruning and robust distillation

open access: yes网络与信息安全学报, 2022
Adversarial training is one of the commonly used defense methods against adversarial attacks, by incorporating adversarial samples into the training process.However, the effectiveness of adversarial training heavily relied on the size of the trained ...
Bin WANG   +6 more
doaj   +3 more sources

Bridged adversarial training

open access: yesNeural Networks, 2023
Adversarial robustness is considered as a required property of deep neural networks. In this study, we discover that adversarially trained models might have significantly different characteristics in terms of margin and smoothness, even they show similar robustness.
Hoki Kim   +3 more
openaire   +3 more sources

Self-Supervised Adversarial Training [PDF]

open access: yesICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2020
Recent work has demonstrated that neural networks are vulnerable to adversarial examples. To escape from the predicament, many works try to harden the model in various ways, in which adversarial training is an effective way which learns robust feature representation so as to resist adversarial attacks.
Chen, Kejiang   +8 more
openaire   +2 more sources

Universal Adversarial Training Using Auxiliary Conditional Generative Model-Based Adversarial Attack Generation

open access: yesApplied Sciences, 2023
While Machine Learning has become the holy grail of modern-day computing, it has many security flaws that have yet to be addressed and resolved. Adversarial attacks are one of these security flaws, in which an attacker appends noise to data samples that ...
Hiskias Dingeto, Juntae Kim
doaj   +1 more source

Efficient Adversarial Training With Transferable Adversarial Examples [PDF]

open access: yes2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020
Adversarial training is an effective defense method to protect classification models against adversarial attacks. However, one limitation of this approach is that it can require orders of magnitude additional training time due to high cost of generating strong adversarial examples during training.
Zheng, Haizhong   +4 more
openaire   +2 more sources

Home - About - Disclaimer - Privacy