126 related articles for article (PubMed ID: 38376968)
1. Adaptive Perturbation for Adversarial Attack.
Yuan Z; Zhang J; Jiang Z; Li L; Shan S
IEEE Trans Pattern Anal Mach Intell; 2024 Feb; PP():. PubMed ID: 38376968
[TBL] [Abstract][Full Text] [Related]
2. Enhancing robustness in video recognition models: Sparse adversarial attacks and beyond.
Mu R; Marcolino L; Ni Q; Ruan W
Neural Netw; 2024 Mar; 171():127-143. PubMed ID: 38091756
[TBL] [Abstract][Full Text] [Related]
3. ABCAttack: A Gradient-Free Optimization Black-Box Attack for Fooling Deep Image Classifiers.
Cao H; Si C; Sun Q; Liu Y; Li S; Gope P
Entropy (Basel); 2022 Mar; 24(3):. PubMed ID: 35327923
[TBL] [Abstract][Full Text] [Related]
4. Strengthening transferability of adversarial examples by adaptive inertia and amplitude spectrum dropout.
Li H; Yu W; Huang H
Neural Netw; 2023 Aug; 165():925-937. PubMed ID: 37441909
[TBL] [Abstract][Full Text] [Related]
5. Generalizable Black-Box Adversarial Attack With Meta Learning.
Yin F; Zhang Y; Wu B; Feng Y; Zhang J; Fan Y; Yang Y
IEEE Trans Pattern Anal Mach Intell; 2024 Mar; 46(3):1804-1818. PubMed ID: 37021863
[TBL] [Abstract][Full Text] [Related]
6. Boosting the transferability of adversarial examples via stochastic serial attack.
Hao L; Hao K; Wei B; Tang XS
Neural Netw; 2022 Jun; 150():58-67. PubMed ID: 35305532
[TBL] [Abstract][Full Text] [Related]
7. Improving the Transferability of Adversarial Examples With a Noise Data Enhancement Framework and Random Erasing.
Xie P; Shi S; Yang S; Qiao K; Liang N; Wang L; Chen J; Hu G; Yan B
Front Neurorobot; 2021; 15():784053. PubMed ID: 34955802
[TBL] [Abstract][Full Text] [Related]
8. An adversarial example attack method based on predicted bounding box adaptive deformation in optical remote sensing images.
Dai L; Wang J; Yang B; Chen F; Zhang H
PeerJ Comput Sci; 2024; 10():e2053. PubMed ID: 38855243
[TBL] [Abstract][Full Text] [Related]
9. Improving Adversarial Robustness via Attention and Adversarial Logit Pairing.
Li X; Goodman D; Liu J; Wei T; Dou D
Front Artif Intell; 2021; 4():752831. PubMed ID: 35156010
[TBL] [Abstract][Full Text] [Related]
10. Gradient Correction for White-Box Adversarial Attacks.
Liu H; Ge Z; Zhou Z; Shang F; Liu Y; Jiao L
IEEE Trans Neural Netw Learn Syst; 2023 Oct; PP():. PubMed ID: 37819820
[TBL] [Abstract][Full Text] [Related]
11. Crafting Adversarial Perturbations via Transformed Image Component Swapping.
Agarwal A; Ratha N; Vatsa M; Singh R
IEEE Trans Image Process; 2022; 31():7338-7349. PubMed ID: 36094979
[TBL] [Abstract][Full Text] [Related]
12. DualFlow: Generating imperceptible adversarial examples by flow field and normalize flow-based model.
Liu R; Jin X; Hu D; Zhang J; Wang Y; Zhang J; Zhou W
Front Neurorobot; 2023; 17():1129720. PubMed ID: 36845066
[TBL] [Abstract][Full Text] [Related]
13. Towards Transferable Adversarial Attacks on Image and Video Transformers.
Wei Z; Chen J; Goldblum M; Wu Z; Goldstein T; Jiang YG; Davis LS
IEEE Trans Image Process; 2023; 32():6346-6358. PubMed ID: 37966925
[TBL] [Abstract][Full Text] [Related]
14. DEFEAT: Decoupled feature attack across deep neural networks.
Huang L; Gao C; Liu N
Neural Netw; 2022 Dec; 156():13-28. PubMed ID: 36228335
[TBL] [Abstract][Full Text] [Related]
15. SMGEA: A New Ensemble Adversarial Attack Powered by Long-Term Gradient Memories.
Che Z; Borji A; Zhai G; Ling S; Li J; Min X; Guo G; Le Callet P
IEEE Trans Neural Netw Learn Syst; 2022 Mar; 33(3):1051-1065. PubMed ID: 33296311
[TBL] [Abstract][Full Text] [Related]
16. Remix: Towards the transferability of adversarial examples.
Zhao H; Hao L; Hao K; Wei B; Cai X
Neural Netw; 2023 Jun; 163():367-378. PubMed ID: 37119676
[TBL] [Abstract][Full Text] [Related]
17. Enhancing adversarial attacks with resize-invariant and logical ensemble.
Shao Y; Zhang Y; Dong W; Zhang Q; Shan P; Guo J; Xu H
Neural Netw; 2024 May; 173():106194. PubMed ID: 38402809
[TBL] [Abstract][Full Text] [Related]
18. Image Adversarial Example Generation Method Based on Adaptive Parameter Adjustable Differential Evolution.
Lin Z; Peng C; Tan W; He X
Entropy (Basel); 2023 Mar; 25(3):. PubMed ID: 36981373
[TBL] [Abstract][Full Text] [Related]
19. Adversarial attack vulnerability of medical image analysis systems: Unexplored factors.
Bortsova G; González-Gonzalo C; Wetstein SC; Dubost F; Katramados I; Hogeweg L; Liefers B; van Ginneken B; Pluim JPW; Veta M; Sánchez CI; de Bruijne M
Med Image Anal; 2021 Oct; 73():102141. PubMed ID: 34246850
[TBL] [Abstract][Full Text] [Related]
20. Attention distraction with gradient sharpening for multi-task adversarial attack.
Liu B; Hu J; Deng W
Math Biosci Eng; 2023 Jun; 20(8):13562-13580. PubMed ID: 37679102
[TBL] [Abstract][Full Text] [Related]
[Next] [New Search]