160 related articles for article (PubMed ID: 36531633)
1. Adversarial training and deep k-nearest neighbors improves adversarial defense of glaucoma severity detection.
Riza Rizky LM; Suyanto S
Heliyon; 2022 Dec; 8(12):e12275. PubMed ID: 36531633
[TBL] [Abstract][Full Text] [Related]
2. Robust Medical Diagnosis: A Novel Two-Phase Deep Learning Framework for Adversarial Proof Disease Detection in Radiology Images.
Haque SBU; Zafar A
J Imaging Inform Med; 2024 Feb; 37(1):308-338. PubMed ID: 38343214
[TBL] [Abstract][Full Text] [Related]
3. Adversarial Examples for Hamming Space Search.
Yang E; Liu T; Deng C; Tao D
IEEE Trans Cybern; 2020 Apr; 50(4):1473-1484. PubMed ID: 30561358
[TBL] [Abstract][Full Text] [Related]
4. Uni-image: Universal image construction for robust neural model.
Ho J; Lee BG; Kang DK
Neural Netw; 2020 Aug; 128():279-287. PubMed ID: 32454372
[TBL] [Abstract][Full Text] [Related]
5. Adversarial Attacks on Time Series.
Karim F; Majumdar S; Darabi H
IEEE Trans Pattern Anal Mach Intell; 2021 Oct; 43(10):3309-3320. PubMed ID: 32286957
[TBL] [Abstract][Full Text] [Related]
6. K-Anonymity inspired adversarial attack and multiple one-class classification defense.
Mygdalis V; Tefas A; Pitas I
Neural Netw; 2020 Apr; 124():296-307. PubMed ID: 32036227
[TBL] [Abstract][Full Text] [Related]
7. Robust image classification against adversarial attacks using elastic similarity measures between edge count sequences.
Oregi I; Del Ser J; Pérez A; Lozano JA
Neural Netw; 2020 Aug; 128():61-72. PubMed ID: 32442627
[TBL] [Abstract][Full Text] [Related]
8. LPF-Defense: 3D adversarial defense based on frequency analysis.
Naderi H; Noorbakhsh K; Etemadi A; Kasaei S
PLoS One; 2023; 18(2):e0271388. PubMed ID: 36745627
[TBL] [Abstract][Full Text] [Related]
9. Towards evaluating the robustness of deep diagnostic models by adversarial attack.
Xu M; Zhang T; Li Z; Liu M; Zhang D
Med Image Anal; 2021 Apr; 69():101977. PubMed ID: 33550005
[TBL] [Abstract][Full Text] [Related]
10. Adversarial attack vulnerability of medical image analysis systems: Unexplored factors.
Bortsova G; González-Gonzalo C; Wetstein SC; Dubost F; Katramados I; Hogeweg L; Liefers B; van Ginneken B; Pluim JPW; Veta M; Sánchez CI; de Bruijne M
Med Image Anal; 2021 Oct; 73():102141. PubMed ID: 34246850
[TBL] [Abstract][Full Text] [Related]
11. How Resilient Are Deep Learning Models in Medical Image Analysis? The Case of the Moment-Based Adversarial Attack (Mb-AdA).
Maliamanis TV; Apostolidis KD; Papakostas GA
Biomedicines; 2022 Oct; 10(10):. PubMed ID: 36289807
[TBL] [Abstract][Full Text] [Related]
12. Crafting Adversarial Perturbations via Transformed Image Component Swapping.
Agarwal A; Ratha N; Vatsa M; Singh R
IEEE Trans Image Process; 2022; 31():7338-7349. PubMed ID: 36094979
[TBL] [Abstract][Full Text] [Related]
13. On the role of deep learning model complexity in adversarial robustness for medical images.
Rodriguez D; Nayak T; Chen Y; Krishnan R; Huang Y
BMC Med Inform Decis Mak; 2022 Jun; 22(Suppl 2):160. PubMed ID: 35725429
[TBL] [Abstract][Full Text] [Related]
14. Adversarial Attack and Defense in Deep Ranking.
Zhou M; Wang L; Niu Z; Zhang Q; Zheng N; Hua G
IEEE Trans Pattern Anal Mach Intell; 2024 Aug; 46(8):5306-5324. PubMed ID: 38349823
[TBL] [Abstract][Full Text] [Related]
15. Adversarial attacks and adversarial training for burn image segmentation based on deep learning.
Chen L; Liang J; Wang C; Yue K; Li W; Fu Z
Med Biol Eng Comput; 2024 May; ():. PubMed ID: 38693327
[TBL] [Abstract][Full Text] [Related]
16. GLH: From Global to Local Gradient Attacks with High-Frequency Momentum Guidance for Object Detection.
Chen Y; Yang H; Wang X; Wang Q; Zhou H
Entropy (Basel); 2023 Mar; 25(3):. PubMed ID: 36981349
[TBL] [Abstract][Full Text] [Related]
17. Adversarial Robustness of Deep Reinforcement Learning Based Dynamic Recommender Systems.
Wang S; Cao Y; Chen X; Yao L; Wang X; Sheng QZ
Front Big Data; 2022; 5():822783. PubMed ID: 35592793
[TBL] [Abstract][Full Text] [Related]
18. Adversarial-Aware Deep Learning System Based on a Secondary Classical Machine Learning Verification Approach.
Alkhowaiter M; Kholidy H; Alyami MA; Alghamdi A; Zou C
Sensors (Basel); 2023 Jul; 23(14):. PubMed ID: 37514582
[TBL] [Abstract][Full Text] [Related]
19. Defending against adversarial attacks on Covid-19 classifier: A denoiser-based approach.
Kansal K; Krishna PS; Jain PB; R S; Honnavalli P; Eswaran S
Heliyon; 2022 Oct; 8(10):e11209. PubMed ID: 36311356
[TBL] [Abstract][Full Text] [Related]
20. ApaNet: adversarial perturbations alleviation network for face verification.
Sun G; Hu H; Su Y; Liu Q; Lu X
Multimed Tools Appl; 2023; 82(5):7443-7461. PubMed ID: 36035322
[TBL] [Abstract][Full Text] [Related]
[Next] [New Search]