These tools will no longer be maintained as of December 31, 2024. Archived website can be found here. PubMed4Hh GitHub repository can be found here. Contact NLM Customer Service if you have questions.


BIOMARKERS

Molecular Biopsy of Human Tumors

- a resource for Precision Medicine *

119 related articles for article (PubMed ID: 38376968)

  • 1. Adaptive Perturbation for Adversarial Attack.
    Yuan Z; Zhang J; Jiang Z; Li L; Shan S
    IEEE Trans Pattern Anal Mach Intell; 2024 Aug; 46(8):5663-5676. PubMed ID: 38376968
    [TBL] [Abstract][Full Text] [Related]  

  • 2. Enhancing robustness in video recognition models: Sparse adversarial attacks and beyond.
    Mu R; Marcolino L; Ni Q; Ruan W
    Neural Netw; 2024 Mar; 171():127-143. PubMed ID: 38091756
    [TBL] [Abstract][Full Text] [Related]  

  • 3. ABCAttack: A Gradient-Free Optimization Black-Box Attack for Fooling Deep Image Classifiers.
    Cao H; Si C; Sun Q; Liu Y; Li S; Gope P
    Entropy (Basel); 2022 Mar; 24(3):. PubMed ID: 35327923
    [TBL] [Abstract][Full Text] [Related]  

  • 4. Strengthening transferability of adversarial examples by adaptive inertia and amplitude spectrum dropout.
    Li H; Yu W; Huang H
    Neural Netw; 2023 Aug; 165():925-937. PubMed ID: 37441909
    [TBL] [Abstract][Full Text] [Related]  

  • 5. Generalizable Black-Box Adversarial Attack With Meta Learning.
    Yin F; Zhang Y; Wu B; Feng Y; Zhang J; Fan Y; Yang Y
    IEEE Trans Pattern Anal Mach Intell; 2024 Mar; 46(3):1804-1818. PubMed ID: 37021863
    [TBL] [Abstract][Full Text] [Related]  

  • 6. Boosting the transferability of adversarial examples via stochastic serial attack.
    Hao L; Hao K; Wei B; Tang XS
    Neural Netw; 2022 Jun; 150():58-67. PubMed ID: 35305532
    [TBL] [Abstract][Full Text] [Related]  

  • 7. An adversarial example attack method based on predicted bounding box adaptive deformation in optical remote sensing images.
    Dai L; Wang J; Yang B; Chen F; Zhang H
    PeerJ Comput Sci; 2024; 10():e2053. PubMed ID: 38855243
    [TBL] [Abstract][Full Text] [Related]  

  • 8. Improving the Transferability of Adversarial Examples With a Noise Data Enhancement Framework and Random Erasing.
    Xie P; Shi S; Yang S; Qiao K; Liang N; Wang L; Chen J; Hu G; Yan B
    Front Neurorobot; 2021; 15():784053. PubMed ID: 34955802
    [TBL] [Abstract][Full Text] [Related]  

  • 9. Improving Adversarial Robustness via Attention and Adversarial Logit Pairing.
    Li X; Goodman D; Liu J; Wei T; Dou D
    Front Artif Intell; 2021; 4():752831. PubMed ID: 35156010
    [TBL] [Abstract][Full Text] [Related]  

  • 10. Gradient Correction for White-Box Adversarial Attacks.
    Liu H; Ge Z; Zhou Z; Shang F; Liu Y; Jiao L
    IEEE Trans Neural Netw Learn Syst; 2023 Oct; PP():. PubMed ID: 37819820
    [TBL] [Abstract][Full Text] [Related]  

  • 11. Crafting Adversarial Perturbations via Transformed Image Component Swapping.
    Agarwal A; Ratha N; Vatsa M; Singh R
    IEEE Trans Image Process; 2022; 31():7338-7349. PubMed ID: 36094979
    [TBL] [Abstract][Full Text] [Related]  

  • 12. DualFlow: Generating imperceptible adversarial examples by flow field and normalize flow-based model.
    Liu R; Jin X; Hu D; Zhang J; Wang Y; Zhang J; Zhou W
    Front Neurorobot; 2023; 17():1129720. PubMed ID: 36845066
    [TBL] [Abstract][Full Text] [Related]  

  • 13. Towards Transferable Adversarial Attacks on Image and Video Transformers.
    Wei Z; Chen J; Goldblum M; Wu Z; Goldstein T; Jiang YG; Davis LS
    IEEE Trans Image Process; 2023; 32():6346-6358. PubMed ID: 37966925
    [TBL] [Abstract][Full Text] [Related]  

  • 14. DEFEAT: Decoupled feature attack across deep neural networks.
    Huang L; Gao C; Liu N
    Neural Netw; 2022 Dec; 156():13-28. PubMed ID: 36228335
    [TBL] [Abstract][Full Text] [Related]  

  • 15. Diffusion Models for Imperceptible and Transferable Adversarial Attack.
    Chen J; Chen H; Chen K; Zhang Y; Zou Z; Shi Z
    IEEE Trans Pattern Anal Mach Intell; 2024 Oct; PP():. PubMed ID: 39405140
    [TBL] [Abstract][Full Text] [Related]  

  • 16. SMGEA: A New Ensemble Adversarial Attack Powered by Long-Term Gradient Memories.
    Che Z; Borji A; Zhai G; Ling S; Li J; Min X; Guo G; Le Callet P
    IEEE Trans Neural Netw Learn Syst; 2022 Mar; 33(3):1051-1065. PubMed ID: 33296311
    [TBL] [Abstract][Full Text] [Related]  

  • 17. Remix: Towards the transferability of adversarial examples.
    Zhao H; Hao L; Hao K; Wei B; Cai X
    Neural Netw; 2023 Jun; 163():367-378. PubMed ID: 37119676
    [TBL] [Abstract][Full Text] [Related]  

  • 18. Enhancing adversarial attacks with resize-invariant and logical ensemble.
    Shao Y; Zhang Y; Dong W; Zhang Q; Shan P; Guo J; Xu H
    Neural Netw; 2024 May; 173():106194. PubMed ID: 38402809
    [TBL] [Abstract][Full Text] [Related]  

  • 19. Image Adversarial Example Generation Method Based on Adaptive Parameter Adjustable Differential Evolution.
    Lin Z; Peng C; Tan W; He X
    Entropy (Basel); 2023 Mar; 25(3):. PubMed ID: 36981373
    [TBL] [Abstract][Full Text] [Related]  

  • 20. Adversarial attack vulnerability of medical image analysis systems: Unexplored factors.
    Bortsova G; González-Gonzalo C; Wetstein SC; Dubost F; Katramados I; Hogeweg L; Liefers B; van Ginneken B; Pluim JPW; Veta M; Sánchez CI; de Bruijne M
    Med Image Anal; 2021 Oct; 73():102141. PubMed ID: 34246850
    [TBL] [Abstract][Full Text] [Related]  

    [Next]    [New Search]
    of 6.