BIOMARKERS

Molecular Biopsy of Human Tumors

- a resource for Precision Medicine *

130 related articles for article (PubMed ID: 37093718)

  • 1. NExT-OOD: Overcoming Dual Multiple-Choice VQA Biases.
    Zhang X; Zhang F; Xu C
    IEEE Trans Pattern Anal Mach Intell; 2024 Apr; 46(4):1913-1931. PubMed ID: 37093718
    [TBL] [Abstract][Full Text] [Related]  

  • 2. Reducing Vision-Answer Biases for Multiple-Choice VQA.
    Zhang X; Zhang F; Xu C
    IEEE Trans Image Process; 2023; 32():4621-4634. PubMed ID: 37556338
    [TBL] [Abstract][Full Text] [Related]  

  • 3. Advancing surgical VQA with scene graph knowledge.
    Yuan K; Kattel M; Lavanchy JL; Navab N; Srivastav V; Padoy N
    Int J Comput Assist Radiol Surg; 2024 May; ():. PubMed ID: 38780829
    [TBL] [Abstract][Full Text] [Related]  

  • 4. Counterfactual Samples Synthesizing and Training for Robust Visual Question Answering.
    Chen L; Zheng Y; Niu Y; Zhang H; Xiao J
    IEEE Trans Pattern Anal Mach Intell; 2023 Nov; 45(11):13218-13234. PubMed ID: 37368813
    [TBL] [Abstract][Full Text] [Related]  

  • 5. Bilinear pooling in video-QA: empirical challenges and motivational drift from neurological parallels.
    Winterbottom T; Xiao S; McLean A; Al Moubayed N
    PeerJ Comput Sci; 2022; 8():e974. PubMed ID: 35721409
    [TBL] [Abstract][Full Text] [Related]  

  • 6. Knowledge-Routed Visual Question Reasoning: Challenges for Deep Representation Embedding.
    Cao Q; Li B; Liang X; Wang K; Lin L
    IEEE Trans Neural Netw Learn Syst; 2022 Jul; 33(7):2758-2767. PubMed ID: 33385313
    [TBL] [Abstract][Full Text] [Related]  

  • 7. Inverse Visual Question Answering: A New Benchmark and VQA Diagnosis Tool.
    Liu F; Xiang T; Hospedales TM; Yang W; Sun C
    IEEE Trans Pattern Anal Mach Intell; 2020 Feb; 42(2):460-474. PubMed ID: 30418897
    [TBL] [Abstract][Full Text] [Related]  

  • 8. MRA-Net: Improving VQA Via Multi-Modal Relation Attention Network.
    Peng L; Yang Y; Wang Z; Huang Z; Shen HT
    IEEE Trans Pattern Anal Mach Intell; 2022 Jan; 44(1):318-329. PubMed ID: 32750794
    [TBL] [Abstract][Full Text] [Related]  

  • 9. Collaborative Modality Fusion for Mitigating Language Bias in Visual Question Answering.
    Lu Q; Chen S; Zhu X
    J Imaging; 2024 Feb; 10(3):. PubMed ID: 38535137
    [TBL] [Abstract][Full Text] [Related]  

  • 10. Loss Re-Scaling VQA: Revisiting the Language Prior Problem From a Class-Imbalance View.
    Guo Y; Nie L; Cheng Z; Tian Q; Zhang M
    IEEE Trans Image Process; 2022; 31():227-238. PubMed ID: 34847029
    [TBL] [Abstract][Full Text] [Related]  

  • 11. General Greedy De-Bias Learning.
    Han X; Wang S; Su C; Huang Q; Tian Q
    IEEE Trans Pattern Anal Mach Intell; 2023 Aug; 45(8):9789-9805. PubMed ID: 37022219
    [TBL] [Abstract][Full Text] [Related]  

  • 12. Multitask Learning for Visual Question Answering.
    Ma J; Liu J; Lin Q; Wu B; Wang Y; You Y
    IEEE Trans Neural Netw Learn Syst; 2023 Mar; 34(3):1380-1394. PubMed ID: 34460390
    [TBL] [Abstract][Full Text] [Related]  

  • 13. Medical Visual Question Answering via Conditional Reasoning and Contrastive Learning.
    Liu B; Zhan LM; Xu L; Wu XM
    IEEE Trans Med Imaging; 2023 May; 42(5):1532-1545. PubMed ID: 37015503
    [TBL] [Abstract][Full Text] [Related]  

  • 14. COIN: Counterfactual Image Generation for Visual Question Answering Interpretation.
    Boukhers Z; Hartmann T; Jürjens J
    Sensors (Basel); 2022 Mar; 22(6):. PubMed ID: 35336415
    [TBL] [Abstract][Full Text] [Related]  

  • 15. Bilateral Cross-Modality Graph Matching Attention for Feature Fusion in Visual Question Answering.
    Cao J; Qin X; Zhao S; Shen J
    IEEE Trans Neural Netw Learn Syst; 2022 Feb; PP():. PubMed ID: 35130171
    [TBL] [Abstract][Full Text] [Related]  

  • 16. Learning to Answer Visual Questions from Web Videos.
    Yang A; Miech A; Sivic J; Laptev I; Schmid C
    IEEE Trans Pattern Anal Mach Intell; 2022 May; PP():. PubMed ID: 35533174
    [TBL] [Abstract][Full Text] [Related]  

  • 17. CRIC: A VQA Dataset for Compositional Reasoning on Vision and Commonsense.
    Gao D; Wang R; Shan S; Chen X
    IEEE Trans Pattern Anal Mach Intell; 2023 May; 45(5):5561-5578. PubMed ID: 36173773
    [TBL] [Abstract][Full Text] [Related]  

  • 18. Multi-Modal Explicit Sparse Attention Networks for Visual Question Answering.
    Guo Z; Han D
    Sensors (Basel); 2020 Nov; 20(23):. PubMed ID: 33255994
    [TBL] [Abstract][Full Text] [Related]  

  • 19. Structured Multimodal Attentions for TextVQA.
    Gao C; Zhu Q; Wang P; Li H; Liu Y; Hengel AVD; Wu Q
    IEEE Trans Pattern Anal Mach Intell; 2022 Dec; 44(12):9603-9614. PubMed ID: 34855584
    [TBL] [Abstract][Full Text] [Related]  

  • 20. A Bi-level representation learning model for medical visual question answering.
    Li Y; Long S; Yang Z; Weng H; Zeng K; Huang Z; Lee Wang F; Hao T
    J Biomed Inform; 2022 Oct; 134():104183. PubMed ID: 36038063
    [TBL] [Abstract][Full Text] [Related]  

    [Next]    [New Search]
    of 7.