130 related articles for article (PubMed ID: 37093718)
1. NExT-OOD: Overcoming Dual Multiple-Choice VQA Biases.
Zhang X; Zhang F; Xu C
IEEE Trans Pattern Anal Mach Intell; 2024 Apr; 46(4):1913-1931. PubMed ID: 37093718
[TBL] [Abstract][Full Text] [Related]
2. Reducing Vision-Answer Biases for Multiple-Choice VQA.
Zhang X; Zhang F; Xu C
IEEE Trans Image Process; 2023; 32():4621-4634. PubMed ID: 37556338
[TBL] [Abstract][Full Text] [Related]
3. Advancing surgical VQA with scene graph knowledge.
Yuan K; Kattel M; Lavanchy JL; Navab N; Srivastav V; Padoy N
Int J Comput Assist Radiol Surg; 2024 May; ():. PubMed ID: 38780829
[TBL] [Abstract][Full Text] [Related]
4. Counterfactual Samples Synthesizing and Training for Robust Visual Question Answering.
Chen L; Zheng Y; Niu Y; Zhang H; Xiao J
IEEE Trans Pattern Anal Mach Intell; 2023 Nov; 45(11):13218-13234. PubMed ID: 37368813
[TBL] [Abstract][Full Text] [Related]
5. Bilinear pooling in video-QA: empirical challenges and motivational drift from neurological parallels.
Winterbottom T; Xiao S; McLean A; Al Moubayed N
PeerJ Comput Sci; 2022; 8():e974. PubMed ID: 35721409
[TBL] [Abstract][Full Text] [Related]
6. Knowledge-Routed Visual Question Reasoning: Challenges for Deep Representation Embedding.
Cao Q; Li B; Liang X; Wang K; Lin L
IEEE Trans Neural Netw Learn Syst; 2022 Jul; 33(7):2758-2767. PubMed ID: 33385313
[TBL] [Abstract][Full Text] [Related]
7. Inverse Visual Question Answering: A New Benchmark and VQA Diagnosis Tool.
Liu F; Xiang T; Hospedales TM; Yang W; Sun C
IEEE Trans Pattern Anal Mach Intell; 2020 Feb; 42(2):460-474. PubMed ID: 30418897
[TBL] [Abstract][Full Text] [Related]
8. MRA-Net: Improving VQA Via Multi-Modal Relation Attention Network.
Peng L; Yang Y; Wang Z; Huang Z; Shen HT
IEEE Trans Pattern Anal Mach Intell; 2022 Jan; 44(1):318-329. PubMed ID: 32750794
[TBL] [Abstract][Full Text] [Related]
9. Collaborative Modality Fusion for Mitigating Language Bias in Visual Question Answering.
Lu Q; Chen S; Zhu X
J Imaging; 2024 Feb; 10(3):. PubMed ID: 38535137
[TBL] [Abstract][Full Text] [Related]
10. Loss Re-Scaling VQA: Revisiting the Language Prior Problem From a Class-Imbalance View.
Guo Y; Nie L; Cheng Z; Tian Q; Zhang M
IEEE Trans Image Process; 2022; 31():227-238. PubMed ID: 34847029
[TBL] [Abstract][Full Text] [Related]
11. General Greedy De-Bias Learning.
Han X; Wang S; Su C; Huang Q; Tian Q
IEEE Trans Pattern Anal Mach Intell; 2023 Aug; 45(8):9789-9805. PubMed ID: 37022219
[TBL] [Abstract][Full Text] [Related]
12. Multitask Learning for Visual Question Answering.
Ma J; Liu J; Lin Q; Wu B; Wang Y; You Y
IEEE Trans Neural Netw Learn Syst; 2023 Mar; 34(3):1380-1394. PubMed ID: 34460390
[TBL] [Abstract][Full Text] [Related]
13. Medical Visual Question Answering via Conditional Reasoning and Contrastive Learning.
Liu B; Zhan LM; Xu L; Wu XM
IEEE Trans Med Imaging; 2023 May; 42(5):1532-1545. PubMed ID: 37015503
[TBL] [Abstract][Full Text] [Related]
14. COIN: Counterfactual Image Generation for Visual Question Answering Interpretation.
Boukhers Z; Hartmann T; Jürjens J
Sensors (Basel); 2022 Mar; 22(6):. PubMed ID: 35336415
[TBL] [Abstract][Full Text] [Related]
15. Bilateral Cross-Modality Graph Matching Attention for Feature Fusion in Visual Question Answering.
Cao J; Qin X; Zhao S; Shen J
IEEE Trans Neural Netw Learn Syst; 2022 Feb; PP():. PubMed ID: 35130171
[TBL] [Abstract][Full Text] [Related]
16. Learning to Answer Visual Questions from Web Videos.
Yang A; Miech A; Sivic J; Laptev I; Schmid C
IEEE Trans Pattern Anal Mach Intell; 2022 May; PP():. PubMed ID: 35533174
[TBL] [Abstract][Full Text] [Related]
17. CRIC: A VQA Dataset for Compositional Reasoning on Vision and Commonsense.
Gao D; Wang R; Shan S; Chen X
IEEE Trans Pattern Anal Mach Intell; 2023 May; 45(5):5561-5578. PubMed ID: 36173773
[TBL] [Abstract][Full Text] [Related]
18. Multi-Modal Explicit Sparse Attention Networks for Visual Question Answering.
Guo Z; Han D
Sensors (Basel); 2020 Nov; 20(23):. PubMed ID: 33255994
[TBL] [Abstract][Full Text] [Related]
19. Structured Multimodal Attentions for TextVQA.
Gao C; Zhu Q; Wang P; Li H; Liu Y; Hengel AVD; Wu Q
IEEE Trans Pattern Anal Mach Intell; 2022 Dec; 44(12):9603-9614. PubMed ID: 34855584
[TBL] [Abstract][Full Text] [Related]
20. A Bi-level representation learning model for medical visual question answering.
Li Y; Long S; Yang Z; Weng H; Zeng K; Huang Z; Lee Wang F; Hao T
J Biomed Inform; 2022 Oct; 134():104183. PubMed ID: 36038063
[TBL] [Abstract][Full Text] [Related]
[Next] [New Search]