These tools will no longer be maintained as of December 31, 2024. Archived website can be found here. PubMed4Hh GitHub repository can be found here. Contact NLM Customer Service if you have questions.
2. Performance of GPT-4 with Vision on Text- and Image-based ACR Diagnostic Radiology In-Training Examination Questions. Hayden N; Gilbert S; Poisson LM; Griffith B; Klochko C Radiology; 2024 Sep; 312(3):e240153. PubMed ID: 39225605 [TBL] [Abstract][Full Text] [Related]
3. Performance of GPT-4V in Answering the Japanese Otolaryngology Board Certification Examination Questions: Evaluation Study. Noda M; Ueno T; Koshu R; Takaso Y; Shimada MD; Saito C; Sugimoto H; Fushiki H; Ito M; Nomura A; Yoshizaki T JMIR Med Educ; 2024 Mar; 10():e57054. PubMed ID: 38546736 [TBL] [Abstract][Full Text] [Related]
4. Evaluation of GPT Large Language Model Performance on RSNA 2023 Case of the Day Questions. Mukherjee P; Hou B; Suri A; Zhuang Y; Parnell C; Lee N; Stroie O; Jain R; Wang KC; Sharma K; Summers RM Radiology; 2024 Oct; 313(1):e240609. PubMed ID: 39352277 [TBL] [Abstract][Full Text] [Related]
5. Exploring the Performance of ChatGPT Versions 3.5, 4, and 4 With Vision in the Chilean Medical Licensing Examination: Observational Study. Rojas M; Rojas M; Burgess V; Toro-Pérez J; Salehi S JMIR Med Educ; 2024 Apr; 10():e55048. PubMed ID: 38686550 [TBL] [Abstract][Full Text] [Related]
6. A Comparison Between GPT-3.5, GPT-4, and GPT-4V: Can the Large Language Model (ChatGPT) Pass the Japanese Board of Orthopaedic Surgery Examination? Nakajima N; Fujimori T; Furuya M; Kanie Y; Imai H; Kita K; Uemura K; Okada S Cureus; 2024 Mar; 16(3):e56402. PubMed ID: 38633935 [TBL] [Abstract][Full Text] [Related]
7. Performance of Progressive Generations of GPT on an Exam Designed for Certifying Physicians as Certified Clinical Densitometrists. Valdez D; Bunnell A; Lim SY; Sadowski P; Shepherd JA J Clin Densitom; 2024; 27(2):101480. PubMed ID: 38401238 [TBL] [Abstract][Full Text] [Related]
8. ChatGPT performance on the American Shoulder and Elbow Surgeons maintenance of certification exam. Fiedler B; Azua EN; Phillips T; Ahmed AS J Shoulder Elbow Surg; 2024 Sep; 33(9):1888-1893. PubMed ID: 38580067 [TBL] [Abstract][Full Text] [Related]
9. Performance of ChatGPT on Nephrology Test Questions. Miao J; Thongprayoon C; Garcia Valencia OA; Krisanapan P; Sheikh MS; Davis PW; Mekraksakit P; Suarez MG; Craici IM; Cheungpasitporn W Clin J Am Soc Nephrol; 2024 Jan; 19(1):35-43. PubMed ID: 37851468 [TBL] [Abstract][Full Text] [Related]
10. Capability of GPT-4V(ision) in the Japanese National Medical Licensing Examination: Evaluation Study. Nakao T; Miki S; Nakamura Y; Kikuchi T; Nomura Y; Hanaoka S; Yoshikawa T; Abe O JMIR Med Educ; 2024 Mar; 10():e54393. PubMed ID: 38470459 [TBL] [Abstract][Full Text] [Related]
11. ChatGPT's diagnostic performance based on textual vs. visual information compared to radiologists' diagnostic performance in musculoskeletal radiology. Horiuchi D; Tatekawa H; Oura T; Shimono T; Walston SL; Takita H; Matsushita S; Mitsuyama Y; Miki Y; Ueda D Eur Radiol; 2025 Jan; 35(1):506-516. PubMed ID: 38995378 [TBL] [Abstract][Full Text] [Related]
12. Performance of ChatGPT Across Different Versions in Medical Licensing Examinations Worldwide: Systematic Review and Meta-Analysis. Liu M; Okuhara T; Chang X; Shirabe R; Nishiie Y; Okada H; Kiuchi T J Med Internet Res; 2024 Jul; 26():e60807. PubMed ID: 39052324 [TBL] [Abstract][Full Text] [Related]
13. Influence of Model Evolution and System Roles on ChatGPT's Performance in Chinese Medical Licensing Exams: Comparative Study. Ming S; Guo Q; Cheng W; Lei B JMIR Med Educ; 2024 Aug; 10():e52784. PubMed ID: 39140269 [TBL] [Abstract][Full Text] [Related]
15. GPT-4 Turbo with Vision fails to outperform text-only GPT-4 Turbo in the Japan Diagnostic Radiology Board Examination. Hirano Y; Hanaoka S; Nakao T; Miki S; Kikuchi T; Nakamura Y; Nomura Y; Yoshikawa T; Abe O Jpn J Radiol; 2024 Aug; 42(8):918-926. PubMed ID: 38733472 [TBL] [Abstract][Full Text] [Related]
16. Comparing the Diagnostic Performance of GPT-4-based ChatGPT, GPT-4V-based ChatGPT, and Radiologists in Challenging Neuroradiology Cases. Horiuchi D; Tatekawa H; Oura T; Oue S; Walston SL; Takita H; Matsushita S; Mitsuyama Y; Shimono T; Miki Y; Ueda D Clin Neuroradiol; 2024 Dec; 34(4):779-787. PubMed ID: 38806794 [TBL] [Abstract][Full Text] [Related]
17. Large Language Models Take on Cardiothoracic Surgery: A Comparative Analysis of the Performance of Four Models on American Board of Thoracic Surgery Exam Questions in 2023. Khalpey Z; Kumar U; King N; Abraham A; Khalpey AH Cureus; 2024 Jul; 16(7):e65083. PubMed ID: 39171020 [TBL] [Abstract][Full Text] [Related]
18. Evaluating GPT-4V's performance in the Japanese national dental examination: A challenge explored. Morishita M; Fukuda H; Muraoka K; Nakamura T; Hayashi M; Yoshioka I; Ono K; Awano S J Dent Sci; 2024 Jul; 19(3):1595-1600. PubMed ID: 39035269 [TBL] [Abstract][Full Text] [Related]
19. Performance of ChatGPT and Bard in self-assessment questions for nephrology board renewal. Noda R; Izaki Y; Kitano F; Komatsu J; Ichikawa D; Shibagaki Y Clin Exp Nephrol; 2024 May; 28(5):465-469. PubMed ID: 38353783 [TBL] [Abstract][Full Text] [Related]
20. Evaluation of Reliability, Repeatability, Robustness, and Confidence of GPT-3.5 and GPT-4 on a Radiology Board-style Examination. Krishna S; Bhambra N; Bleakney R; Bhayana R Radiology; 2024 May; 311(2):e232715. PubMed ID: 38771184 [TBL] [Abstract][Full Text] [Related] [Next] [New Search]