BIOMARKERS

Molecular Biopsy of Human Tumors

- a resource for Precision Medicine *

235 related articles for article (PubMed ID: 35523917)

  • 1. Demystifying the Black Box: The Importance of Interpretability of Predictive Models in Neurocritical Care.
    Moss L; Corsar D; Shaw M; Piper I; Hawthorne C
    Neurocrit Care; 2022 Aug; 37(Suppl 2):185-191. PubMed ID: 35523917
    [TBL] [Abstract][Full Text] [Related]  

  • 2. Opening the Black Box: The Promise and Limitations of Explainable Machine Learning in Cardiology.
    Petch J; Di S; Nelson W
    Can J Cardiol; 2022 Feb; 38(2):204-213. PubMed ID: 34534619
    [TBL] [Abstract][Full Text] [Related]  

  • 3. Opening the black box: interpretable machine learning for predictor finding of metabolic syndrome.
    Zhang Y; Zhang X; Razbek J; Li D; Xia W; Bao L; Mao H; Daken M; Cao M
    BMC Endocr Disord; 2022 Aug; 22(1):214. PubMed ID: 36028865
    [TBL] [Abstract][Full Text] [Related]  

  • 4. Artificial Intelligence and Machine Learning Applications in Critically Ill Brain Injured Patients.
    Vitt JR; Mainali S
    Semin Neurol; 2024 Jun; 44(3):342-356. PubMed ID: 38569520
    [TBL] [Abstract][Full Text] [Related]  

  • 5. Analysis and interpretability of machine learning models to classify thyroid disease.
    Akter S; Mustafa HA
    PLoS One; 2024; 19(5):e0300670. PubMed ID: 38820460
    [TBL] [Abstract][Full Text] [Related]  

  • 6. Opening the black box of AI-Medicine.
    Poon AIF; Sung JJY
    J Gastroenterol Hepatol; 2021 Mar; 36(3):581-584. PubMed ID: 33709609
    [TBL] [Abstract][Full Text] [Related]  

  • 7. Big Data/AI in Neurocritical Care: Maybe/Summary.
    Suarez JI
    Neurocrit Care; 2022 Aug; 37(Suppl 2):166-169. PubMed ID: 34966957
    [TBL] [Abstract][Full Text] [Related]  

  • 8. On the interpretability of machine learning-based model for predicting hypertension.
    Elshawi R; Al-Mallah MH; Sakr S
    BMC Med Inform Decis Mak; 2019 Jul; 19(1):146. PubMed ID: 31357998
    [TBL] [Abstract][Full Text] [Related]  

  • 9. Improving Clinical Translation of Machine Learning Approaches Through Clinician-Tailored Visual Displays of Black Box Algorithms: Development and Validation.
    Wongvibulsin S; Wu KC; Zeger SL
    JMIR Med Inform; 2020 Jun; 8(6):e15791. PubMed ID: 32515746
    [TBL] [Abstract][Full Text] [Related]  

  • 10. Algorithmic Decision-Making Based on Machine Learning from Big Data: Can Transparency Restore Accountability?
    de Laat PB
    Philos Technol; 2018; 31(4):525-541. PubMed ID: 30873341
    [TBL] [Abstract][Full Text] [Related]  

  • 11. A brief history of artificial intelligence embryo selection: from black-box to glass-box.
    Lee T; Natalwala J; Chapple V; Liu Y
    Hum Reprod; 2024 Feb; 39(2):285-292. PubMed ID: 38061074
    [TBL] [Abstract][Full Text] [Related]  

  • 12. Ethical, Legal, and Financial Considerations of Artificial Intelligence in Surgery.
    Morris MX; Song EY; Rajesh A; Asaad M; Phillips BT
    Am Surg; 2023 Jan; 89(1):55-60. PubMed ID: 35978473
    [TBL] [Abstract][Full Text] [Related]  

  • 13. Designing an Interpretability-Based Model to Explain the Artificial Intelligence Algorithms in Healthcare.
    Ennab M; Mcheick H
    Diagnostics (Basel); 2022 Jun; 12(7):. PubMed ID: 35885463
    [TBL] [Abstract][Full Text] [Related]  

  • 14. SMILE: systems metabolomics using interpretable learning and evolution.
    Sha C; Cuperlovic-Culf M; Hu T
    BMC Bioinformatics; 2021 May; 22(1):284. PubMed ID: 34049495
    [TBL] [Abstract][Full Text] [Related]  

  • 15. Explainable, trustworthy, and ethical machine learning for healthcare: A survey.
    Rasheed K; Qayyum A; Ghaly M; Al-Fuqaha A; Razi A; Qadir J
    Comput Biol Med; 2022 Oct; 149():106043. PubMed ID: 36115302
    [TBL] [Abstract][Full Text] [Related]  

  • 16. Explainable Machine Learning Framework for Image Classification Problems: Case Study on Glioma Cancer Prediction.
    Pintelas E; Liaskos M; Livieris IE; Kotsiantis S; Pintelas P
    J Imaging; 2020 May; 6(6):. PubMed ID: 34460583
    [TBL] [Abstract][Full Text] [Related]  

  • 17. MS-CPFI: A model-agnostic Counterfactual Perturbation Feature Importance algorithm for interpreting black-box Multi-State models.
    Cottin A; Zulian M; Pécuchet N; Guilloux A; Katsahian S
    Artif Intell Med; 2024 Jan; 147():102741. PubMed ID: 38184354
    [TBL] [Abstract][Full Text] [Related]  

  • 18. Balancing accuracy and interpretability of machine learning approaches for radiation treatment outcomes modeling.
    Luo Y; Tseng HH; Cui S; Wei L; Ten Haken RK; El Naqa I
    BJR Open; 2019; 1(1):20190021. PubMed ID: 33178948
    [TBL] [Abstract][Full Text] [Related]  

  • 19. Artificial Intelligence and Black-Box Medical Decisions: Accuracy versus Explainability.
    London AJ
    Hastings Cent Rep; 2019 Jan; 49(1):15-21. PubMed ID: 30790315
    [TBL] [Abstract][Full Text] [Related]  

  • 20. Precision oncology: a review to assess interpretability in several explainable methods.
    Gimeno M; Sada Del Real K; Rubio A
    Brief Bioinform; 2023 Jul; 24(4):. PubMed ID: 37253690
    [TBL] [Abstract][Full Text] [Related]  

    [Next]    [New Search]
    of 12.