213 related articles for article (PubMed ID: 36632860)
1. Fine-tuning BERT for automatic ADME semantic labeling in FDA drug labeling to enhance product-specific guidance assessment.
Shi Y; Wang J; Ren P; ValizadehAslani T; Zhang Y; Hu M; Liang H
J Biomed Inform; 2023 Feb; 138():104285. PubMed ID: 36632860
[TBL] [Abstract][Full Text] [Related]
2. Information Extraction From FDA Drug Labeling to Enhance Product-Specific Guidance Assessment Using Natural Language Processing.
Shi Y; Ren P; Zhang Y; Gong X; Hu M; Liang H
Front Res Metr Anal; 2021; 6():670006. PubMed ID: 34179681
[TBL] [Abstract][Full Text] [Related]
3. A Fine-Tuned Bidirectional Encoder Representations From Transformers Model for Food Named-Entity Recognition: Algorithm Development and Validation.
Stojanov R; Popovski G; Cenikj G; Koroušić Seljak B; Eftimov T
J Med Internet Res; 2021 Aug; 23(8):e28229. PubMed ID: 34383671
[TBL] [Abstract][Full Text] [Related]
4. PharmBERT: a domain-specific BERT model for drug labels.
ValizadehAslani T; Shi Y; Ren P; Wang J; Zhang Y; Hu M; Zhao L; Liang H
Brief Bioinform; 2023 Jul; 24(4):. PubMed ID: 37317617
[TBL] [Abstract][Full Text] [Related]
5. Comparing Pre-trained and Feature-Based Models for Prediction of Alzheimer's Disease Based on Speech.
Balagopalan A; Eyre B; Robin J; Rudzicz F; Novikova J
Front Aging Neurosci; 2021; 13():635945. PubMed ID: 33986655
[No Abstract] [Full Text] [Related]
6. Extracting comprehensive clinical information for breast cancer using deep learning methods.
Zhang X; Zhang Y; Zhang Q; Ren Y; Qiu T; Ma J; Sun Q
Int J Med Inform; 2019 Dec; 132():103985. PubMed ID: 31627032
[TBL] [Abstract][Full Text] [Related]
7. Bidirectional Encoder Representations from Transformers-like large language models in patient safety and pharmacovigilance: A comprehensive assessment of causal inference implications.
Wang X; Xu X; Liu Z; Tong W
Exp Biol Med (Maywood); 2023 Nov; 248(21):1908-1917. PubMed ID: 38084745
[TBL] [Abstract][Full Text] [Related]
8. Multi-Label Classification in Patient-Doctor Dialogues With the RoBERTa-WWM-ext + CNN (Robustly Optimized Bidirectional Encoder Representations From Transformers Pretraining Approach With Whole Word Masking Extended Combining a Convolutional Neural Network) Model: Named Entity Study.
Sun Y; Gao D; Shen X; Li M; Nan J; Zhang W
JMIR Med Inform; 2022 Apr; 10(4):e35606. PubMed ID: 35451969
[TBL] [Abstract][Full Text] [Related]
9. Investigation of improving the pre-training and fine-tuning of BERT model for biomedical relation extraction.
Su P; Vijay-Shanker K
BMC Bioinformatics; 2022 Apr; 23(1):120. PubMed ID: 35379166
[TBL] [Abstract][Full Text] [Related]
10. GT-Finder: Classify the family of glucose transporters with pre-trained BERT language models.
Ali Shah SM; Taju SW; Ho QT; Nguyen TT; Ou YY
Comput Biol Med; 2021 Apr; 131():104259. PubMed ID: 33581474
[TBL] [Abstract][Full Text] [Related]
11. When BERT meets Bilbo: a learning curve analysis of pretrained language model on disease classification.
Li X; Yuan W; Peng D; Mei Q; Wang Y
BMC Med Inform Decis Mak; 2022 Apr; 21(Suppl 9):377. PubMed ID: 35382811
[TBL] [Abstract][Full Text] [Related]
12. Automatic text classification of actionable radiology reports of tinnitus patients using bidirectional encoder representations from transformer (BERT) and in-domain pre-training (IDPT).
Li J; Lin Y; Zhao P; Liu W; Cai L; Sun J; Zhao L; Yang Z; Song H; Lv H; Wang Z
BMC Med Inform Decis Mak; 2022 Jul; 22(1):200. PubMed ID: 35907966
[TBL] [Abstract][Full Text] [Related]
13. BertSRC: transformer-based semantic relation classification.
Lee Y; Son J; Song M
BMC Med Inform Decis Mak; 2022 Sep; 22(1):234. PubMed ID: 36068535
[TBL] [Abstract][Full Text] [Related]
14. Fine-Tuning Bidirectional Encoder Representations From Transformers (BERT)-Based Models on Large-Scale Electronic Health Record Notes: An Empirical Study.
Li F; Jin Y; Liu W; Rawat BPS; Cai P; Yu H
JMIR Med Inform; 2019 Sep; 7(3):e14830. PubMed ID: 31516126
[TBL] [Abstract][Full Text] [Related]
15. Adapting Bidirectional Encoder Representations from Transformers (BERT) to Assess Clinical Semantic Textual Similarity: Algorithm Development and Validation Study.
Kades K; Sellner J; Koehler G; Full PM; Lai TYE; Kleesiek J; Maier-Hein KH
JMIR Med Inform; 2021 Feb; 9(2):e22795. PubMed ID: 33533728
[TBL] [Abstract][Full Text] [Related]
16. Drug knowledge discovery via multi-task learning and pre-trained models.
Li D; Xiong Y; Hu B; Tang B; Peng W; Chen Q
BMC Med Inform Decis Mak; 2021 Nov; 21(Suppl 9):251. PubMed ID: 34789238
[TBL] [Abstract][Full Text] [Related]
17. WavBERT: Exploiting Semantic and Non-semantic Speech using Wav2vec and BERT for Dementia Detection.
Zhu Y; Obyat A; Liang X; Batsis JA; Roth RM
Interspeech; 2021; 2021():3790-3794. PubMed ID: 37063977
[TBL] [Abstract][Full Text] [Related]
18. Semantic Textual Similarity in Japanese Clinical Domain Texts Using BERT.
Mutinda FW; Yada S; Wakamiya S; Aramaki E
Methods Inf Med; 2021 Jun; 60(S 01):e56-e64. PubMed ID: 34237783
[TBL] [Abstract][Full Text] [Related]
19. Relation classification via BERT with piecewise convolution and focal loss.
Liu J; Duan X; Zhang R; Sun Y; Guan L; Lin B
PLoS One; 2021; 16(9):e0257092. PubMed ID: 34506554
[TBL] [Abstract][Full Text] [Related]
20. RxBERT: Enhancing drug labeling text mining and analysis with AI language modeling.
Wu L; Gray M; Dang O; Xu J; Fang H; Tong W
Exp Biol Med (Maywood); 2023 Nov; 248(21):1937-1943. PubMed ID: 38166420
[TBL] [Abstract][Full Text] [Related]
[Next] [New Search]