171 related articles for article (PubMed ID: 38321999)
1. Survey of transformers and towards ensemble learning using transformers for natural language processing.
Zhang H; Shafiq MO
J Big Data; 2024; 11(1):25. PubMed ID: 38321999
[TBL] [Abstract][Full Text] [Related]
2. Measurement of Semantic Textual Similarity in Clinical Texts: Comparison of Transformer-Based Models.
Yang X; He X; Zhang H; Ma Y; Bian J; Wu Y
JMIR Med Inform; 2020 Nov; 8(11):e19735. PubMed ID: 33226350
[TBL] [Abstract][Full Text] [Related]
3. Transformers-sklearn: a toolkit for medical language understanding with transformer-based models.
Yang F; Wang X; Ma H; Li J
BMC Med Inform Decis Mak; 2021 Jul; 21(Suppl 2):90. PubMed ID: 34330244
[TBL] [Abstract][Full Text] [Related]
4. RadBERT: Adapting Transformer-based Language Models to Radiology.
Yan A; McAuley J; Lu X; Du J; Chang EY; Gentili A; Hsu CN
Radiol Artif Intell; 2022 Jul; 4(4):e210258. PubMed ID: 35923376
[TBL] [Abstract][Full Text] [Related]
5. Clinical concept extraction using transformers.
Yang X; Bian J; Hogan WR; Wu Y
J Am Med Inform Assoc; 2020 Dec; 27(12):1935-1942. PubMed ID: 33120431
[TBL] [Abstract][Full Text] [Related]
6. Multi-Label Classification in Patient-Doctor Dialogues With the RoBERTa-WWM-ext + CNN (Robustly Optimized Bidirectional Encoder Representations From Transformers Pretraining Approach With Whole Word Masking Extended Combining a Convolutional Neural Network) Model: Named Entity Study.
Sun Y; Gao D; Shen X; Li M; Nan J; Zhang W
JMIR Med Inform; 2022 Apr; 10(4):e35606. PubMed ID: 35451969
[TBL] [Abstract][Full Text] [Related]
7. Protected Health Information Recognition by Fine-Tuning a Pre-training Transformer Model.
Oh SH; Kang M; Lee Y
Healthc Inform Res; 2022 Jan; 28(1):16-24. PubMed ID: 35172087
[TBL] [Abstract][Full Text] [Related]
8. Sequence-to-sequence pretraining for a less-resourced Slovenian language.
Ulčar M; Robnik-Šikonja M
Front Artif Intell; 2023; 6():932519. PubMed ID: 37056912
[TBL] [Abstract][Full Text] [Related]
9. Identify diabetic retinopathy-related clinical concepts and their attributes using transformer-based natural language processing methods.
Yu Z; Yang X; Sweeting GL; Ma Y; Stolte SE; Fang R; Wu Y
BMC Med Inform Decis Mak; 2022 Sep; 22(Suppl 3):255. PubMed ID: 36167551
[TBL] [Abstract][Full Text] [Related]
10. A multimodal approach to cross-lingual sentiment analysis with ensemble of transformer and LLM.
Miah MSU; Kabir MM; Sarwar TB; Safran M; Alfarhood S; Mridha MF
Sci Rep; 2024 Apr; 14(1):9603. PubMed ID: 38671064
[TBL] [Abstract][Full Text] [Related]
11. A Question-and-Answer System to Extract Data From Free-Text Oncological Pathology Reports (CancerBERT Network): Development Study.
Mitchell JR; Szepietowski P; Howard R; Reisman P; Jones JD; Lewis P; Fridley BL; Rollison DE
J Med Internet Res; 2022 Mar; 24(3):e27210. PubMed ID: 35319481
[TBL] [Abstract][Full Text] [Related]
12. Text classification models for the automatic detection of nonmedical prescription medication use from social media.
Al-Garadi MA; Yang YC; Cai H; Ruan Y; O'Connor K; Graciela GH; Perrone J; Sarker A
BMC Med Inform Decis Mak; 2021 Jan; 21(1):27. PubMed ID: 33499852
[TBL] [Abstract][Full Text] [Related]
13. Semantic relational machine learning model for sentiment analysis using cascade feature selection and heterogeneous classifier ensemble.
Yenkikar A; Babu CN; Hemanth DJ
PeerJ Comput Sci; 2022; 8():e1100. PubMed ID: 36262147
[TBL] [Abstract][Full Text] [Related]
14. Deep Learning Approach for Negation and Speculation Detection for Automated Important Finding Flagging and Extraction in Radiology Report: Internal Validation and Technique Comparison Study.
Weng KH; Liu CF; Chen CJ
JMIR Med Inform; 2023 Apr; 11():e46348. PubMed ID: 37097731
[TBL] [Abstract][Full Text] [Related]
15. SentiMedQAer: A Transfer Learning-Based Sentiment-Aware Model for Biomedical Question Answering.
Zhu X; Chen Y; Gu Y; Xiao Z
Front Neurorobot; 2022; 16():773329. PubMed ID: 35360832
[TBL] [Abstract][Full Text] [Related]
16. Evaluation of clinical named entity recognition methods for Serbian electronic health records.
Kaplar A; Stošović M; Kaplar A; Brković V; Naumović R; Kovačević A
Int J Med Inform; 2022 Aug; 164():104805. PubMed ID: 35653828
[TBL] [Abstract][Full Text] [Related]
17. COVID-Twitter-BERT: A natural language processing model to analyse COVID-19 content on Twitter.
Müller M; Salathé M; Kummervold PE
Front Artif Intell; 2023; 6():1023281. PubMed ID: 36998290
[TBL] [Abstract][Full Text] [Related]
18. A comparative study of pretrained language models for long clinical text.
Li Y; Wehbe RM; Ahmad FS; Wang H; Luo Y
J Am Med Inform Assoc; 2023 Jan; 30(2):340-347. PubMed ID: 36451266
[TBL] [Abstract][Full Text] [Related]
19. Explainable clinical coding with in-domain adapted transformers.
López-García G; Jerez JM; Ribelles N; Alba E; Veredas FJ
J Biomed Inform; 2023 Mar; 139():104323. PubMed ID: 36813154
[TBL] [Abstract][Full Text] [Related]
20. Integrating Multimodal Information in Large Pretrained Transformers.
Rahman W; Hasan MK; Lee S; Zadeh A; Mao C; Morency LP; Hoque E
Proc Conf Assoc Comput Linguist Meet; 2020 Jul; 2020():2359-2369. PubMed ID: 33782629
[TBL] [Abstract][Full Text] [Related]
[Next] [New Search]