152 related articles for article (PubMed ID: 36242862)
21. Pretrained Transformer Language Models Versus Pretrained Word Embeddings for the Detection of Accurate Health Information on Arabic Social Media: Comparative Study.
Albalawi Y; Nikolov NS; Buckley J
JMIR Form Res; 2022 Jun; 6(6):e34834. PubMed ID: 35767322
[TBL] [Abstract][Full Text] [Related]
22. A transformer architecture based on BERT and 2D convolutional neural network to identify DNA enhancers from sequence information.
Le NQK; Ho QT; Nguyen TT; Ou YY
Brief Bioinform; 2021 Sep; 22(5):. PubMed ID: 33539511
[TBL] [Abstract][Full Text] [Related]
23. MultiGran-SMILES: multi-granularity SMILES learning for molecular property prediction.
Jiang J; Zhang R; Zhao Z; Ma J; Liu Y; Yuan Y; Niu B
Bioinformatics; 2022 Sep; 38(19):4573-4580. PubMed ID: 35961025
[TBL] [Abstract][Full Text] [Related]
24. Generating contextual embeddings for emergency department chief complaints.
Chang D; Hong WS; Taylor RA
JAMIA Open; 2020 Jul; 3(2):160-166. PubMed ID: 32734154
[TBL] [Abstract][Full Text] [Related]
25. Relation Classification for Bleeding Events From Electronic Health Records Using Deep Learning Systems: An Empirical Study.
Mitra A; Rawat BPS; McManus DD; Yu H
JMIR Med Inform; 2021 Jul; 9(7):e27527. PubMed ID: 34255697
[TBL] [Abstract][Full Text] [Related]
26. Automatic text classification of actionable radiology reports of tinnitus patients using bidirectional encoder representations from transformer (BERT) and in-domain pre-training (IDPT).
Li J; Lin Y; Zhao P; Liu W; Cai L; Sun J; Zhao L; Yang Z; Song H; Lv H; Wang Z
BMC Med Inform Decis Mak; 2022 Jul; 22(1):200. PubMed ID: 35907966
[TBL] [Abstract][Full Text] [Related]
27. Fine-tuning of BERT Model to Accurately Predict Drug-Target Interactions.
Kang H; Goo S; Lee H; Chae JW; Yun HY; Jung S
Pharmaceutics; 2022 Aug; 14(8):. PubMed ID: 36015336
[TBL] [Abstract][Full Text] [Related]
28. Prediction of RNA-protein interactions using a nucleotide language model.
Yamada K; Hamada M
Bioinform Adv; 2022; 2(1):vbac023. PubMed ID: 36699410
[TBL] [Abstract][Full Text] [Related]
29. RadioBERT: A deep learning-based system for medical report generation from chest X-ray images using contextual embeddings.
Kaur N; Mittal A
J Biomed Inform; 2022 Nov; 135():104220. PubMed ID: 36229001
[TBL] [Abstract][Full Text] [Related]
30. LMPhosSite: A Deep Learning-Based Approach for General Protein Phosphorylation Site Prediction Using Embeddings from the Local Window Sequence and Pretrained Protein Language Model.
Pakhrin SC; Pokharel S; Pratyush P; Chaudhari M; Ismail HD; Kc DB
J Proteome Res; 2023 Aug; 22(8):2548-2557. PubMed ID: 37459437
[TBL] [Abstract][Full Text] [Related]
31. Extracting clinical named entity for pituitary adenomas from Chinese electronic medical records.
Fang A; Hu J; Zhao W; Feng M; Fu J; Feng S; Lou P; Ren H; Chen X
BMC Med Inform Decis Mak; 2022 Mar; 22(1):72. PubMed ID: 35321705
[TBL] [Abstract][Full Text] [Related]
32. Use of BERT (Bidirectional Encoder Representations from Transformers)-Based Deep Learning Method for Extracting Evidences in Chinese Radiology Reports: Development of a Computer-Aided Liver Cancer Diagnosis Framework.
Liu H; Zhang Z; Xu Y; Wang N; Huang Y; Yang Z; Jiang R; Chen H
J Med Internet Res; 2021 Jan; 23(1):e19689. PubMed ID: 33433395
[TBL] [Abstract][Full Text] [Related]
33. BERT-Promoter: An improved sequence-based predictor of DNA promoter using BERT pre-trained model and SHAP feature selection.
Le NQK; Ho QT; Nguyen VN; Chang JS
Comput Biol Chem; 2022 Aug; 99():107732. PubMed ID: 35863177
[TBL] [Abstract][Full Text] [Related]
34. BertMCN: Mapping colloquial phrases to standard medical concepts using BERT and highway network.
Kalyan KS; Sangeetha S
Artif Intell Med; 2021 Feb; 112():102008. PubMed ID: 33581833
[TBL] [Abstract][Full Text] [Related]
35. An improved multi-modal representation-learning model based on fusion networks for property prediction in drug discovery.
Wu J; Su Y; Yang A; Ren J; Xiang Y
Comput Biol Med; 2023 Oct; 165():107452. PubMed ID: 37690287
[TBL] [Abstract][Full Text] [Related]
36. IUP-BERT: Identification of Umami Peptides Based on BERT Features.
Jiang L; Jiang J; Wang X; Zhang Y; Zheng B; Liu S; Zhang Y; Liu C; Wan Y; Xiang D; Lv Z
Foods; 2022 Nov; 11(22):. PubMed ID: 36429332
[TBL] [Abstract][Full Text] [Related]
37. GT-Finder: Classify the family of glucose transporters with pre-trained BERT language models.
Ali Shah SM; Taju SW; Ho QT; Nguyen TT; Ou YY
Comput Biol Med; 2021 Apr; 131():104259. PubMed ID: 33581474
[TBL] [Abstract][Full Text] [Related]
38. Integrating Multimodal Information in Large Pretrained Transformers.
Rahman W; Hasan MK; Lee S; Zadeh A; Mao C; Morency LP; Hoque E
Proc Conf Assoc Comput Linguist Meet; 2020 Jul; 2020():2359-2369. PubMed ID: 33782629
[TBL] [Abstract][Full Text] [Related]
39. Symptom-BERT: Enhancing Cancer Symptom Detection in EHR Clinical Notes.
Zeinali N; Albashayreh A; Fan W; White SG
J Pain Symptom Manage; 2024 May; ():. PubMed ID: 38789092
[TBL] [Abstract][Full Text] [Related]
40. GeneralizedDTA: combining pre-training and multi-task learning to predict drug-target binding affinity for unknown drug discovery.
Lin S; Shi C; Chen J
BMC Bioinformatics; 2022 Sep; 23(1):367. PubMed ID: 36071406
[TBL] [Abstract][Full Text] [Related]
[Previous] [Next] [New Search]