168 related articles for article (PubMed ID: 33638635)
1. BERT4Bitter: a bidirectional encoder representations from transformers (BERT)-based model for improving the prediction of bitter peptides.
Charoenkwan P; Nantasenamat C; Hasan MM; Manavalan B; Shoombuatong W
Bioinformatics; 2021 Sep; 37(17):2556-2562. PubMed ID: 33638635
[TBL] [Abstract][Full Text] [Related]
2. iBitter-SCM: Identification and characterization of bitter peptides using a scoring card method with propensity scores of dipeptides.
Charoenkwan P; Yana J; Schaduangrat N; Nantasenamat C; Hasan MM; Shoombuatong W
Genomics; 2020 Jul; 112(4):2813-2822. PubMed ID: 32234434
[TBL] [Abstract][Full Text] [Related]
3. IUP-BERT: Identification of Umami Peptides Based on BERT Features.
Jiang L; Jiang J; Wang X; Zhang Y; Zheng B; Liu S; Zhang Y; Liu C; Wan Y; Xiang D; Lv Z
Foods; 2022 Nov; 11(22):. PubMed ID: 36429332
[TBL] [Abstract][Full Text] [Related]
4. BERT-Kcr: prediction of lysine crotonylation sites by a transfer learning method with pre-trained BERT models.
Qiao Y; Zhu X; Gong H
Bioinformatics; 2022 Jan; 38(3):648-654. PubMed ID: 34643684
[TBL] [Abstract][Full Text] [Related]
5. Umami-BERT: An interpretable BERT-based model for umami peptides prediction.
Zhang J; Yan W; Zhang Q; Li Z; Liang L; Zuo M; Zhang Y
Food Res Int; 2023 Oct; 172():113142. PubMed ID: 37689906
[TBL] [Abstract][Full Text] [Related]
6. Multi-Label Classification in Patient-Doctor Dialogues With the RoBERTa-WWM-ext + CNN (Robustly Optimized Bidirectional Encoder Representations From Transformers Pretraining Approach With Whole Word Masking Extended Combining a Convolutional Neural Network) Model: Named Entity Study.
Sun Y; Gao D; Shen X; Li M; Nan J; Zhang W
JMIR Med Inform; 2022 Apr; 10(4):e35606. PubMed ID: 35451969
[TBL] [Abstract][Full Text] [Related]
7. Predicting protein-peptide binding residues via interpretable deep learning.
Wang R; Jin J; Zou Q; Nakai K; Wei L
Bioinformatics; 2022 Jun; 38(13):3351-3360. PubMed ID: 35604077
[TBL] [Abstract][Full Text] [Related]
8. BERT-Kgly: A Bidirectional Encoder Representations From Transformers (BERT)-Based Model for Predicting Lysine Glycation Site for
Liu Y; Liu Y; Wang GA; Cheng Y; Bi S; Zhu X
Front Bioinform; 2022; 2():834153. PubMed ID: 36304324
[TBL] [Abstract][Full Text] [Related]
9. Transfer Learning for Sentiment Classification Using Bidirectional Encoder Representations from Transformers (BERT) Model.
Areshey A; Mathkour H
Sensors (Basel); 2023 May; 23(11):. PubMed ID: 37299959
[TBL] [Abstract][Full Text] [Related]
10. iBitter-Fuse: A Novel Sequence-Based Bitter Peptide Predictor by Fusing Multi-View Features.
Charoenkwan P; Nantasenamat C; Hasan MM; Moni MA; Lio' P; Shoombuatong W
Int J Mol Sci; 2021 Aug; 22(16):. PubMed ID: 34445663
[TBL] [Abstract][Full Text] [Related]
11. AMP-BERT: Prediction of antimicrobial peptide function based on a BERT model.
Lee H; Lee S; Lee I; Nam H
Protein Sci; 2023 Jan; 32(1):e4529. PubMed ID: 36461699
[TBL] [Abstract][Full Text] [Related]
12. TRP-BERT: Discrimination of transient receptor potential (TRP) channels using contextual representations from deep bidirectional transformer based on BERT.
Ali Shah SM; Ou YY
Comput Biol Med; 2021 Oct; 137():104821. PubMed ID: 34508974
[TBL] [Abstract][Full Text] [Related]
13. Deep Learning Approach for Negation and Speculation Detection for Automated Important Finding Flagging and Extraction in Radiology Report: Internal Validation and Technique Comparison Study.
Weng KH; Liu CF; Chen CJ
JMIR Med Inform; 2023 Apr; 11():e46348. PubMed ID: 37097731
[TBL] [Abstract][Full Text] [Related]
14. DTI-BERT: Identifying Drug-Target Interactions in Cellular Networking Based on BERT and Deep Learning Method.
Zheng J; Xiao X; Qiu WR
Front Genet; 2022; 13():859188. PubMed ID: 35754843
[TBL] [Abstract][Full Text] [Related]
15. Prediction of RNA-protein interactions using a nucleotide language model.
Yamada K; Hamada M
Bioinform Adv; 2022; 2(1):vbac023. PubMed ID: 36699410
[TBL] [Abstract][Full Text] [Related]
16. Meta-iAVP: A Sequence-Based Meta-Predictor for Improving the Prediction of Antiviral Peptides Using Effective Feature Representation.
Schaduangrat N; Nantasenamat C; Prachayasittikul V; Shoombuatong W
Int J Mol Sci; 2019 Nov; 20(22):. PubMed ID: 31731751
[TBL] [Abstract][Full Text] [Related]
17. A transformer architecture based on BERT and 2D convolutional neural network to identify DNA enhancers from sequence information.
Le NQK; Ho QT; Nguyen TT; Ou YY
Brief Bioinform; 2021 Sep; 22(5):. PubMed ID: 33539511
[TBL] [Abstract][Full Text] [Related]
18. iUmami-SCM: A Novel Sequence-Based Predictor for Prediction and Analysis of Umami Peptides Using a Scoring Card Method with Propensity Scores of Dipeptides.
Charoenkwan P; Yana J; Nantasenamat C; Hasan MM; Shoombuatong W
J Chem Inf Model; 2020 Dec; 60(12):6666-6678. PubMed ID: 33094610
[TBL] [Abstract][Full Text] [Related]
19. PD-BertEDL: An Ensemble Deep Learning Method Using BERT and Multivariate Representation to Predict Peptide Detectability.
Wang H; Wang J; Feng Z; Li Y; Zhao H
Int J Mol Sci; 2022 Oct; 23(20):. PubMed ID: 36293242
[TBL] [Abstract][Full Text] [Related]
20. iDNA-ABT: advanced deep learning model for detecting DNA methylation with adaptive features and transductive information maximization.
Yu Y; He W; Jin J; Xiao G; Cui L; Zeng R; Wei L
Bioinformatics; 2021 Dec; 37(24):4603-4610. PubMed ID: 34601568
[TBL] [Abstract][Full Text] [Related]
[Next] [New Search]