123 related articles for article (PubMed ID: 38405777)
1. TopoFormer: Multiscale Topology-enabled Structure-to-Sequence Transformer for Protein-Ligand Interaction Predictions.
Chen D; Liu J; Wei GW
Res Sq; 2024 Feb; ():. PubMed ID: 38405777
[TBL] [Abstract][Full Text] [Related]
2. A topology-based network tree for the prediction of protein-protein binding affinity changes following mutation.
Wang M; Cang Z; Wei GW
Nat Mach Intell; 2020; 2(2):116-123. PubMed ID: 34170981
[TBL] [Abstract][Full Text] [Related]
3. TIDAL: Topology-Inferred Drug Addiction Learning.
Zhu Z; Dou B; Cao Y; Jiang J; Zhu Y; Chen D; Feng H; Liu J; Zhang B; Zhou T; Wei GW
J Chem Inf Model; 2023 Mar; 63(5):1472-1489. PubMed ID: 36826415
[TBL] [Abstract][Full Text] [Related]
4. Persistent Cohomology for Data With Multicomponent Heterogeneous Information.
Cang Z; Wei GW
SIAM J Math Data Sci; 2020; 2(2):396-418. PubMed ID: 34222831
[TBL] [Abstract][Full Text] [Related]
5. Persistent topological Laplacian analysis of SARS-CoV-2 variants.
Wei X; Chen J; Wei GW
J Comput Biophys Chem; 2023 Aug; 22(5):569-587. PubMed ID: 37829318
[TBL] [Abstract][Full Text] [Related]
6. Persistent topological Laplacian analysis of SARS-CoV-2 variants.
Wei X; Chen J; Guo-Wei W
ArXiv; 2023 Apr; ():. PubMed ID: 36748007
[TBL] [Abstract][Full Text] [Related]
7. Leveraging transformers-based language models in proteome bioinformatics.
Le NQK
Proteomics; 2023 Dec; 23(23-24):e2300011. PubMed ID: 37381841
[TBL] [Abstract][Full Text] [Related]
8. Integration of persistent Laplacian and pre-trained transformer for protein solubility changes upon mutation.
Wee J; Chen J; Xia K; Wei GW
ArXiv; 2023 Nov; ():. PubMed ID: 37961732
[TBL] [Abstract][Full Text] [Related]
9. Integration of persistent Laplacian and pre-trained transformer for protein solubility changes upon mutation.
Wee J; Chen J; Xia K; Wei GW
Comput Biol Med; 2024 Feb; 169():107918. PubMed ID: 38194782
[TBL] [Abstract][Full Text] [Related]
10. Transformers-sklearn: a toolkit for medical language understanding with transformer-based models.
Yang F; Wang X; Ma H; Li J
BMC Med Inform Decis Mak; 2021 Jul; 21(Suppl 2):90. PubMed ID: 34330244
[TBL] [Abstract][Full Text] [Related]
11. Transformer versus traditional natural language processing: how much data is enough for automated radiology report classification?
Yang E; Li MD; Raghavan S; Deng F; Lang M; Succi MD; Huang AJ; Kalpathy-Cramer J
Br J Radiol; 2023 Sep; 96(1149):20220769. PubMed ID: 37162253
[TBL] [Abstract][Full Text] [Related]
12. Do it the transformer way: A comprehensive review of brain and vision transformers for autism spectrum disorder diagnosis and classification.
Alharthi AG; Alzahrani SM
Comput Biol Med; 2023 Dec; 167():107667. PubMed ID: 37939407
[TBL] [Abstract][Full Text] [Related]
13. Predicting Semantic Similarity Between Clinical Sentence Pairs Using Transformer Models: Evaluation and Representational Analysis.
Ormerod M; Martínez Del Rincón J; Devereux B
JMIR Med Inform; 2021 May; 9(5):e23099. PubMed ID: 34037527
[TBL] [Abstract][Full Text] [Related]
14. Analyzing Transfer Learning of Vision Transformers for Interpreting Chest Radiography.
Usman M; Zia T; Tariq A
J Digit Imaging; 2022 Dec; 35(6):1445-1462. PubMed ID: 35819537
[TBL] [Abstract][Full Text] [Related]
15. AGL-Score: Algebraic Graph Learning Score for Protein-Ligand Binding Scoring, Ranking, Docking, and Screening.
Nguyen DD; Wei GW
J Chem Inf Model; 2019 Jul; 59(7):3291-3304. PubMed ID: 31257871
[TBL] [Abstract][Full Text] [Related]
16. Persistent spectral theory-guided protein engineering.
Qiu Y; Wei GW
Nat Comput Sci; 2023 Feb; 3(2):149-163. PubMed ID: 37637776
[TBL] [Abstract][Full Text] [Related]
17. Persistent Laplacian projected Omicron BA.4 and BA.5 to become new dominating variants.
Chen J; Qiu Y; Wang R; Wei GW
Comput Biol Med; 2022 Dec; 151(Pt A):106262. PubMed ID: 36379191
[TBL] [Abstract][Full Text] [Related]
18. Virtual screening of DrugBank database for hERG blockers using topological Laplacian-assisted AI models.
Feng H; Wei GW
Comput Biol Med; 2023 Feb; 153():106491. PubMed ID: 36599209
[TBL] [Abstract][Full Text] [Related]
19. AMMU: A survey of transformer-based biomedical pretrained language models.
Kalyan KS; Rajasekharan A; Sangeetha S
J Biomed Inform; 2022 Feb; 126():103982. PubMed ID: 34974190
[TBL] [Abstract][Full Text] [Related]
20. Macromolecular crowding: chemistry and physics meet biology (Ascona, Switzerland, 10-14 June 2012).
Foffi G; Pastore A; Piazza F; Temussi PA
Phys Biol; 2013 Aug; 10(4):040301. PubMed ID: 23912807
[TBL] [Abstract][Full Text] [Related]
[Next] [New Search]