File Download
There are no files associated with this item.
Links for fulltext
(May Require Subscription)
- Publisher Website: 10.1093/bib/bbad409
- Scopus: eid_2-s2.0-85177472553
- PMID: 37974507
- Find via
Supplementary
- Citations:
- Appears in Collections:
Article: Efficient prediction of peptide self-assembly through sequential and graphical encoding
Title | Efficient prediction of peptide self-assembly through sequential and graphical encoding |
---|---|
Authors | |
Keywords | aggregation propensity coarse-grained molecular dynamics computational biology deep learning graph encoding self-assembly peptide sequence encoding |
Issue Date | 2023 |
Citation | Briefings in Bioinformatics, 2023, v. 24, n. 6, article no. bbad409 How to Cite? |
Abstract | In recent years, there has been an explosion of research on the application of deep learning to the prediction of various peptide properties, due to the significant development and market potential of peptides. Molecular dynamics has enabled the efficient collection of large peptide datasets, providing reliable training data for deep learning. However, the lack of systematic analysis of the peptide encoding, which is essential for artificial intelligence-assisted peptide-related tasks, makes it an urgent problem to be solved for the improvement of prediction accuracy. To address this issue, we first collect a high-quality, colossal simulation dataset of peptide self-assembly containing over 62 000 samples generated by coarse-grained molecular dynamics. Then, we systematically investigate the effect of peptide encoding of amino acids into sequences and molecular graphs using state-of-the-art sequential (i.e. recurrent neural network, long short-term memory and Transformer) and structural deep learning models (i.e. graph convolutional network, graph attention network and GraphSAGE), on the accuracy of peptide self-assembly prediction, an essential physiochemical process prior to any peptide-related applications. Extensive benchmarking studies have proven Transformer to be the most powerful sequence-encoding-based deep learning model, pushing the limit of peptide self-assembly prediction to decapeptides. In summary, this work provides a comprehensive benchmark analysis of peptide encoding with advanced deep learning models, serving as a guide for a wide range of peptide-related predictions such as isoelectric points, hydration free energy, etc. |
Persistent Identifier | http://hdl.handle.net/10722/355023 |
ISSN | 2023 Impact Factor: 6.8 2023 SCImago Journal Rankings: 2.143 |
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Liu, Zihan | - |
dc.contributor.author | Wang, Jiaqi | - |
dc.contributor.author | Luo, Yun | - |
dc.contributor.author | Zhao, Shuang | - |
dc.contributor.author | Li, Wenbin | - |
dc.contributor.author | Li, Stan Z. | - |
dc.date.accessioned | 2025-03-21T09:10:40Z | - |
dc.date.available | 2025-03-21T09:10:40Z | - |
dc.date.issued | 2023 | - |
dc.identifier.citation | Briefings in Bioinformatics, 2023, v. 24, n. 6, article no. bbad409 | - |
dc.identifier.issn | 1467-5463 | - |
dc.identifier.uri | http://hdl.handle.net/10722/355023 | - |
dc.description.abstract | In recent years, there has been an explosion of research on the application of deep learning to the prediction of various peptide properties, due to the significant development and market potential of peptides. Molecular dynamics has enabled the efficient collection of large peptide datasets, providing reliable training data for deep learning. However, the lack of systematic analysis of the peptide encoding, which is essential for artificial intelligence-assisted peptide-related tasks, makes it an urgent problem to be solved for the improvement of prediction accuracy. To address this issue, we first collect a high-quality, colossal simulation dataset of peptide self-assembly containing over 62 000 samples generated by coarse-grained molecular dynamics. Then, we systematically investigate the effect of peptide encoding of amino acids into sequences and molecular graphs using state-of-the-art sequential (i.e. recurrent neural network, long short-term memory and Transformer) and structural deep learning models (i.e. graph convolutional network, graph attention network and GraphSAGE), on the accuracy of peptide self-assembly prediction, an essential physiochemical process prior to any peptide-related applications. Extensive benchmarking studies have proven Transformer to be the most powerful sequence-encoding-based deep learning model, pushing the limit of peptide self-assembly prediction to decapeptides. In summary, this work provides a comprehensive benchmark analysis of peptide encoding with advanced deep learning models, serving as a guide for a wide range of peptide-related predictions such as isoelectric points, hydration free energy, etc. | - |
dc.language | eng | - |
dc.relation.ispartof | Briefings in Bioinformatics | - |
dc.subject | aggregation propensity | - |
dc.subject | coarse-grained molecular dynamics | - |
dc.subject | computational biology | - |
dc.subject | deep learning | - |
dc.subject | graph encoding | - |
dc.subject | self-assembly peptide | - |
dc.subject | sequence encoding | - |
dc.title | Efficient prediction of peptide self-assembly through sequential and graphical encoding | - |
dc.type | Article | - |
dc.description.nature | link_to_subscribed_fulltext | - |
dc.identifier.doi | 10.1093/bib/bbad409 | - |
dc.identifier.pmid | 37974507 | - |
dc.identifier.scopus | eid_2-s2.0-85177472553 | - |
dc.identifier.volume | 24 | - |
dc.identifier.issue | 6 | - |
dc.identifier.spage | article no. bbad409 | - |
dc.identifier.epage | article no. bbad409 | - |
dc.identifier.eissn | 1477-4054 | - |