Interpretable single-cell transcription factor prediction based on deep learning with attention mechanism. 2023

Meiqin Gong, and Yuchen He, and Maocheng Wang, and Yongqing Zhang, and Chunli Ding
West China Second University Hospital, Sichuan University, Chengdu 610041, China.

Predicting the transcription factor binding site (TFBS) in the whole genome range is essential in exploring the rule of gene transcription control. Although many deep learning methods to predict TFBS have been proposed, predicting TFBS using single-cell ATAC-seq data and embedding attention mechanisms needs to be improved. To this end, we present IscPAM, an interpretable method based on deep learning with an attention mechanism to predict single-cell transcription factors. Our model adopts the convolution neural network to extract the data feature and optimize the pre-trained model. In particular, the model obtains faster training and prediction due to the embedded attention mechanism. For datasets, we take ATAC-seq, ChIP-seq, and DNA sequences data for the pre-trained model, and single-cell ATAC-seq data is used to predict the TF binding graph in the given cell. We verify the interpretability of the model through ablation experiments and sensitivity analysis. IscPAM can efficiently predict the combination of whole genome transcription factors in single cells and study cellular heterogeneity through chromatin accessibility of related diseases.

UI MeSH Term Description Entries
D002843 Chromatin The material of CHROMOSOMES. It is a complex of DNA; HISTONES; and nonhistone proteins (CHROMOSOMAL PROTEINS, NON-HISTONE) found within the nucleus of a cell. Chromatins
D005786 Gene Expression Regulation Any of the processes by which nuclear, cytoplasmic, or intercellular factors influence the differential control (induction or repression) of gene action at the level of transcription or translation. Gene Action Regulation,Regulation of Gene Expression,Expression Regulation, Gene,Regulation, Gene Action,Regulation, Gene Expression
D000077321 Deep Learning Supervised or unsupervised machine learning methods that use multiple layers of data representations generated by nonlinear transformations, instead of individual task-specific ALGORITHMS, to build and train neural network models. Hierarchical Learning,Learning, Deep,Learning, Hierarchical
D014157 Transcription Factors Endogenous substances, usually proteins, which are effective in the initiation, stimulation, or termination of the genetic transcription process. Transcription Factor,Factor, Transcription,Factors, Transcription
D016571 Neural Networks, Computer A computer architecture, implementable in either hardware or software, modeled after biological neural networks. Like the biological system in which the processing capability is a result of the interconnection strengths between arrays of nonlinear processing nodes, computerized neural networks, often called perceptrons or multilayer connectionist models, consist of neuron-like units. A homogeneous group of units makes up a layer. These networks are good at pattern recognition. They are adaptive, performing tasks by example, and thus are better for decision-making than are linear learning machines or cluster analysis. They do not require explicit programming. Computational Neural Networks,Connectionist Models,Models, Neural Network,Neural Network Models,Neural Networks (Computer),Perceptrons,Computational Neural Network,Computer Neural Network,Computer Neural Networks,Connectionist Model,Model, Connectionist,Model, Neural Network,Models, Connectionist,Network Model, Neural,Network Models, Neural,Network, Computational Neural,Network, Computer Neural,Network, Neural (Computer),Networks, Computational Neural,Networks, Computer Neural,Networks, Neural (Computer),Neural Network (Computer),Neural Network Model,Neural Network, Computational,Neural Network, Computer,Neural Networks, Computational,Perceptron

Related Publications

Meiqin Gong, and Yuchen He, and Maocheng Wang, and Yongqing Zhang, and Chunli Ding
January 2023, Bioinformatics (Oxford, England),
Meiqin Gong, and Yuchen He, and Maocheng Wang, and Yongqing Zhang, and Chunli Ding
July 2021, Proteins,
Meiqin Gong, and Yuchen He, and Maocheng Wang, and Yongqing Zhang, and Chunli Ding
May 2024, Journal of chemical information and modeling,
Meiqin Gong, and Yuchen He, and Maocheng Wang, and Yongqing Zhang, and Chunli Ding
October 2023, Nature communications,
Meiqin Gong, and Yuchen He, and Maocheng Wang, and Yongqing Zhang, and Chunli Ding
May 2024, Journal of environmental management,
Meiqin Gong, and Yuchen He, and Maocheng Wang, and Yongqing Zhang, and Chunli Ding
January 2023, IEEE/ACM transactions on computational biology and bioinformatics,
Meiqin Gong, and Yuchen He, and Maocheng Wang, and Yongqing Zhang, and Chunli Ding
January 2022, Bioinformatics (Oxford, England),
Meiqin Gong, and Yuchen He, and Maocheng Wang, and Yongqing Zhang, and Chunli Ding
February 2022, Patterns (New York, N.Y.),
Meiqin Gong, and Yuchen He, and Maocheng Wang, and Yongqing Zhang, and Chunli Ding
January 2024, IEEE journal of biomedical and health informatics,
Meiqin Gong, and Yuchen He, and Maocheng Wang, and Yongqing Zhang, and Chunli Ding
February 2021, BMC bioinformatics,
Copied contents to your clipboard!