- Research
- Open access
- Published:
CacPred: a cascaded convolutional neural network for TF-DNA binding prediction
BMC Genomics volume 26, Article number: 264 (2025)
Abstract
Background
Transcription factors (TFs) regulate the genes’ expression by binding to DNA sequences. Aligned TFBSs of the same TF are seen as cis-regulatory motifs, and substantial computational efforts have been invested to find motifs. In recent years, convolutional neural networks (CNNs) have succeeded in TF-DNA binding prediction, but existing DL methods’ accuracy needs to be improved and convolution function in TF-DNA binding prediction should be further explored.
Results
We develop a cascaded convolutional neural network model named CacPred to predict TF-DNA binding on 790 Chromatin immunoprecipitation-sequencing (ChIP-seq) datasets and seven ChIP-nexus (chromatin immunoprecipitation experiments with nucleotide resolution through exonuclease, unique barcode, and single ligation) datasets. We compare CacPred to six existing DL models across nine standard evaluation metrics. Our results indicate that CacPred outperforms all comparison models for TF-DNA binding prediction, and the average accuracy (ACC), matthews correlation coefficient (MCC), and the area of eight metrics radar (AEMR) are improved by 3.3%, 9.2%, and 6.4% on 790 ChIP-seq datasets. Meanwhile, CacPred improves the average ACC, MCC, and AEMR of 5.5%, 16.8%, and 12.9% on seven ChIP-nexus datasets. To explain the proposed method, motifs are used to show features CacPred learned. In light of the results, CacPred can find some significant motifs from input sequences.
Conclusions
This paper indicates that CacPred performs better than existing models on ChIP-seq data. Seven ChIP-nexus datasets are also analyzed, and they coincide with results that our proposed method performs the best on ChIP-seq data. CacPred only is equipped with the convolutional algorithm, demonstrating that pooling processing of the existing models leads to losing some sequence information. Some significant motifs are found, showing that CacPred can learn features from input sequences. In this study, we demonstrate that CacPred is an effective and feasible model for predicting TF-DNA binding. CacPred is freely available at https://github.com/zhangsq06/CacPred.
Background
Transcription factors (TFs) act as a crucial role in gene expression, cellular processes and transcriptional regulatory networks by binding to transcription factor binding sites (TFBSs) [1, 2]. TF-DNA binding prediction and identifying TFBSs are fundamental challenges for revealing the regulatory mechanisms of TFs and TF’s cooperation [3, 4]. Chromatin immunoprecipitation-sequencing (ChIP-seq) combines ChIP technology and high-throughput sequencing to obtain a TF binding region on genomic sequences. Meanwhile, ChIP-nexus combines exonuclease, specific barcodes, and a single ligation step, and adds an efficient DNA self-circularization step in the library preparation process that achieves a single nucleotide resolution [5]. TFBSs are short and conserved sequences, which increases the difficulty of position prediction via computational methods [6]. The aligned TFBSs are a regulatory motif, which can be represented by the position weight matrix (PWM) [7]. Although some public databases contain documented motifs, lots of unknown motifs and potential TF regulatory mechanisms need to be discovered.
Substantial computational efforts have been invested in predicting TF-DNA binding and finding motifs [8]. For example, MEME-ChIP employed expectation–maximization and DREME to discover Ab initio motifs [9], and gkm-SVM ultilized gap-ker and support vector machine (SVM) to combine multiple similar k-mers into more interpretable PWMs [10]. Because of the complexity of the TF binding mechanisms and the generation of large-scale genomic sequencing data, these models are difficult to handle large-scale data and reveal complex regulatory mechanisms of TFs.
Deep learning (DL) algorithms including convolutional neural networks (CNNs) [11, 12], recurrent neural networks (RNNs) [13, 14], and deep belief networks (DBNs) [15], have exhibited tremendous progress and obtained record-breaking performance in biological applications including cancer classification [16, 17], protein model quality assessment [18,19,20], and lesion recognition [21]. Meanwhile, previous research has proved DL is a feasible method for TF-DNA binding prediction and motif finding [20, 22]. DeepBind, a method proposed by Alipanahi et al. in 2015, was the first DL model to utilize a CNN to find DNA motifs [23]. DeepBind employed the convolutional kernels as motif detectors, which gives new insight into motif finding. DeepBind achieved better performance than MEME-ChIP and gkm-SVM [23]. Inspired by DeepBind, more DL models are developed for TF-DNA binding prediction and motif finding, such as DeeperBind, Basset, DeepHistone, and TBiNet, et al. Especially, DeeperBind employed CNNs and RNNs, which was developed based on DeepBind model. Among all DL models, a recently developed DL model, named DESSO, is the first model to utilize features of DNA shape (HelT, MGW, ProT, and Roll) and combine the binomial hypothetical test with CNNs to find DNA sequence and shape motifs [20]. Because RNN can capture the information within sequences, RNN is also used to find motifs and predict TF-DNA binding [24,25,26,27,28]. Our previous research assessed 20 DL models across 871 ChIP-seq datasets and defined an area of eight metrics radar (AEMR) score to evaluate the performance of these models [20]. The existing 20 DL models all employed convolutional layer, which demonstrates that convolution plays a critical role in TF-DNA binding predicting and motif finding. Our results indicated that DeepHistone is the top model for sequence classification, and DESSO is the top model for motif finding, respectively [29]. Through the previous research, we found that DL methods have a great advantage over the traditional methods. Meanwhile, we also found that the convolution’s function should be further explored and existing model’s accuracy need to be improved.
This paper proposes a cascaded convolutional neural network (CacPred) for TF-DNA binding prediction. Based on our previous research, six competitive models are selected as comparison models. In addition, evaluation metrics including precision, recall, F1_score, accuracy (ACC), specificity, Matthews correlation coefficient (MCC), area under the receiver operating characteristic curve (AUC) and area under the precision-recall curve (PRC) are used to assess DL models’ performance. Meanwhile, AEMR is selected as an overall score to rank all DL models. First, the CacPred model is assessed on 790 ChIP-seq datasets covering 261 TFs, and CacPred improves the average ACC of 3.3%, MCC of 9.2%, and AEMR of 6.4%, in predicting TF-DNA binding. Then, to verify the generalization of CacPred, CacPred is tested on seven ChIP-nexus datasets and improves the average ACC, MCC, and AEMR of 5.5%, 16.8%, and 12.9%. To explain the CacPred model, motifs are used to represent the features the CacPred learned. Our results demonstrate that motifs CacPred found are significant by comparing them to the motif database (HOCOMOCO.v11).
Results
Datasets and preprocessing
The experimental data includes 790 ChIP-seq datasets covering 261 TFs and seven ChIP-nexus datasets (Table. S1) covering seven TFs in this paper [30]. The 790 ChIP-seq datasets contain 690 ENCODE ChIP-seq datasets covering 161 TFs and 100 ChIP-seq datasets of Cistrome database covering 100 TFs [31]. All sequences in each sub-dataset are fixed with 1,001bps around their centers, and ranked in the decreasing order of original signal scores, which are all positive samples with label ‘1’. For a sub-dataset, we define a sequence to be a negative sample, which has matched GC-content to a positive sample and doesn’t overlap with any peaks in positive samples. So, the ratio of positive and negative samples is 1:1. This paper selects negative samples with the 1,001-bp-long sequences from the human genome, which are all negative sample with the label ‘0’. Each negative sequence is labeled as ‘0’, meaning that TFs can't bind to them. CacPred needs two inputs, i.e. forward sequence and reverse complementary sequence, each of which must be binary vectors. So, each sequence is encoded as a \(M = 4 \times 1001\) matrix, i.e. A = [1, 0, 0, 0], G = [0, 1, 0, 0], C = [0, 0, 1, 0], T = [0, 0, 0, 1].
Experimental setup
CacPred is optimized to minimize the average loss from the BCEloss by the Adadelta algorithm [32, 33]. To avoid the overfitting issue, dropout is used in CacPred model [34]. For a sub-dataset, 80% of samples are set as a training set, and 20% of samples are set as a testing set. The hyper-parameters contain dropout ratio, batch size, and learning rate in our experiments, which are optimized by three-fold cross-validation on the training set. Epochs of training were set to 20, the AUC of the validation set is calculated. When the AUC is highest in the validation set, hyper-parameters are saved and applied to the testing data. The CacPred model is implemented by Pytorch [35]. This study selects Basset [36], DeepHistone [22], DESSO, DeepBind, DeeperBind, TBiNet [10] as comparison models, based on previous research. The metrics that contain precision, recall, F1_score, ACC, specificity, MCC, AUC, PRC, and AEMR (formula 2) are used to assess models’ performance. To explain CacPred, motifs are used to show features that CacPred learned. The workflow of the experimental setting is shown in Fig. 1.
Performance on 790 ChIP-seq datasets
CacPred model is compared to six existing models on 790 ChIP-seq datasets. We calculate AEMRs of all models on the testing data and rank them by the average AEMR scores. Figure 2 shows the CacPred model obtains an AEMR score of 2.49 (Fig. 2g), outperforms DeepHistone under the AEMR score, and improves the AEMR score by 6.4%. DeepBind model is the first model to find motifs and predict the TF-DNA binding, which obtains the AEMR of 1.75. TBiNet obtains the AEMR of 1.52, which is lower than DeepBind model.
CacPred yields the highest score under precision, recall, F1_score, specificity, ACC, MCC, AUC, and PRC metrics, and achieves the average 0.923, 0.922, 0.942, 0.932, 0.945, 0.915, 0.965, and 0.964 scores, respectively (Fig. 3). And CacPred improves the MCC score of 9.2% in our evaluation. In light of our results, the performance of DL models except DeepBind is consistent. DeepBind yields a higher score than TBiNet under precision, recall, F1_score, specificity, ACC, MCC, and AUC, but the PRC score of DeepBind is lower than TBiNet. Figure 3 also gives the standard deviations (STD) of all models across precision, recall, F1_score, specificity, ACC, MCC, AUC, and PRC. CacPred achieves the lowest STD of precision, recall, F1_score, ACC, MCC, AUC, and PRC than others. Furthermore, TBiNet obtains the highest STD of eight metrics, which demonstrates the stability of TBiNet needs to be improved. Further, this study tries to validate CacPred on cross-cell type TF binding data and select the ChIP-seq data of ETS1 from 690 ENCODE ChIP-seq datasets. The CacPred is trained on K562 of ETS1 and is tested on GM12878 and K549 of ETS1. Our results show that CacPred outperforms the comparison models on the AEMR score (Table. S2).
Validating CacPred on ChIP-nexus datasets
To further test the performance of CacPred, all models are trained and tested on the seven ChIP-nexus datasets in the above way. Based on our results, CacPred outperforms all comparison models (Table 1). CacPred obtains the highest score under precision, recall, F1_score, specificity, ACC, MCC, AUC, and PRC metrics, and achieves the average 0.98, 0.98, 0.97, 0.96, 0.98, 0.97, 0.98, and 0.97 scores, respectively. For the AEMR score, CacPred also obtains the highest score of 2.35, which improves the AEMR by 12.9%. Meanwhile, this paper takes the dataset numbered GSM407277 as an example to show the PRC and AUC curve, the CacPred model achieves the highest PRC and AUC of 0.988 and 0.989 (Fig. 4).
Explaining CacPred model
To explain CacPred model, we show features CacPred model learned in a visualized way. CacPred learns features of DNA sequences by the first layer. Each convolutional kernel in the first layer is seen as a motif detector. For each sub-dataset, the forward DNA sequences are fed to the trained CacPred and calculate the outputs of the first layer. Each value of the outputs can represent the importance of each fragment of the input sequence, and the length of each fragment is equal to the width of the convolutional kernel. This study then selects a maximum value of the vector as the activated score (> 0) to obtain the activated sequence (a fragment of the input sequence). After that, this study aligns the set of activated sequences and counts nucleotide occurrences in the set of aligned activated sequences to obtain PWMs. Finally, the underlying TFs’ binding motifs are identified by querying the HOCOMOCO v11 database via the TOMTOM v5.1.0 tool [37]. The matched motifs are significant if their P-values are less than 0.05. The dataset wgEncodeEH001833 is taken as an example (Fig. 5), and five significant motifs are found and visualized by WebLogo [38].
Discussion
The TF-DNA binding prediction and motif finding are key steps to analyzing and understanding TFs’ functions. This study proposes a cascaded convolutional neural network model (CacPred) for predicting TF-DNA binding and finding motifs. Existing DL models are selected as comparison models, which are compared with CacPred. Our evaluation metrics contain the precision, recall, F1_score, specificity, ACC, MCC, AUC, and PRC, and they are summed as the AEMR score. All comparison models are tested across 790 ENCODE ChIP-seq data and seven ChIP-nexus data. CacPred achieves all the highest metrics, among which the average ACC, MCC, and AEMR are improved by 3.3%, 9.2%, and 6.4% on ChIP-seq data. And the CacPred improves the average ACC, MCC, and AEMR of 5.5%, 16.8%, and 12.9% on ChIP-nexus data. In this study, convolutional algorithms and pooling are applied to all the comparison models, but our proposed method only employs the convolutional algorithms. CacPred achieves the best performance, we reason that the pooling process may lose some sequence information while models are trained. Meanwhile, forward sequence and reverse complementary sequence are fed to CacPred, which provides more sequence information than one of them.
Meanwhile, this study further explores models’ ability to predict TF-DNA binding on cross-cell type ChIP-seq data, and our results show that CacPred also achieved the highest AEMR. To the best of our knowledge, Transcription factor binding preference is easily influenced by different cell types. Considering cross-cell type ChIP-seq data, researchers need to develop new DL frameworks fusing the characterization of different cell types. To interpret CacPred model, some significant motifs are used to show the features CacPred learned, which demonstrates CacPred can automatically learn meaningful features from input sequences.
In this study, CacPred is trained on each sub-dataset, which is suitable for binary classification of each sub-dataset. If we want to apply CacPred to multiple classification on all datasets, CacPred needs to be improved. And owing to the limitation of the width of convolutional kernels, CacPred only can identify TFBS with fixed length. So we should develop convolutional kernels with varied widths to identify TFBS with different lengths. Most of TFs directly bind to the DNA sequences, but TFs sometimes bind indirectly to motifs of other TFs. Assay for Transposase-Accessible Chromatin using sequencing (ATAC-seq) data can simultaneously detect hundreds of TF motif occurrences, but de novo motif discovery tools for ATAC-seq data are lacking [39]. ATAC-seq provides more information to reveal cooperative TF interactions, but the existing models limit the ability to learn motif syntax that promotes TF cooperativity. In recent years, more and more attention has been paid to the expansion of DL methods on graphs. The ideas of CNNs, RNNs, and encoders are applied to graphs, so graph neural networks (GNNs) are developed [40]. GNNs contain graph convolutional networks (GCNs) [41], graph attention networks [42], graph autoencoders [43], etc., which have successes in gene–gene interactions [44]. Considering the successful application of GNNs, this paper infers that GNNs have great potential in motif finding and revealing TFs’ cooperativity.
Conclusions
This paper introduced CacPred, which utilized cascaded CNN to predict TF-DNA binding from ChIP-seq and ChIP-nexus data. The CacPred significantly improved the AEMR score compared with existing models. And owing to the limitation of the width of convolutional kernels, CacPred only can identify TFBS with fixed length. However, CacPred only employed the convolutional algorithm, and the existing models used convolution and pooling. In light of the experimental results, the existing models may lose some important information in pooling processing. In this study, we demonstrate that CacPred is an effective and feasible model for predicting TF-DNA binding. CacPred is also a potential tool for the other classification tasks in bioinformatics.
Methods
In this paper, we develop a DL framework named CacPred for TF-DNA binding prediction and finding motifs (Fig. 6). CacPred consists of six layers, i.e., a convolutional layer, a transposed convolutional layer, a combined layer, two concatenated convolutional layers, and a fully connected layer (Fig. 5). CacPred utilized the forward sequences and reverse complementary sequences as inputs, where each input sequence is encoded into a \(M = 4 \times 1001\) matrix. The first layer employs two different convolutional layers with 4*16 convolutional kernels to accept the forward sequences and the reverse complementary sequences, and they contain 32 convolutional kernels without sharing parameters respectively. The output of the first layer can be given by formulas (3) and (4).
where, \(M_{1}\) and \(M_{2}\) represent forward sequences and reverse complementary sequences encoded matrix, respectively. The \(conv_{11} ( \cdot )\) and \(conv_{12} ( \cdot )\) represent a convolutional layer of the first layer respectively; \({\text{Re}} LU( \cdot )\) represents the rectified linear unit function; \(C_{11}\) and \(C_{12}\) are two outputs of the first layer.
The framework of the CacPred model. The CacPred needs the forward sequences and reverse complementary sequences as inputs, which are encoded into a matrix, respectively. Each red and cadet blue wide-sided rectangle represents a convolution; Each red and cadet blue narrow-sided represents a transposed convolutional algorithm; each green wide-sided rectangle represents a concatenated convolution; σ is the sigmoid function
The second layer used two transposed convolutional layers, each of them containing 32 convolutional kernels of size 1*16 and a stride of 1. The output of the second layer can be given by formulas (5) and (6).
where \(trans\_conv_{21} ( \cdot )\) and \(trans\_conv_{22} ( \cdot )\) are transposed convolutional layers; \(C_{21}\) and \(C_{22}\) represents the outputs of \(trans\_conv_{21} ( \cdot )\) and \(trans\_conv_{22} ( \cdot )\) respectively.
The third layer is a combined layer, which combines the outputs of the second layer as a matrix, and the output of the third layer can be given by formula (7).
The fourth and fifth layers are concatenated convolutional layers with the convolutional algorithm. The fourth layer employs eight convolutional kernels of size 1*15, and the fifth layer employs 16 convolutional kernels of size 1*15. The outputs of the fourth and fifth layers are given by formula (8) and formula (9).
The sixth layer of CacPred is a fully connected layer with 1,001 neurons and the output can be defined as:
where \(\hat{y}\) represents the output of CacPred; \(\omega\) represents the weight matrix; \(b\) represents the bias.
CacPred selects the Binary Cross Entropy as the loss function (BCELoss):
where \(y\) represents the true label; the \(\log ( \cdot )\) represents the logarithmic function.
Data availability
All data analyzed can be downloaded from http://bmbl.sdstate.edu/DESSO/ and http://cistrome.org/db/#/. The accession number of seven ChIP-nexus datasets is listed in the supplementary material Table S2. The source code of the manuscript is available at https://github.com/zhangsq06/CacPred.
Abbreviations
- TFs:
-
Transcription factors
- DL:
-
Deep learning
- CNN:
-
Convolutional neural network
- ChIP-seq:
-
Chromatin immunoprecipitation-sequencing
- ChIP-nexus:
-
Chromatin immunoprecipitation experiments with nucleotide resolution through exonuclease, unique barcode, and single ligation
- ACC:
-
Accuracy
- MCC:
-
Matthews correlation coefficient
- AEMR:
-
Area of eight metrics radar
- TFBSs:
-
Transcription factor binding sites
- PWM:
-
Position weight matrix
- SVM:
-
Support vector machine
- RNNs:
-
Recurrent neural networks
- DBNs:
-
Deep belief networks
- AUC:
-
Area under the receiver operating characteristic curve
- PRC:
-
Area under the precision-recall curve
- STD:
-
Standard deviations
- BCELoss:
-
Binary cross entropy loss function
- GNNs:
-
Graph neural networks
- GCNs:
-
Graph convolutional networks
References
Chen CY, Chen ST, Fuh CS, Juan HF, Huang HC. Coregulation of transcription factors and microRNAs in human transcriptional regulatory network. Bmc Bioinform. 2011;12(Suppl 1):1–10.
Bannister A., Miska, E. Regulation of gene expression by transcription factor acetylation. CMLS, Cell Mol Life Sci. 2000;57(8-9):1184–92.
Delgado F M, Gómez-Vela F. Computational methods for gene regulatory networks reconstruction and analysis: a review[J]. Artif Intell Med. 2019;95:133–45.
Nie CS, Xiao. Cooperative binding of transcription factors in the human genome. Genomics. 2020; 112(5):3427–34.
He Q, Johnston J, Zeitlinger J. ChIP-nexus enables improved detection of in vivo transcription factor binding footprints. Nat Biotechnol. 2015;33(4):395–401.
Zhou J, Lu Q, Gui L, Xu R, Wang H. MTTFsite: Cross-cell-type TF Binding Site Prediction by using Multi-task Learning. Bioinform. 2019;35(24):5067–77.
Wright H, Cohen A, Sönmez K, Yochum G, Mcweeney S. Occupancy Classification of Position Weight Matrix-Inferred Transcription Factor Binding Sites. PLoS ONE. 2011;6(11): e26160.
Guo Y, Tian K, Zeng H, Guo X, Gifford DK. A novel k-mer set memory (KSM) motif representation improves regulatory variant prediction. Genome Res. 2018;28(6):891–900.
Bailey TL. MEME-ChIP: motif analysis of large DNA datasets. Bioinformatics. 2011;27(12):1696–7.
Park S, Koh Y, Jeon H, Kim H, Kang J. Enhancing the interpretability of transcription factor binding site prediction using attention mechanism. Sci Rep. 2020;10(1):1–10.
Xu M, Ning C, Chen T, Rui J. DeepEnhancer: Predicting enhancers by convolutional neural networks. 2016 IEEE International Conference on Bioinformatics and Biomedicine (BIBM). IEEE; 2016. p. 637–644.
Xin G, Jie Z, Zhi W, Hakonarson H. DeepPolyA: A Convolutional Neural Network Approach for Polyadenylation Site Prediction. IEEE Access. 2018;6(99):24340–9.
Graves A, Mohamed AR, Hinton G. Speech Recognition with Deep Recurrent Neural Networks. IEEE International Conference on Acoustics. 2013. p. 6645–6649.
Pollastri G, Przybylski D, Rost B, Baldi P. Improving the prediction of protein secondary structure in three and eight classes using recurrent neural networks and profiles. Proteins-structure Function & Bioinformatics. 2010;47(2):228–35.
Pan X, Shen HB. RNA-protein binding motifs mining with a new hybrid deep learning based cross-domain knowledge integration approach. BMC Bioinformatics. 2017;18(1):136.
Johnson JM, Khoshgoftaar TM. Survey on deep learning with class imbalance[J]. J Big Data. 2019;6(1):1–54.
Gerazov B, Conceicao R C. Deep learning for tumour classification in homogeneous breast tissue in medical microwave imaging[C]//IEEE EUROCON 2017-17th International Conference on Smart Technologies. IEEE; 2017. p. 564–569.
Jing X, Xu J. Improved protein model quality assessment by integrating sequential and pairwise features using deep learning[J]. Bioinformatics. 2020;36(22-23):5361–7.
Nguyen SP, Yi S, Dong X. DL-PRO: A novel deep learning method for protein model quality assessment. IEEE. 2014:2071–8.
Cao R, Bhattacharya D, Hou J, et al. DeepQA: improving the estimation of single protein model quality with deep belief networks[J]. BMC Bioinformatics. 2016;17:1–9.
Seeja RD, Suresh A. Deep learning based skin lesion segmentation and classification of melanoma using support vector machine (SVM)[J]. Asian Pac J Cancer Prev: APJCP. 2019;20(5):1555.
Yin Q, Wu M, Liu Q, Lv H, Jiang R. DeepHistone: a deep learning approach to predicting histone modifications. BMC Genomics. 2019;20(S2):11–23.
Alipanahi B, Delong A, Weirauch MT, Frey BJ. Predicting the sequence specificities of DNA- and RNA-binding proteins by deep learning. Nat Biotechnol. 2015;33(8):831–8.
Quang D, Xie X. FactorNet: A deep learning framework for predicting cell type specific transcription factor binding from nucleotide-resolution sequential data. Methods. 2019;166:40–7.
Hassanzadeh HR, Wang M. DeeperBind: Enhancing Prediction of Sequence Specificities of DNA Binding Proteins. IEEE International conference on bioinformatics and biomedicine (BIBM). IEEE; 2016. p. 178–183.
Angermueller C, Lee HJ, Reik W, Stegle O. DeepCpG: accurate prediction of single-cell DNA methylation states using deep learning. Genome Biol. 2017;18(1):1–13.
Daniel Q, Xie X. DanQ: a hybrid convolutional and recurrent deep neural network for quantifying the function of DNA sequences. Nucleic Acids Res. 2016;11:e107–e107.
Zhang Q, Shen Z, Huang DS. Modeling in-vivo protein-DNA binding by combining multiple-instance learning with a hybrid deep neural network. Sci Rep. 2019;9(1):1–10.
Zhang S, Ma A, Zhao J, Xu D, Ma Q, Wang Y. Assessing deep learning methods in cis-regulatory motif finding based on genomic sequencing data. Brief Bioinform. 2021;23(1):1–10.
Auerbach RK, Chen B, Butte AJ. Relating genes to function: identifying enriched transcription factors using the ENCODE ChIP-Seq significance tool. Bioinformatics. 2013;29(15):1922–4.
Mei S, Qin Q, Wu Q, Sun H, Zheng R, Zang C, Zhu M, Wu J, Shi X, Taing L. Cistrome Data Browser: a data portal for ChIP-Seq and chromatin accessibility data in human and mouse. Nucleic Acids Res. 2016;45(1):D658–D662.
Wang Y, Zhang S, Yang L, Yang S, Ma Q. Measurement of Conditional Relatedness Between Genes Using Fully Convolutional Neural Network. Front Genet. 2019;10:1009.
Tang S, Shen C, Wang D, et al. Adaptive deep feature learning network with Nesterov momentum and its application to rotating machinery fault diagnosis[J]. Neurocomputing. 2018;305:1–14.
Hemke R, Buckless CG, Tsao A, Wang B, Torriani M. Deep learning for automated segmentation of pelvic muscles, fat, and bone from CT studies for body composition assessment. Skeletal Radiol. 2020;49(3):387–95.
Mishra P. Introduction to PyTorch, Tensors, and Tensor Operations[M]//PyTorch Recipes: A Problem-Solution Approach to Build, Train and Deploy Neural Network Models. Berkeley, CA: Apress; 2022. p. 1–28.
Kelley DR, Snoek J, Rinn JL. Basset: Learning the regulatory code of the accessible genome with deep convolutional neural networks. Genome Res. 2016:gr.200535.200115.
Bailey TL, Boden M, Buske FA, Frith M, Grant CE, Clementi L, Ren J, Li WW, Noble WS: MEME SUITE: tools for motif discovery and searching. Nucleic acids res. 2009, 37(suppl_2):W202-W208.
Crooks G. E: WebLogo: A Sequence Logo Generator. Genome Res. 2004;14(6):1188–90.
Miskimen KL, Chan ER, Haines JL. Assay for Transposase‐Accessible Chromatin Using Sequencing (ATAC‐seq) Data Analysis. Curr Protocols in Human Genet. 2017:92(1):20.24. 21–20.24. 13.
Li X, Zhou Y, Dvornek NC, et al. Pooling regularized graph neural network for fmri biomarker analysis[C]//Medical Image Computing and Computer Assisted Intervention–MICCAI 2020: 23rd International Conference. Springer International Publishing; 2020. p. 625–635.
Abu-El-Haija S, Kapoor A, Perozzi B, et al. N-gcn: Multi-scale graph convolution for semi-supervised node classification[C]//uncertainty in artificial intelligence. PMLR; 2020. p. 841–851.
Velickovic P, Cucurull G, Casanova A, et al. Graph attention networks[J]. Stat. 2017;1050(20):10–48550.
Salha G, Hennequin R, Remy JB, et al. Fastgae: Scalable graph autoencoders with stochastic subgraph decoding [J]. Neural Netw. 2021;142:1–19.
Ye Y, Bar-Joseph Z. GCNG: graph convolutional networks for inferring gene interaction from spatial transcriptomics data. Genome Biol. 2020;21(1):1–16.
Acknowledgements
Not applicable.
About this supplement
This article has been published as part of BMC Genomics, Volume 26 Supplement 2, 2025: 17th International Symposium on Bioinformatics Research and Applications. The full contents of the supplement are available at https://biomedcentral-bmcgenomics.publicaciones.saludcastillayleon.es/articles/supplements/volume-26-supplement-2.
Funding
This work was supported by the National Natural Science Foundation of China (Nos. 62302218, 62072212), the Development Project of Jilin Province of China (Nos. 20220508125RC, 2020C003). This work was also supported by Jilin Province Key Laboratory of Big Data Intelligent Computing (No. 20180622002JC).
Author information
Authors and Affiliations
Contributions
SZ conceived the project. SZ and XX collected the data and performed the experiments. YW and AM designed the study. SZ and ZL wrote the manuscript. All authors read and approved the final manuscript for publication.
Corresponding authors
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent to publication
Not applicable.
Competing interests
The authors declare that there is no conflict of interest regarding the publication of this paper.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Zhang, S., Ma, A., Xie, X. et al. CacPred: a cascaded convolutional neural network for TF-DNA binding prediction. BMC Genomics 26 (Suppl 2), 264 (2025). https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12864-025-11399-y
Received:
Accepted:
Published:
DOI: https://doiorg.publicaciones.saludcastillayleon.es/10.1186/s12864-025-11399-y