 Research
 Open Access
 Published:
A multiomics supervised autoencoder for pancancer clinical outcome endpoints prediction
BMC Medical Informatics and Decision Making volume 20, Article number: 129 (2020)
Abstract
Background
With the rapid development of sequencing technologies, collecting diverse types of cancer omics data become more costeffective. Many computational methods attempted to represent and fuse multiple omics into a comprehensive view of cancer. However, different types of omics are related and heterogeneous. Most of the existing methods do not consider the difference between omics, so the biological knowledge of individual omics may not be fully excavated. And for a given task (e.g. predicting overall survival), these methods prefer to use sample similarity or domain knowledge to learn a more reasonable representation of omics, but it’s not enough.
Methods
For the purpose of learning more useful representation for individual omics and fusing them to improve the prediction ability, we proposed an autoencoderbased method named MOSAE (Multiomics Supervised Autoencoder). In our method, a specific autoencoder were designed for each omics according to their size of dimension to generate omicsspecific representations. Then, a supervised autoencoder was constructed based on specific autoencoder by using labels to enforce each specific autoencoder to learn both omicsspecific and taskspecific representations. Finally, representations of different omics that generate from supervised autoencoders were fused in a traditional but powerful way, and the fused representation was used for subsequent predictive tasks.
Results
We applied our method over TCGA PanCancer dataset to predict four different clinical outcome endpoints (OS, PFI, DFI, and DSS). Compared with traditional and stateoftheart methods, MOSAE achieved better predictive performance. We also tested the effects of each improvement, which all have a positive effect on predictive performance.
Conclusions
Predicting clinical outcome endpoints are very important for precision medicine and personalized medicine. And multiomics fusion is an effective way to solve this problem. MOSAE is a powerful multiomics fusion method, which can generate both omicsspecific and taskspecific representation for given endpoint predictive tasks and improve the predictive performance.
Background
Introduction
Driven by highthroughput sequencing technologies, many cancer genomics programs have been established to generate omics data, so the cancer data grows almost exponentially in volume, variety and complexity [1]. Among these largescale sequencing studies, The Cancer Genome Atlas (TCGA) is the most famous one, which generates a rich resource of multiomics data and provides more than 30 cancer types. In TCGA, each patient holds multiomics profiling, including DNA methylation, protein expression, gene expression RNASeq, miRNA mature strand expression and so on. These different types of omics are related and heterogeneous. For a single patient, different omics are associated with the same trait and biological connections exist between different omics, but different omics provide different molecular level information for the same trait and they are also different in quantitative and descriptive forms. Therefore, integration analysis of multiomics is a great challenge, and powerful integration methods may promote the exploration of pathogenesis of cancer by taking advantage of different omics, and furthermore, accelerating the development of precision medicine and personalized medicine.
In most cancers, the number of patients is only a few hundred, but many omics has tens of thousands of dimensions, which means for a specific task, most omics features are noise which causes the curse of dimensionality problem. Furthermore, if we use multiple highdimension omics simultaneously, the problem will be more serious. Recently, many works [2] focus on alleviating the curse of dimensionality problem, but in real TCGA datasets, not all omics are highdimension such as the dimension of protein expression is only a few hundred in the TCGA PanCancer dataset. Moreover, recent works [3] utilize sample similarity to achieve the idea that similar samples should have similar representations, and utilizing domain knowledge to construct associations between and within omics. These methods are powerful in solving clustering problem, but for predicting, the representations that generate from the sample similarity and domain knowledge are not power enough. Because under specific task, there is still a lot of noise (e.g. taskindependent information) in these representations, which hinders the performance of these models [4].
Based on above observations, we proposed a model named Multiomics Supervised Autoencoder (MOSAE), an autoencoderbased multiomics fusion method, which designs different autoencoder structures for different omics based on dimensional differences and uses label information to enforce autoencoder to equip the ability that the representations generate from autoencoders are associated with subsequent tasks. So each omics will have a specific supervised autoencoder, and the representations produce from the supervised autoencoders will be fused in a simple but powerful manner. Key contributions of this paper are summarized as follows:

1.
We observed that different omics contain different biological knowledge, and only a fraction of this knowledge is useful for subsequent tasks. Therefore, we constructed an omics and taskspecific structure of autoencoder (named supervised autoencoder) to explore knowledge of each omics. Each omics has a unique supervised autoencoder, and representations generate from these supervised autoencoder contain both omics and taskspecific biological knowledge.

2.
We observed that traditional fusion methods such as concatenation are not suitable for integrating multiomics, and average is a better alternative. This method averaging the representations by element, hence, knowledge from different omics are enforced to have the same meanings in the same position of dimension.

3.
We redesigned the loss function to guarantees the availability of our structure. Specifically, for the supervised autoencoders of each omics, both prediction error and reconstruction error were constructed. And a single prediction error that based on fused representation were constructed. This loss function can prevent information leakage.

4.
We verified our method by predicting four different Pancancer clinical outcome endpoints. The results shown that MOSAE achieved better results than traditional and stateoftheart methods, and has robust generalization ability.
Related work
We summarized recent multiomics learning methods into two part: representation of omics and fusion between multiple representations of each sample.
For representing of omics, autoencoder is widely used, which is a deep learning method for dimension reduction of highdimensional omics data [5], representation of cancer patients [6] and even fusion of multiomics data [7]. And in order to increase the ability of representation of autoencoder, a large number of variants of autoencoder were developed [8]. utilizes nonlinear data selfexpressiveness to learn the hidden layer of autoencoder, which is the representation of patient [3]. utilizes feature interaction network and patient similarity network to constrain the training objective of autoencoder that alleviate overfitting and curse of dimensionality problem. Most of above methods are unsupervised or semisupervised and they perform the same processing on different omics, which reduced the representation ability of autoencoder. Therefore, our method MOSAE employed labels to train autoencoder for each omics and designed the structure of autoencoder according to the characteristics of different omics.
For fusing multiple representations, concatenation is the most common method, which concatenate different omics representations into a single vector directly, but concatenation is often unworkable [1]. Therefore, a plethora of methods have been developed to integrate/fuse multiple omics data. And [9] divides integrative methods into four categories: networkfree nonBayesian (NFNBY), networkfree Bayesian (NFBY), networkbased nonBayesian (NBNBY) and networkbased Bayesian (NBBY) methods. They are classified by whether employing a prior on data probability distribution or graphs to model interactions. In NFNBY, Partial Least Squares (PLS)based methods are used widely. For example, sPLS [10] is a sparse vision of PLS, and Multiblock PLS [11] performing PLS on a multiomics dataset. In addition, multiomics genewise weights is another popular method in NFNBY that integrate different omics into a score for each gene. In NFBY, iCluster [12] is an innovative method that capture the similarities among different omics by minimize withincluster variance. In NBNBY, stSVM [13] utilizes diffusion kernels to random walk on each network with restarts, and Similarity Network Fusion (SNF) [14] fuses patient similarity networks (samplesample network) iteratively from each omics. Affinity Network Fusion (ANF) [15] is an upgrade of SNF that employ state transition matrix to obtain affinity matrix (similarity matrix) and fused weighted view by a ‘smooth’ procedure. Our method belongs to NBNBY that no prior is assumed on data and no graph is used for model interactions, which means our model structure is simpler than others. However, our experimental results show that our fusion method outperforms concatenation and some NBNBYbased methods [3], and we believe our fusion method may contain some real biological meaning.
Methods
Our method is based on autoencoder so we give a brief introduction to it, and then we divided our method into three part: specific autoencoder, supervised autoencoder and multiomics fusion, and discussed them in later sections.
Autoencoder
Autoencoder is an unsupervised neural network method that applies backpropagation, setting the output values to be equal to the inputs. And one of the hidden layer of autoencoder is considered as the representation of the inputs. If the hidden layer has fewer neural units than the input layer, we treat the hidden layer as a compressed knowledge representation of the original input, otherwise we treat the hidden layer as a ‘diversity’ representation that map the original space to a higher dimensional space.
Usually, autoencoder is divided into two processes, encoder and decoder. Suppose the original input is X ∈ ℝ^{N × p}, a samplefeature matrix with N samples and p features. An onelayer neural network with parameter Θ_{e} is regarded as encoder:
H is usually referred to as latent representation of input X. The encoder maps N samples from pdimension space to kdimension space. And another onelayer neural network with parameter Θ_{d} is regarded as decoder:
\( \overset{\sim }{X} \) is reconstruction representation which has the same shape as X. The decoder maps N samples from kdimension space back to pdimension space, and it should be noted that X and \( \overset{\sim }{X} \) are different. The whole process of autoencoder can be expressed as:
Finally, the objective function of autoencoder can be formulated with Frobenius norm:
The objective function is also called reconstruction error, which try to penalize the difference between X and \( \overset{\sim }{X} \). And the latent representation H is generally used in subsequent tasks as the representation of input, because it’s widely believed that H retains input information in a better form. And there is an interesting fact that if all neural networks in autoencoder is linear and the dimension of H is less than the dimension of X, we would observe a similar dimensionality reduction as observed in principal component analysis (PCA).
Specific autoencoder
As mentioned before, different omics have different properties, more specifically, the dimension of some omics are high and other are low. This situation is very common, but current methods are not considered. In order to deal with this situation, we thought that the omics with highdimension contain more information than the omics with lowdimension, because highdimension omics contains more genes (or protein), and more genes (or protein) mean more complete descriptions of a patient.
Therefore, for highdimension omics (hold more information), we need compress them to a lower dimension space. In addition, the compression step can avoid overfitting. On the contrary, for lowdimension omics, the information is less, so we need to map original omics nonlinearity into higher dimension, and produce more nonlinear combinations of original features. Therefore, in our specific autoencoder, we suppose that there are M types of omics X = {X^{(1)}, …, X^{(i)}, …, X^{(M)}}, in which \( {X}^{(i)}\in {\mathbb{R}}^{N\times {p}^{(i)}} \) represents the i^{th} omics. And then the M types of omics have been divided into highdimension group X^{(high)} and lowdimension group X^{(low)}. Without loss of generality they were formulated as:
And autoencoders for each omics were formulated as:
\( {F}_{\uppi^{(i)}}\left(\bullet \right) \) is a ‘compression’ neural network and \( {G}_{\uppi^{(i)}}\left(\bullet \right) \) is an ‘expansion’ neural network, they have formed omicsspecific layers. And the decoder structure is same for all omics. H^{(i)} is the new latent representation of i^{th} omics that will be used in subsequent tasks, and all H^{(i)} have the same dimension k. The loss function become:
In fact, the structures are different in all omics not only in highdimension or lowdimension omics, because different autoencoder have different parameters.
Supervised autoencoder
Now, we have got the autoencoders for each omics, but the latent representations produced by those autoencoders may not good enough to represent the omics for a given task. Follow above ideas, highdimension omics hold more information, but under a given task, only a few information are useful and others are considered as noises. So many methods attempted to enforce autoencoder to learn more specific information but their representation ability is insufficient. And in this paper, we thought that using labels is the best way to do so. Therefore, we reformulated the loss function of the specific autoencoder as:
\( Predictor\left({H}^{(i)},{\varTheta}_p^{(i)}\ \right) \) is a supervised neural network with parameter \( {\varTheta}_p^{(i)} \) and input H^{(i)}. Y is the label vector of given task. And \( \mathcal{L} \) is the prediction loss (cross entropy or mean squared error). The loss function encourages that the representation produced by autoencoder should hold omics and taskspecific knowledge, and α is used to adjust propensity. From another perspective, the supervised autoencoder is a common supervised deep neural network with reconstruction error, but in our method, we focus on the representation rather than prediction, so we called it supervised autoencoder.
Multiomics fusion
After above processing, we can generate a very powerful representation for each omics, and they can be used directly. But different omics describe different aspects about the same trait, the representation of a singleomics is not comprehensive for a trait. Therefore, we used a very simple but powerful way, average, to fuse those representations. In supervised autoencoder, we got H^{(i)} for i^{th} omics, and all M representations have the same dimension k, so we fused them as:
Again, we used labels to enforce the H^{(fusion)} contains taskspecific information. And the loss function becomes:
Y^{(fusion)} is the result produced by a neural network with parameter Θ_{f} and input H^{(fusion)}. All labels were processed uniformly in this loss function, so there is no information leakage. And α, β are used to adjust propensity, they were set to 1 in this paper. Average is a very simple operation of fusion, but we believe it plays an important role: the elements in each H^{(i)} may have different meanings, but the average operation enforces the elements in the corresponding positions to have the same meanings. The fused representation may represent a higher level of biological information (e.g. pathway) than the molecular level.
An illustration of the whole framework of MOSAE is depicted in Fig. 1. The framework is the same for four different tasks.
Results
Dataset
We used TCGA PanCancer data to verify our method, which downloaded from UCSC Xena (https://xenabrowser.net/datapages/). There are many types of omics in TCGA PanCancer, and we selected four of them to verify our method, including DNA methylation, miRNA sequencing (miRNASeq), RNA sequencing (RNAseq) and protein expression (RPPA). Besides, we used four prediction tasks (binary) in our experiment: overall survival (OS), diseasespecific survival (DSS), progressionfree interval (PFI) and diseasefree interval (DFI). They are clinical outcome endpoints, and specific definitions can be found in [16].
For each task, samples with both above four types of omics were selected, and we obtained 5983, 5799, 5983 and 3191 samples for OS, DSS, PFI and DFI, respectively. And for features in each omics types, we removed variables with low variance or low mean and filtered out outliers. Because of the difference between tasks, the number of features in each omics under each task is different. And the final number of features we used is shown in Table 1.
Evaluation metric
The proportion of the sample in each task is unbalanced, for example, there are 5983 samples in OS, of which 1692 are positive samples and 4291 are negative samples. Therefore, we selected Area Under the Receiver Operating Characteristic Curve (ROC AUC) as the evaluation metric. In biostatistics, the metric is known as Cindex which ranges from 0.5 to 1, and the value over 0.7 indicate that our model may be a good model. We implemented this metric using scikitlearn (https://scikitlearn.org). Besides, we used 5fold cross validation to calculate the metric, so there is a standard deviation for each result.
Comparison with other methods
In order to verify the performance of our method MOSAE, we chosen six common machine learning methods as baselines to compare with MOSAE. Those baselines include SVM, DecisionTree, Naïve Bayes, kNN, RandomForest, and AdaBoost. And the input of baselines is the concatenation of multiple omics, and they were implemented using scikitlearn. Multiview Factorization AutoEncoder predicted OS and PFI using the same dataset and metric as us, and achieved stateoftheart results. So we used the results in their paper [3] directly to compare with ours. For MOSAE, there are four supervised autoencoders for four omics, and the number of input units is equivalent to the number of input features. DNA methylation and RNASeq belong to highdimension group so the number of the second layer of their corresponding autoencoder was set to 100 as a ‘compression’ neural network, and miRNASeq and RPPA belong to lowdimension group so the second layer was set to 1000 as an ‘expansion’ neural network. The latent representation layer was set to 400 for all autoencoders. The ROC AUC scores are shown in Table 2, and the standard deviation is in brackets. At the same time, in order to show the difference between different methods more intuitively, we visualized the ROC AUC scores in Fig. 2. Our algorithm achieves the best results under all tasks.
Further performance analysis
In this section, we experimented to examine the performance of various improvements in MOSAE. Since our method is based on autoencoder, we choose a plain autoencoder as the baseline, which is a three layers neural network, and we used the concatenation of multiomics as input, and used the middle layer as latent representation.
In order to show the results more intuitively, we divide MOSAE into four modules. The first module is multiomics (MO), which mean each omics has its own autoencoder. The second module is fusion, which determines the way latent representations are merged. Here we used two ways, concatenating (Cat) and averaging (Ave). And for concatenation, the size of each latent representation is set to 100, so the size of fusion representation is 400, which is equivalent to the size of average. The third module is supervised (Sup), which means using labels for latent representation. The fourth module is specific (Spec), which means the network structure will be adjusted according to the size of the dimension. So MOSAE can be expressed as (MO + Ave + Sup + Spec). The ROC AUC scores are shown in Table 3 and visualized in Fig. 3. Every module gives a steady improvement in performance for predicting all tasks.
Discussion
Performance of MOSAE
The results (Table 2, Fig. 2) compared with other methods shown that MOSAE has a significant improvement over the baseline approaches for predicting all tasks, and results are as stable as baselines. Therefore, we can draw a conclusion that our method outperforms traditional methods, and MOSAE can be generalized to different tasks. Besides, we observed that the results of PFI and DFI are worse than OS and DSS. We think this is because PFI and DFI have shortterm clinical followup intervals. And the amount of data in DFI is much smaller than PFI, so DFI is even worse. In addition, Multiview Factorization AutoEncoder [3] is an autoencoderbased method, which belongs to NBNBY methods that utilizes graph to model domain knowledge of features and utilizes sample similarity matrix to fuse different view. But its results for OS and PFI are 0.766 and 0.724, respectively, which is worse than ours (OS: 0.7830 and PFI: 0.7325). This may indicated that our fusion methods and other improvements are more effective than Multiview Factorization AutoEncoder.
From the results in Table 3 and Fig. 3, we can found that every module gives a steady improvement in performance for predicting all tasks, and Ave and Sup are the most prominent. The results indicated that the latent representations and fusion representations produced by MOSAE are more powerful for predicting tasks, and average is a better fusion method than concatenation.
Reconstruction error
From another perspective, MOSAE can be considered as a multiview neural network with reconstruction error. To examined the role of reconstruction error in MOSAE, we used plain neural network (NN) as the comparison object, the input is the same as plain autoencoder (AE). And MO, Ave, Sup and Spec are the same as above, for example, (AE + MO + Ave + Sup + Spec) is MOSAE, this is a litter different from above, the purpose is to make a more clearly comparison. And (NN + MO + Ave + Sup + Spec) means there is an independent neural network for each omics (MO), the representations that generate from each neural networks were fused by averaging (Ave), those independent neural network were train with labels (Sup) and the structure of neural network were dependent on the dimension of omics (Spec). From the results (Table 4), we can see that NN outperforms AE, which means the reconstruction error may not work in predicting, but our Sup module changed this situation. This change proves that adding label information to each autoencoder (with reconstruction error) can improve the representation ability. But there is an exception. For OS, reconstruction error did not work well, this may due to the good data quality of OS, which can be confirmed in [16].
Conclusions
Predicting clinical outcome endpoints are very important for precision medicine and personalized medicine. And multiomics fusion is an effective way to solve this problem. In this paper, we developed an autoencoderbased method named MOSAE to fuse multiomics to predict clinical outcome. Firstly, we utilized the difference of omics to design specific structure of autoencoders for different omics, then we employed labels to enforce autoencdoers to learning taskspecific representations. Finally, we fused those representations by averaging, which is a simple but powerful operation. MOSAE has been verified in PanCancer dataset, and the results shown that MOSAE outperforms traditional and stateoftheart methods in all tasks. Every improvement in MOSAE has improved the performance of prediction, and MOSAE has robust generalization ability. Our feature work will focus on designing new structure of autoencoder and developing more powerful and interpretable fusion methods.
Availability of data and materials
The datasets analysed during the current study are available in UCSC Xena (https://xenabrowser.net/datapages/).
Abbreviations
 MOSAE:

Multiomics Supervised Autoencoder
 TCGA:

The Cancer Genome Atlas
 NFNBY:

NetworkFree NonBayesian
 NFBY:

NetworkFree Bayesian
 NBNBY:

NetworkBased NonBayesian
 NBBY:

NetworkBased Bayesian
 PLS:

Partial Least Squares
 SNF:

Similarity Network Fusion
 ANF:

Affinity Network Fusion
 ROC AUC:

Area Under the Receiver Operating Characteristic Curve
 OS:

Overall survival
 DSS:

Diseasespecific survival
 PFI:

Progressionfree interval
 DFI:

Diseasefree interval
References
Li Y, Wu FX, Ngom A. A review on machine learning principles for multiview biological data integration. Brief Bioinform. 2016;19:325–40.
Meng C, Zeleznik OA, Thallinger GG, Kuster B, Gholami AM, Culhane AC. Dimension reduction techniques for the integrative analysis of multiomics data. Brief Bioinform. 2016;17:628–41.
Ma T, Zhang A. Multiview factorization AutoEncoder with network constraints for multiomic integrative analysis. In: 2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM); 2018. p. 702–7.
Locatello F, Bauer S, Lucic M, Rätsch G, Gelly S, Schölkopf B, Bachem O. Challenging common assumptions in the unsupervised learning of disentangled representations. arXiv preprint arXiv:1811.12359; 2018.
Chen L, Cai C, Chen V, et al. Learning a hierarchical representation of the yeast transcriptomic machinery using an autoencoder model. BMC Bioinformatics. 2016. https://doi.org/10.1186/s1285901508521.
Chaudhary K, Poirion OB, Lu L, Garmire LX. Deep learning–based multiomics integration robustly predicts survival in liver cancer. Clin Cancer Res. 2018;24:1248–59.
Miotto R, Li L, Kidd BA, Dudley JT. Deep patient: an unsupervised representation to predict the future of patients from the electronic health records. Sci Rep. 2016;6:26094.
Yang B, Liu S, Pang S, Pang C, Shang X. Deep Subspace Similarity Fusion for the Prediction of Cancer Subtypes. In: 2018 IEEE International Conference on Bioinformatics and Biomedicine (BIBM); 2018. p. 566–71.
Bersanelli M, Mosca E, Remondini D, Giampieri E, Sala C, Castellani G, Milanesi L. Methods for the integration of multiomics data: mathematical aspects. BMC Bioinformatics. 2016;17:S15.
Lê Cao KA, González I, Déjean S. integrOmics: an R package to unravel relationships between two omics datasets. Bioinformatics. 2009;25:2855–6.
Li W, Zhang S, Liu CC, Zhou XJ. Identifying multilayer gene regulatory modules from multidimensional genomic data. Bioinformatics. 2012;28:2458–66.
Shen R, Olshen AB, Ladanyi M. Integrative clustering of multiple genomic data types using a joint latent variable model with application to breast and lung cancer subtype analysis. Bioinformatics. 2009;25:2906–12.
Cun Y, Fröhlich H. Netclass: an rpackage for network based, integrative biomarker signature discovery. Bioinformatics. 2014;30:1325–6.
Wang B, Mezlini AM, Demir F, Fiume M, Tu Z, Brudno M, HaibeKains B, Goldenberg A. Similarity network fusion for aggregating data types on a genomic scale. Nat Methods. 2014;11:333.
Ma T, Zhang A. Integrate multiomic data using affinity network fusion (anf) for cancer patient clustering. In: 2017 IEEE International Conference on Bioinformatics and Biomedicine (BIBM); 2017. p. 398–403.
Liu J, Lichtenberg T, Hoadley KA, Poisson LM, Lazar AJ, Cherniack AD, Kovatich AJ, Benz CC, Levine DA, Lee AV, Omberg L. An integrated TCGA pancancer clinical data resource to drive highquality survival outcome analytics. Cell. 2018;173:400–16.
Acknowledgements
Not applicable.
About this supplement
This article has been published as part of BMC Medical Informatics and Decision Making Volume 20 Supplement 3, 2020: Health Information Processing. The full contents of the supplement are available online at https://bmcmedinformdecismak.biomedcentral.com/articles/supplements/volume20supplement3 .
Funding
Publication costs are funded by Natural Science Foundation of Guangdong Province of China (2015A030308017). This study was supported by Natural Science Foundation of Guangdong Province of China (2015A030308017), National Natural Science Foundation of China (61976239), Innovation Foundation of Highend Scientific Research Institutions of Zhongshan City of China (2019AG031). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Author information
Authors and Affiliations
Contributions
KT conceived the study, performed the experiments and wrote the paper. WH performed some of the experiments and modified the paper. JH advised on the algorithms and modified the paper. SD guided the project, advised on the algorithms and modified the paper. All authors read and reviewed the final manuscript.
Corresponding author
Ethics declarations
Ethics approval and consent to participate
Not applicable.
Consent for publication
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
About this article
Cite this article
Tan, K., Huang, W., Hu, J. et al. A multiomics supervised autoencoder for pancancer clinical outcome endpoints prediction. BMC Med Inform Decis Mak 20 (Suppl 3), 129 (2020). https://doi.org/10.1186/s1291102011143
Published:
DOI: https://doi.org/10.1186/s1291102011143
Keywords
 Multicomics
 Autoencoder
 Fusion
 Representation
 PanCancer
 Endpoints