Cooperative learning for multi-view analysis

12/23/2021
by   Daisy Yi Ding, et al.
3

We propose a new method for supervised learning with multiple sets of features ("views"). The multi-view problem is especially important in biology and medicine, where "-omics" data such as genomics, proteomics and radiomics are measured on a common set of samples. Cooperative learning combines the usual squared error loss of predictions with an "agreement" penalty to encourage the predictions from different data views to agree. By varying the weight of the agreement penalty, we get a continuum of solutions that include the well-known early and late fusion approaches. Cooperative learning chooses the degree of agreement (or fusion) in an adaptive manner, using a validation set or cross-validation to estimate test set prediction error. One version of our fitting procedure is modular, where one can choose different fitting mechanisms (e.g. lasso, random forests, boosting, neural networks) appropriate for different data views. In the setting of cooperative regularized linear regression, the method combines the lasso penalty with the agreement penalty. The method can be especially powerful when the different data views share some underlying relationship in their signals that can be exploited to strengthen signal, while each view has its idiosyncratic noise that needs to be reduced. We show that cooperative learning achieves higher predictive accuracy on simulated data and real multiomics examples of cancer stage and treatment response prediction. Leveraging aligned signals and allowing flexible fitting mechanisms for different modalities, cooperative learning offers a powerful approach to multiomics data fusion.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/02/2023

Semi-supervised Cooperative Learning for Multiomics Data Fusion

Multiomics data fusion integrates diverse data modalities, ranging from ...
research
11/06/2018

Stacked Penalized Logistic Regression for Selecting Views in Multi-View Learning

In multi-view learning, features are organized into multiple sets called...
research
03/10/2019

Lasso tuning through the flexible-weighted bootstrap

Regularized regression approaches such as the Lasso have been widely ado...
research
12/11/2019

Integrative Generalized Convex Clustering Optimization and Feature Selection for Mixed Multi-View Data

In mixed multi-view data, multiple sets of diverse features are measured...
research
05/19/2020

Fast cross-validation for multi-penalty ridge regression

Prediction based on multiple high-dimensional data types needs to accoun...
research
10/10/2018

Principal component-guided sparse regression

We propose a new method for supervised learning, especially suited to wi...
research
06/07/2020

What needles do sparse neural networks find in nonlinear haystacks

Using a sparsity inducing penalty in artificial neural networks (ANNs) a...

Please sign up or login with your details

Forgot password? Click here to reset