**Multi-Task Structure Learning**

In the paradigm of multi-task learning, multiple related prediction tasks are learned jointly by sharing some information. This calls to specify when to share (i.e., decide when some tasks are related), what to share (i.e., determine the form through which sharing among the related tasks occur) and how to share (i.e., specify concrete ways to share knowledge among tasks). More formally, multi-task learning is defined as follows.

**Multi-Task Learning.**
Let \(T\) learning tasks \(\{\mathcal{T}_t\}_{t=1}^T\) where each task \(\mathcal{T}_t\) is associated to a training set \(\mathcal{D}_t = \{x_{t,i},y_{t,i}\}_{i=1}^{n_t}\), a model parameter \(w_t\in\mathbb{R}^d\) which need to be learned, and a loss function \(\mathcal{L}(\cdot,\mathcal{D}_t)\) evaluating the prediction performance of \(w_t\) on the training set. In the following, we denote by \(W=[w_1\cdots w_T]\in\mathbb{R}^{d\times T}\) the collection of model parameters. Multi-task learning aims to help improve the learning of a model for \(\mathcal{T}_t\) by using the knowledge contained in all or some of the \(T\) tasks.

A major challenge in multi-task learning consist in finding groups of related tasks in order to prevent that two unrelated tasks do not end up influencing each other, hence worsening the performance of both tasks. In what follows, I will present various types of approaches addressing that challenge.

## 1. Factorization approaches and sparse coding techniques

One line of research relies on the assumption that the parameter matrix \(W\) can be factorized as \(W=LS\) where \(L\) denotes the latent basis tasks and \(S\) is a matrix containing the weights of linear combination for each task. Hence, the predictor \(w_t\) for the \(t\)-th task is given by \(Ls_t\) where \(s_t\) is the \(t\)-th column of \(S\). Several assumptions can be made on \(L\) and \(S\).

To the best of my knowledge, the work of [Kumar and Daumé, 2012] initiated these factorization techniques in multi-task learning by taking its inspiration from the low dimensional subspace assumption of [Argyriou et al., 2008] where it was achieved by imposing a trace-constraint, encouraging sparsity on the singular values of \(W\). The same assumption is enforced there by factorizing \(W=LS\) with a small number \(k\) of latent basis tasks so that \(L\) and \(S\) are low-rank matrices. In addition, each linear combination is assumed to be sparse in the latent bases and the overlap in the sparsity patterns of two tasks controls the amount of sharing between them. A similar work has been conducted in [Maurer et al., 2013] where the authors highlight the connection between such factorization technique and sparse coding. They additionally present a probabilistic analysis which complements well with the practical insights in the above work. The same factorization is found in [Barzilai and Crammer, 2015] but the approach differs by the fact that each task is associated to a single latent basis task. Departing from these works, [Zhong et al., 2016] considers instead a full-rank assumption on \(L\) along with a row-sparsity assumption on \(S\) that encourages related tasks to share a subset of basis tasks. More recently, the work [Jeong and Jun, 2018] brought the low-rank assumption up to date with sparsity assumptions between and within the rows of \(L\) and the use of the \(k\)-support norm on columns of \(S\) to impose sparsity while considering possible correlations.

**GO-MTL** [Kumar and Daumé, 2012]
. The authors assume that there are \(k < T\) latent basis tasks, i.e., \(L\in\mathbb{R}^{d\times k}\) and \(S\in\mathbb{R}^{k\times T}\) are two low-rank matrices. In addition, they penalize the complexity of \(L\) and assume that each task is represented by a sparse combination of the latent tasks. This results in the following optimization problem.

Since the model allow two tasks from different groups to overlap by having one or more bases in common, it is called *Grouping and Overlap in Multi-Task Learning* by the authors.

**ASAP-MT** [Barzilai and Crammer, 2015]
. The authors also assume that there are \(k < T\) latent basis tasks, i.e., \(L\in\mathbb{R}^{d\times k}\) and \(S\in\mathbb{R}^{k\times T}\). They penalize the complexity of \(L\) and enforce that the \(t\)-th column of the matrix \(S\), denoted by \(s_t\in\mathbb{R}^k\) associates task \(t\) with one of the \(k\) clusters. For example, if the \(k\)-th element of \(s_t\) is one, and all other elements of \(s_t\) are zero, we would say that \(t\) is associated with cluster \(k\).

Here the authors considers the hinge loss and logistic loss, and use some slack variables to optimize the objective. The optimization problem is then recast as *A SAddle Point convex optimization problem* which gives its name to the method.

**FMTL** [Zhong et al., 2016]
. Motivated by GO-MTL, the authors propose the *Flexible Multi-Task Learning* paradigm to identify the task grouping and overlap without imposing any specific structure assumptions, e.g., the number of latent basis tasks. Instead of predetermining the size of latent basis tasks and constraining the subspace to be low rank, the authors use a full rank subspace and introduce two regularization terms to the corresponding representation matrix \(S\) of the learning tasks. The first is a \(\ell_{2,1}\)-norm regularization term, which introduces row-sparsity on \(S\) that encourages related tasks to share a subset of basis tasks. The second term is column-orthogonality that prevents unrelated tasks
from sharing common basis. The constraint ensure that the latent basis tasks are orthogonal and form a subspace in \(\mathbb{R}^d\).

Since \(L\) is orthogonal and full rank, we can easily derive the representation matrix as \(S=L^{-1}W=L^\top W\) and \(S^\top S = W^\top W\). Thus we have

\[\underset{L,W}{\mathrm{minimize}} \; \sum_{t=1}^T \mathcal{L}(w_t,\mathcal{D}_t) + \lambda_1 \| L^\top W \|_{2,1} + \lambda_2 \| W^\top W\|_F^2\quad \text{s.t.}\quad L^\top L = I_{d\times d}\]**VSTG-MTL** [Jeong and Jun, 2018]
. The *Variable Selection and Task Grouping for Multi-Task Learning* method also assumes that \(L\) and \(S\) are low-rank matrices but differs from the regularization terms. It encourages the sparsities between and within the latent task vectors and relies on the k-support norm to induce less sparse weighting vector \(v_t\) than that from the \(\ell_1\) norm and similarly enhances the overlaps in the
task groups.

The group structure among tasks are decided by the sparsity patterns on the weighting vector \(s_t\). Tasks with same sparsity patterns on the weighting vector \(s_t\) belong to the same group, whereas those with the orthogonal ones belong to disjoint groups. Two groups are regard as being overlapped if their sparsity patterns are not orthogonal, i.e., they partially share the latent bases.

## 2. Decomposition approaches and dirty models

Another set of approaches called decomposition approaches assume that the parameter matrix can be decomposed as the sum of multiple component matrices, i.e., \(W=\sum_{h=1}^H W^{(h)}\). Depending on the sparsity patterns or closeness of some components, groups of related tasks can be a posteriori discovered.

In [Jalali et al., 2010] the authors proposes a dirty models with element-wise sparsity in \(W^{(1)}\) and block-structured row-sparsity in \(W^{(2)}\) so that the resulting sparsity patterns of the sum unveils the related tasks. On the clean side, the work of [Chen et al., 2011] captures the relationship of multiple related tasks using a low-rank structure in \(W^{(1)}\) while identifying the outlier tasks using a group-sparse structure in \(W^{(2)}\). A multi-level decomposition is considered in [Han and Zhang, 2015] where a \(\ell_2\) norm encourages the closeness of pairs of task at each level. A group structure can then be a posteriori recovered level-wise.

**Dirty** [Jalali et al., 2010]
. This work considers a model where features are assumed to be either active for all tasks, or inactive for most of the tasks. To do so, the parameter matrix is written as \(W = W^{(1)} + W^{(2)}\) with different sparsity assumptions for each: element-wise sparsity in \(W^{(1)}\) and block-structured row-sparsity in \(W^{(2)}\).

As a result, certain rows of \(W\) would have many non-zero entries, corresponding to features shared by several tasks, while certain rows would be elementwise sparse, corresponding to those features which are relevant for some tasks but not all, while certain rows would have all zero entries, corresponding to those features that are not relevant to any task. On the contrary, a *clean* model would just use one type of sparsity assumption and not multiple.

**RMTL** [Chen et al., 2011]
. In the *Robust Multi-Task Learning* approach, the parameter matrix is written as \(W = W^{(1)} + W^{(2)}\) where \(W^{(1)}\) is supposed to be low-rank and \(W^{(2)}\) is assumed to be group sparse. The corresponding optimization problem reads

If the \(t\)-th task is from the related tasks group, \(w_t^{(2)}\) is expected to be a zero-vector and hence \(w_t\) obeys the specified low-rank structure constraint; on the other hand, if the \(t\)-th task is from the outlier tasks group, \(w_t^{(2)}\) is expected to be non-zero.

**MeTaG** [Han and Zhang, 2015]
. The *Multi-level TAsk
Grouping* method aims to learn the multi-level grouping structure instead of only one level among tasks.

The regularization imposes a \(\ell_2\) norm on the pairwise difference among the column vectors in \(W_h\), which encourages each pair of columns \(w_{t}^{(h)}\) and \(w_{t'}^{(h)}\) in \(W^{(h)}\) to be identical. If this happens, then the \(t\)-th and \(t'\)-th tasks belong to a task group at the \(h\)-th level.

## 3. Clustering based on representative tasks

In order to exploit the relationships between tasks, this kind of approaches writes the model parameter of each task \(w_t\) as a linear combination of some representative tasks selected from the tasks pool or from the whole the tasks pool, i.e. \(w_t\approx W c_t\), where \(C\) is a \(T\times T\) task correlation matrix and \(c_t\) is the combination coefficients. The grouping structure of the tasks is embedded in \(C\) meaning that tasks that select a common representative task are regarded as a group, and all tasks can be clustered into different groups based on their representative tasks.

In [Lee et al., 2016] , the authors proposed to represent each task by a sparse non-negative linear combination of all the other tasks. The method allow for asymmetric information transfer between the tasks, such that the amount of information transfer from a confident predictor to a less confident one is larger than the other way around. However, since the combination coefficients are restricted to be positive, this prevent negatively correlated tasks from being clustered together and hence from sharing information. To cope with this issue, [Liu and Pan, 2017] relaxed the positive restriction so that the method can capture both positive and negative correlations amongst tasks. In addition, the correlation matrix was restricted to be block-diagonal with a trace Lasso norm penalty. More recently, [Yao et al., 2019] provides a more robust representation by applying the \(\ell_{1,2}\)-norm regularization twice: once to the representation difference \(W-WC\) and once to the correlation matrix \(C\) in order to select a few representative tasks.

**AMTL** [Lee et al., 2016]
. The *Asymmetric Multi-Task Learning* method relies on an asymmetric matrix \(C\) (such that \(c_{t,t}=0\)) which allows for asymmetric information transfer between the tasks, such that the amount of information transfer from a confident predictor to a less confident one is larger than the other way around. The authors considers the following optimization problem

where \(\Lambda(W)=\mathrm{Diag}( \{\mathcal{L}(w_t, \mathcal{D}_t)\}_{t=1}^T)\) so that the first regularization term enforces each task parameter to be reconstructed as a sparse combination of other tasks selected based on the task-wise loss.

**GAMTL** [Liu and Pan, 2017]
. The *Group Adaptive Multi-Task Learning* method uses one regularization term to enforce that each task is a linear combination of other tasks while the other term is the trace Lasso penalty imposed on the task relationship vectors \(\{c_t\}\)’s in \(C\).

**RCMTL** [Yao et al., 2019]
. The *Robust Clustered Multi-Task Learning* approach encourages only the relevant tasks to share common information with each other by using \(\ell_{1,2}\) penalties.

## 4. Indicators and Groupwise regularization approaches

Another line of research to unveils groups of related tasks relies on the sum of \(L\) penalties terms, each one of them applied to a different group of tasks. The \(L\) groups are explicitly modeled by \(L\) diagonal matrices \(Q_1,\ldots,Q_L\) where \(Q_l\in\{0,1\}^{T\times T}\) is the indicator matrix for the \(l\)-th group. The goal is then to estimate these indicator matrices so that the group-wise penalty best regularizes the parameter matrix.

The first work that initiated these approaches in the multi-task setting is [Kang et al., 2011] where the authors applied the trace norm squared groupwise in order that tasks parameters within each group lie in a low dimensional subspace. More recently, [Kshirsagar et al., 2017] followed the same strategy but used the \(\ell_{1,2}\) norm instead so that task relatedness is intended as shared sparsity, meaning that tasks in a group all have similar relevant features. Since in both cases the objective is smooth, its optimization is carried alternatively over the parameter matrix \(W\) and a relaxed version \(\{Q_l\}_l\)’s. Departing from these methods, [Frecon et al., 2020] considers the groupwise trace norm penalty and designed a smooth continuous bilevel scheme where the \(\{Q_l\}_l\) are estimated at the upper-level so that the corresponding parameter matrix, estimated at the lower-level, generalizes well to unseen data.

**Whom** [Kang et al., 2011]
. This method assumes that tasks form disjoint groups and the tasks parameters within each group lie in a low dimensional subspace.

**Group-MTL** [Kshirsagar et al., 2017]
. In this work, task relatedness is intended as shared sparsity, meaning that tasks in a group all have similar relevant features, i.e., the same zeros in their parameter vectors.

**BiGMTL** [Frecon et al., 2020]
. The *Bilevel Grouping in Multi-Task Learning* method decouples the learning of the groups and the parameter matrix in two levels. The upper-level problem optimize the groups so that the corresponding parameter matrix, learned at the lower-level problem, performs well on a validation set. This work also assumes tasks parameters within each group lie in a low dimensional subspace by using the trace norm penalty.

## 5. Graph-based regularized approaches

Here we focus on a popular multi-task framework denoted as graph-based regularized framework. In this framework, task relations are represented as a graph whose nodes are the tasks and the weighted edges encode some knowledge over similarities between tasks. Here, the relationship between the tasks is usually modeled by a task covariance matrix \(\Sigma\) or a task precision matrix \(\Omega=\Sigma^{-1}\).

**Clustered-MTL** [Jacob et al., 2009]
. This work relies on the hypothesis that the parameters of tasks within a cluster lie close to each other in \(\ell_2\) norm sense. The penalties are made of three terms: \(\mathcal{R}_{\text{mean}}\) which measures on average how large the weight vectors are, \(\mathcal{R}_{\text{between}}(W,\mathcal{G})\) which quantifies how close to each other the different groups are, and \(\mathcal{R}_{\text{within}}(W,\mathcal{G})\)
which measures the compactness of the groups

By adopting some convex relaxation technique, the convex objective function can be formulated as

\[\underset{W,\Sigma}{\mathrm{minimize}}\; \sum_{t=1}^T \mathcal{L}(w_t,\mathcal{D}_t) + \lambda \mathrm{tr}{(W 1 1^\top W^\top )} + \mathrm{tr}{( \tilde{W} \Sigma^{-1} \tilde{W}^\top )}\quad\text{s.t.}\quad \begin{cases}\tilde{W} = W (\mathbf{1}-11^\top / T),\\ \alpha \mathbf{1} \preceq \Sigma \preceq \beta \mathbf{1},\\ \mathrm{tr}{(\Sigma)} = \gamma\end{cases}\]**MTLapTR** [Fei and Huan, 2012]
. The *Multi-Task Laplacian regularization with Task Relationship inference* method combines sparse regularization and task relationship modeling in the sense that the method selects a small subset of features and the feature sets of two closely related tasks have common features. To do so, the authors suggest the following objective function:

\(\underset{W,\Sigma}{\mathrm{minimize}}\; \sum_{t=1}^T \mathcal{L}(w_t,\mathcal{D}_t) + \lambda_1\|W\|_1 + \frac{\lambda_2}{2} \mathrm{tr}(W^\top L W) + \frac{\lambda_3}{2}\mathrm{tr}(W\Sigma^{-1}W^\top)\quad\text{s.t.}\quad \begin{cases}
\Sigma \succeq 0 \\ \mathrm{tr}(\Sigma)=k
\end{cases}\)
where \(\mathrm{tr}(W^\top LW)\) is a Tikhonov regularization term conveniently
written in matrix format in terms of the graph Laplacian \(L\) matrix in order to impose smoothness across features meaning that the selected features tend to be connected in the feature graph. One down side is that \(L\) should be specified *a priori*.

**MTRL** [Zhang and Yeung, 2014]
. The *Multi-Task Relationship Learning* method estimates task dependence (in the form of a task covariance matrix) along with task specific coefficients. A convex relaxation is proposed to ease optimization.

**CoGraph-MTL** [Flamary et al., 2014]
. The purpose of the *Constrained Graph-regularized Multi-Task Learning* method is to learn the adjacency matrix of the task relations graph, jointly with the task decision function parameters, while making the graph the more interpretable as possible. The interpretability of the adjacency matrix is achieved by incorporating some sparsity-inducing regularization \(\mathcal{R}\). The authors consider a model that induces pairwise similarity between tasks. A bilevel framework for learning task similarities in multi-task learning problems is designed where the similarities are learned so as to optimize a proxy on the generalization errors of all tasks. The proposed optimization problem reads

where \(\Sigma(\lambda) = \mathrm{Diag}( \{\lambda_t\}_{t=1}^T) + L(\lambda)\) plays the role of a tasks covariance matrix (under some constraints which can be included in \(\mathcal{R}\)) with \(L(\lambda)\) being the Laplacian matrix.

**MSSL** [Gonçalves et al., 2016]
. The *Multi-task Sparse Structure Learning* method assumes that i) the features across tasks (rows \(w_j\) of the matrix \(W\)) follows a multivariate Gaussian distribution with zero mean and covariance matrix \(\Sigma\), i.e., \(w_j \sim \mathcal{N}(0,\Sigma)\), and that ii) \(W\) and the precision matrix \(\Omega=\Sigma^{-1}\) are sparse.

The authors also propose an extension to copulas.

**BMSL** [Gonçalves et al., 2019]
. This work introduces the *Bayesian Multitask with Structure Learning* model and propose a Gibbs sampler for inference. Here the authors considers two types of priors on the precision matrix i) a diffuse Wishart prior so that all tasks are assumed to be equally related a priori and ii) a Bayesian graphical LASSO prior to impose sparsity in the task relatedness.

**GGMTL** [Alesiani et al., 2020]
. The inner problem (model learning) aims at finding the optimal model for a given structure (i.e. graph or equivalently the adjency matrix \(e\)), while the outer problem (structure learning) aims at minimizing a cost function that includes two terms: (1) the learned model’s accuracy on the validation data, and (2) the sparseness of the graph. The sparseness of the graph is captured with three terms: (a) the \(\ell_2^2\) norm of the edge values, measuring the energy of the graph, (b) the \(\ell_1\) norm measuring the sparseness of the edges, and (c) \(H(e)\) measuring the entropy of the edges.

where \(L(e)\) is the Laplacian matrix, so the regularization is the Dirichlet energy

\[\mathrm{tr}(W^\top L(e) W) = \sum_{i,j\in G} e_{i,j}\|w_i-w_j\|_2^2,\]and \(H(e)\) is the un-normalized entropy of the edge values .

## References

**[Alesiani et al., 2020]**F. Alesiani, S. Yu, A. Shaker and W. Yin. "Towards Interpretable Multi-Task Learning Using Bilevel Programming". ArXiv Computing Research Repository (CoRR) (2020)**[Argyriou et al., 2008]**A. Argyriou, T. Evgeniou M. Pontil. "Convex multi-task feature learning". Machine learning (2008)**[Barzilai and Crammer, 2015]**A. Barzilai and K. Crammer. "Convex Multi-Task Learning by Clustering". Proceedings of the International Conference on Artificial Intelligence and Statistics (AISTATS) (2015)**[Chen et al., 2011]**J. Chen, J. Zhou and J. Ye. "Integrating low-rank and group-sparse structures for robust multi-task learning". Proceedings of the International Conference on Knowledge discovery and data mining (KDD) (2011)**[Fei and Huan, 2012]**H. Fei and J. Huan. "Structured feature selection and task relationship inference for multi-task learning". Knowledge and Information Systems (KAIS) (2012)**[Flamary et al., 2014]**R. Flamary, A. Rakotomamonjy and G. Gasso. "Learning Constrained Task Similarities in Graph-Regularized Multi-Task Learning". Regularization, Optimization, Kernels, and Support Vector Machines (2014)**[Frecon et al., 2020]**J. Frecon, S. Salzo and M. Pontil. "Unveiling groups of related tasks in multi-task learning". Proceedings of the International Conference on Pattern Recognition (ICPR) (2020)**[Gonçalves et al., 2016]**A.R. Gonçalves, F.J. Von Zuben and A. Banerjee. "Multi-task Sparse Structure Learning with Gaussian Copula Models". Journal of Machine Learning Research (JMLR) (2016)**[Gonçalves et al., 2019]**A.R. Gonçalves, P. Ray, B. Soper, D. Widemann, M. Nygård, J.F. Nygård and A.P. Sales. "Bayesian multitask learning regression for heterogeneous patient cohorts". Journal of Biomedical Informatics (JBI) (2019)**[Han and Zhang, 2015]**L. Han Y. Zhang. "Learning multi-level task groups in multi-task learning". Proceedings of the Conference on Artificial Intelligence (AAAI) (2015)**[Jacob et al., 2009]**L. Jacob, J.-P. Vert and F.R. Bach. "Clustered Multi-Task Learning - A Convex Formulation". Proceedings of the International Conference on Neural Information Processing Systems (NIPS) (2009)**[Jalali et al., 2010]**A. Jalali, S. Sujay, C. Ruan and P. Ravikumar. "A Dirty Model for Multi-task Learning". Proceedings of the International Conference on Neural Information Processing Systems (NIPS) (2010)**[Jeong and Jun, 2018]**J.-Y. Jeong and C.-H. Jun. "Variable Selection and Task Grouping for Multi-Task Learning". Proceedings of the International Conference on Knowledge Discovery & Data Mining (KDD) (2018)**[Kang et al., 2011]**Z. Kang, K. Grauman and F. Sha. "Learning with Whom to Share in Multi-task Feature Learning". Proceedings of the International Conference on Machine Learning (ICML) (2011)**[Kshirsagar et al., 2017]**M. Kshirsagar, E. Yang, A.C. Lozano. "Learning Task Clusters via Sparsity Grouped Multitask Learning". Proceedings of the European Conference on Machine Learning and Knowledge Discovery in Databases (ECML PKDD) (2017)**[Kumar and Daumé, 2012]**A. Kumar and H. Daumé. "Learning Task Grouping and Overlap in Multi-Task Learning". Proceedings of the International Conference on Machine Learning (ICML) (2012)**[Lee et al., 2016]**G. Lee, E. Yang and S. Hwang. "Asymmetric Multi-task Learning Based on Task Relatedness and Loss". Proceedings of the International Conference on Machine Learning (ICML) (2016)**[Liu and Pan, 2017]**S. Liu and S.J. Pan. "Adaptive Group Sparse Multi-task Learning via Trace Lasso". Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI) (2017)**[Maurer et al., 2013]**A. Maurer, M. Pontil and B. Romera-Paredes. "Sparse Coding for Multitask and Transfer Learning". Proceedings of the International Conference on Machine Learning (ICML) (2013)**[Yao et al., 2019]**Y. Yao, J. Cao and H. Chen. "Robust Task Grouping with Representative Tasks for Clustered Multi-Task Learning". Proceedings of the International Conference on Knowledge Discovery & Data Mining (KDD) (2019)**[Zhang and Yeung, 2014]**Y. Zhang and D.-Y. Yeung. "A Regularization Approach to Learning Task Relationships in Multitask Learning". ACM Transactions on Knowledge Discovery from Data (TKDD) (2014)**[Zhong et al., 2016]**S. Zhong, J. Pu, Y.-G. Jiang, R. Feng and X. Xue. "Flexible multi-task learning with latent task grouping". Neurocomputing (2016)