Structured sparsity regularization is a class of methods, and an area of research in statistical learning theory, that extend and generalize sparsity regularization learning methods.[1] Both sparsity and structured sparsity regularization methods seek to exploit the assumption that the output variable (i.e., response, or dependent variable) to be learned can be described by a reduced number of variables in the input space (i.e., the domain, space of features or explanatory variables). Sparsity regularization methods focus on selecting the input variables that best describe the output. Structured sparsity regularization methods generalize and extend sparsity regularization methods, by allowing for optimal selection over structures like groups or networks of input variables in .[2][3]
Common motivation for the use of structured sparsity methods are model interpretability, high-dimensional learning (where dimensionality of may be higher than the number of observations ), and reduction of computational complexity.[4] Moreover, structured sparsity methods allow to incorporate prior assumptions on the structure of the input variables, such as overlapping groups,[2] non-overlapping groups, and acyclic graphs.[3] Examples of uses of structured sparsity methods include face recognition,[5] magnetic resonance image (MRI) processing,[6] socio-linguistic analysis in natural language processing,[7] and analysis of genetic expression in breast cancer.[8]
^Rosasco, Lorenzo; Poggio, Tomasso (December 2014). A Regularization Tour of Machine Learning, MIT-9.520 Lectures Notes.
^ abCite error: The named reference groupLasso was invoked but never defined (see the help page).
^ abCite error: The named reference latentLasso was invoked but never defined (see the help page).
^Cite error: The named reference LR18 was invoked but never defined (see the help page).
^Jia, Kui; et al. (2012). "Robust and Practical Face Recognition via Structured Sparsity". In Andrew Fitzgibbon; Svetlana Lazebnik; Pietro Perona; Yoichi Sato; Cordelia Schmid (eds.). Computer Vision – ECCV 2012: 12th European Conference on Computer Vision, Florence, Italy, October 7-13, 2012 Proceedings, Part IV.
^Chen, Chen; et al. (2012). "Compressive Sensing MRI with Wavelet Tree Sparsity". Proceedings of the 26th Annual Conference on Neural Information Processing Systems. Vol. 25. Curran Associates. pp. 1115–1123.
^Eisenstein, Jacob; et al. (2011). "Discovering Sociolinguistic Associations with Structured Sparsity". Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics.
^Jacob, Laurent; et al. (2009). "Group Lasso with Overlap and Graph Lasso". Proceedings of the 26th International Conference on Machine Learning.
and 24 Related for: Structured sparsity regularization information
variables). Sparsityregularization methods focus on selecting the input variables that best describe the output. Structuredsparsityregularization methods...
regularization problems where the regularization penalty may not be differentiable. One such example is ℓ 1 {\displaystyle \ell _{1}} regularization (also...
matrix regularization generalizes notions of vector regularization to cases where the object to be learned is a matrix. The purpose of regularization is to...
shrinkage. There are several variations to the basic sparse approximation problem. Structuredsparsity: In the original version of the problem, any of the...
way is a relaxed version of the k-sparse autoencoder. Instead of forcing sparsity, we add a sparsityregularization loss, then optimize for min θ , ϕ...
under which recovery is possible. The first one is sparsity, which requires the signal to be sparse in some domain. The second one is incoherence, which...
Manifold regularization adds a second regularization term, the intrinsic regularizer, to the ambient regularizer used in standard Tikhonov regularization. Under...
and other metrics. Regularization perspectives on support-vector machines interpret SVM as a special case of Tikhonov regularization, specifically Tikhonov...
kernel Predictive analytics Regularization perspectives on support vector machines Relevance vector machine, a probabilistic sparse-kernel model identical...
similar to dropout as it introduces dynamic sparsity within the model, but differs in that the sparsity is on the weights, rather than the output vectors...
codes. The regularization and kernel theory literature for vector-valued functions followed in the 2000s. While the Bayesian and regularization perspectives...
\mathbf {\Gamma } } . The local sparsity constraint allows stronger uniqueness and stability conditions than the global sparsity prior, and has shown to be...
l_{1}} -regularization techniques, such as sparse regression, LASSO, and l 1 {\displaystyle l_{1}} -SVM Regularized trees, e.g. regularized random forest...
Bayesian statistics of graphical models, false discovery rates, and regularization. She is the Louis Block Professor of statistics at the University of...
for efficient computation Parallel tree structure boosting with sparsity Efficient cacheable block structure for decision tree training XGBoost works...
successfully used RLHF for this goal have noted that the use of KL regularization in RLHF, which aims to prevent the learned policy from straying too...
representation error (over the input data), together with L1 regularization on the weights to enable sparsity (i.e., the representation of each data point has only...
language model. Skip-gram language model is an attempt at overcoming the data sparsity problem that preceding (i.e. word n-gram language model) faced. Words represented...
Hessenberg decomposition and QR decomposition based approaches with regularization have begun to attract attention In 2017, Google Scholar Blog published...
regarded as a generalization of the kernel perceptron algorithm with regularization. The sequential minimal optimization (SMO) algorithm used to learn support...
does not work for high-dimensional data without regularization, while FPCA has a built-in regularization due to the smoothness of the functional data and...
case where no regularization has been integrated, by the singular values of matrix F {\displaystyle F} . Of course, the use of regularization (or other kinds...