Please use this identifier to cite or link to this item: https://hdl.handle.net/20.500.14279/29873
Title: Rethinking Bayesian Learning for Data Analysis: The Art of Prior and Inference in Sparsity-Aware Modeling
Authors: Cheng, Lei 
Yin, Feng 
Theodoridis, Sergios 
Chatzis, Sotirios P. 
Chang, Tsung-Hui 
Major Field of Science: Engineering and Technology
Field Category: Mechanical Engineering
Keywords: Statistics - Machine Learning;Statistics - Machine Learning;Computer Science - Learning;eess.IV;eess.SP
Issue Date: Nov-2022
Source: IEEE Signal Processing Magazine, 2022, vol. 39, no. 6, pp. 18-52
Volume: 39
Issue: 6
Start page: 18
End page: 52
Journal: IEEE Signal Processing Magazine 
Abstract: Sparse modeling for signal processing and machine learning has been at the focus of scientific research for over two decades. Among others, supervised sparsity-aware learning comprises two major paths paved by: a) discriminative methods and b) generative methods. The latter, more widely known as Bayesian methods, enable uncertainty evaluation w.r.t. the performed predictions. Furthermore, they can better exploit related prior information and naturally introduce robustness into the model, due to their unique capacity to marginalize out uncertainties related to the parameter estimates. Moreover, hyper-parameters associated with the adopted priors can be learnt via the training data. To implement sparsity-aware learning, the crucial point lies in the choice of the function regularizer for discriminative methods and the choice of the prior distribution for Bayesian learning. Over the last decade or so, due to the intense research on deep learning, emphasis has been put on discriminative techniques. However, a come back of Bayesian methods is taking place that sheds new light on the design of deep neural networks, which also establish firm links with Bayesian models and inspire new paths for unsupervised learning, such as Bayesian tensor decomposition. The goal of this article is two-fold. First, to review, in a unified way, some recent advances in incorporating sparsity-promoting priors into three highly popular data modeling tools, namely deep neural networks, Gaussian processes, and tensor decomposition. Second, to review their associated inference techniques from different aspects, including: evidence maximization via optimization and variational inference methods. Challenges such as small data dilemma, automatic model structure search, and natural prediction uncertainty evaluation are also discussed. Typical signal processing and machine learning tasks are demonstrated.
URI: https://hdl.handle.net/20.500.14279/29873
ISSN: 10535888
DOI: 10.1109/MSP.2022.3198201
Rights: © Copyright IEEE
Type: Article
Affiliation : Cyprus University of Technology 
The Chinese University of Hong Kong 
Shenzhen Research Institute of Big Data 
National and Kapodistrian University of Athens 
Publication Type: Peer Reviewed
Appears in Collections:Άρθρα/Articles

CORE Recommender
Show full item record

SCOPUSTM   
Citations

68
checked on Mar 14, 2024

WEB OF SCIENCETM
Citations

49
Last Week
3
Last month
11
checked on Nov 1, 2023

Page view(s)

161
Last Week
0
Last month
1
checked on Nov 21, 2024

Google ScholarTM

Check

Altmetric


This item is licensed under a Creative Commons License Creative Commons