Project Details
Projekt Print View

Random Matrix Approaches to Approximate Bayesian Inference in Machine Learning

Subject Area Image and Language Processing, Computer Graphics and Visualisation, Human Computer Interaction, Ubiquitous and Wearable Computing
Electronic Semiconductors, Components and Circuits, Integrated Systems, Sensor Technology, Theoretical Electrical Engineering
Statistical Physics, Nonlinear Dynamics, Complex Systems, Soft and Fluid Matter, Biological Physics
Term from 2018 to 2022
Project identifier Deutsche Forschungsgemeinschaft (DFG) - Project number 407712271
 
The Bayesian paradigm provides important methods for learning from data. It combines a probabilistic model for the generation of data together with prior knowledge over likely parameters within a probability distribution over the parameters of the model. However, practical applications of this idea to models with a large number of parameters are often plagued by computational problems related to the intractability of high--dimensional probability distributions. Approximate inference methods of machine learning provide algorithms for approximating such distributions by simpler ones - typically by multivariate Gaussian distributions. These inference methods yield often excellent results in applications. But, the update of covariance matrices (which give important information on uncertainties and dependencies between variables) of these Gaussian distributions within the iterative inference algorithms requires matrix operations per iteration which grows cubic in the number of parameters of the model. This makes the applications of such methods problematic when the number of variables is large. Hence, further approximations are necessary. And these may deteriorate the quality of the predictions. A second relevant problem is the fact that there is no guarantee of convergence for some popular inference algorithms. It is unclear if the failure to converge is an artefact of the algorithm or is related to the complexity of the Bayesian model. Motivated by recent research in the fields of information theory and statistical physics, this project will address these problems from a new angle. Assuming that data matrices can be considered as random (in a mathematically well-defined way), results of random matrix theory suggest novel ways to efficiently approximate the required matrix operations. These approximations are expected to perform well in the asymptotic limit when matrices are large. Random matrix methods will also provide new ways for analyzing the performance of iterative inference algorithms for large problems under certain statistical assumptions on the data. We will use these random matrix techniques to speed up existing algorithms as well as designing novel algorithms with optimized convergence properties. We will investigate the quality and robustness of such methods. Finally, we will validate our approach on various Bayesian models in machine learning and compare the performance with that of competing methods on simulated as well as real data.
DFG Programme Research Grants
International Connection Denmark, Switzerland
 
 

Additional Information

Textvergrößerung und Kontrastanpassung