Export Citations
Save this search
Please login to be able to save your searches and receive alerts for new content matching your search criteria.
- research-articleMay 2024
Optimal preconditioning and fisher adaptive langevin sampling
NIPS '23: Proceedings of the 37th International Conference on Neural Information Processing SystemsArticle No.: 1280, Pages 29449–29460We define an optimal preconditioning for the Langevin diffusion by analytically optimizing the expected squared jumped distance. This yields as the optimal preconditioning an inverse Fisher information covariance matrix, where the covariance matrix is ...
- research-articleApril 2024
Gradient estimation with discrete stein operators
NIPS '22: Proceedings of the 36th International Conference on Neural Information Processing SystemsArticle No.: 1873, Pages 25829–25841Gradient estimation—approximating the gradient of an expectation with respect to the parameters of a distribution—is central to the solution of many machine learning problems. However, when the distribution is discrete, most common gradient estimators ...
- research-articleNovember 2022
Variance reduction for Metropolis–Hastings samplers
AbstractWe introduce a general framework that constructs estimators with reduced variance for random walk Metropolis and Metropolis-adjusted Langevin algorithms. The resulting estimators require negligible computational cost and are derived in a post-...
- research-articleApril 2022
Sequential changepoint detection in neural networks with checkpoints
AbstractWe introduce a framework for online changepoint detection and simultaneous model learning which is applicable to highly parametrized models, such as deep neural networks. It is based on detecting changepoints across time by sequentially performing ...
- research-articleJune 2024
Entropy-based adaptive Hamiltonian Monte Carlo
NIPS '21: Proceedings of the 35th International Conference on Neural Information Processing SystemsArticle No.: 2182, Pages 28482–28495Hamiltonian Monte Carlo (HMC) is a popular Markov Chain Monte Carlo (MCMC) algorithm to sample from an unnormalized probability distribution. A leapfrog integrator is commonly used to implement HMC in practice, but its performance can be sensitive to the ...
-
- research-articleMay 2021
Large scale multi-label learning using Gaussian processes
AbstractWe introduce a Gaussian process latent factor model for multi-label classification that can capture correlations among class labels by using a small set of latent Gaussian process functions. To address computational challenges, when the number of ...
- research-articleDecember 2019
Gradient-based adaptive Markov chain Monte Carlo
NIPS'19: Proceedings of the 33rd International Conference on Neural Information Processing SystemsDecember 2019, Article No.: 1409, Pages 15730–15739We introduce a gradient-based learning method to automatically adapt Markov chain Monte Carlo (MCMC) proposal distributions to intractable targets. We define a maximum entropy regularised objective function, referred to as generalised speed measure, ...
- ArticleAugust 2017
Bayesian Boolean matrix factorisation
ICML'17: Proceedings of the 34th International Conference on Machine Learning - Volume 70Pages 2969–2978Boolean matrix factorisation aims to decompose a binary data matrix into an approximate Boolean product of two low rank, binary matrices: one containing meaningful patterns, the other quantifying how the observations can be expressed as a combination of ...
- ArticleDecember 2016
One-vs-each approximation to softmax for scalable estimation of probabilities
NIPS'16: Proceedings of the 30th International Conference on Neural Information Processing SystemsPages 4168–4176The softmax representation of probabilities for categorical variables plays a prominent role in modern machine learning with numerous applications in areas such as large scale classification, neural language modeling and recommendation systems. However, ...
- ArticleDecember 2016
The generalized reparameterization gradient
NIPS'16: Proceedings of the 30th International Conference on Neural Information Processing SystemsPages 460–468The reparameterization gradient has become a widely used method to obtain Monte Carlo gradients to optimize the variational objective. However, this technique does not easily apply to commonly used distributions such as beta or gamma without further ...
- research-articleJuly 2016
First learn then earn: optimizing mobile crowdsensing campaigns through data-driven user profiling
MobiHoc '16: Proceedings of the 17th ACM International Symposium on Mobile Ad Hoc Networking and ComputingPages 271–280https://doi.org/10.1145/2942358.2942369We study the optimal design of mobile crowdsensing campaigns in terms of the aggregate quality of contributions attracted for a set of tasks. The interaction of the campaign with users is realized through a mobile app interface that recommends tasks to ...
- ArticleJune 2016
Overdispersed black-box variational inference
UAI'16: Proceedings of the Thirty-Second Conference on Uncertainty in Artificial IntelligencePages 647–656We introduce overdispersed black-box variational inference, a method to reduce the variance of the Monte Carlo estimator of the gradient in black-box variational inference. Instead of taking samples from the variational distribution, we use importance ...
- research-articleMay 2016
Short-Term Load Forecasting using a Cluster of Neural Networks for the Greek Energy Market
SETN '16: Proceedings of the 9th Hellenic Conference on Artificial IntelligenceArticle No.: 15, Pages 1–6https://doi.org/10.1145/2903220.2903222In the context of the liberalization of the Greek Energy Market, load forecasting is essential in various system programming procedures. Short-term load forecasting extends from one to seven days, although in this paper a model is proposed for the next ...
- articleJanuary 2016
Variational inference for latent variables and uncertain inputs in Gaussian processes
The Gaussian process latent variable model (GP-LVM) provides a flexible approach for non-linear dimensionality reduction that has been widely applied. However, the current approach for training GP-LVMs is based on maximum likelihood, where the latent ...
- ArticleDecember 2015
Inference for determinantal point processes without spectral knowledge
NIPS'15: Proceedings of the 29th International Conference on Neural Information Processing Systems - Volume 2Pages 3393–3401Determinantal point processes (DPPs) are point process models that naturally encode diversity between the points of a given realization, through a positive definite kernel K. DPPs possess desirable properties, such as exact sampling or analyticity of the ...
- ArticleDecember 2015
Local expectation gradients for black box variational inference
NIPS'15: Proceedings of the 29th International Conference on Neural Information Processing Systems - Volume 2Pages 2638–2646We introduce local expectation gradients which is a general purpose stochastic variational inference algorithm for constructing stochastic gradients by sampling from the variational distribution. This algorithm divides the problem of estimating the ...
- ArticleDecember 2014
Hamming ball auxiliary sampling for factorial hidden Markov models
NIPS'14: Proceedings of the 28th International Conference on Neural Information Processing Systems - Volume 2Pages 2960–2968We introduce a novel sampling algorithm for Markov chain Monte Carlo-based Bayesian inference for factorial hidden Markov models. This algorithm is based on an auxiliary variable construction that restricts the model space allowing iterative exploration ...
- ArticleJune 2014
Doubly stochastic variational bayes for non-conjugate inference
ICML'14: Proceedings of the 31st International Conference on International Conference on Machine Learning - Volume 32Pages II-1971–II-1980We propose a simple and effective variational inference algorithm based on stochastic optimisation that can be widely applied for Bayesian non-conjugate inference in continuous parameter spaces. This algorithm is based on stochastic approximation and ...
- ArticleDecember 2013
Variational inference for Mahalanobis distance metrics in Gaussian process regression
NIPS'13: Proceedings of the 27th International Conference on Neural Information Processing Systems - Volume 1Pages 279–287We introduce a novel variational method that allows to approximately integrate out kernel hyperparameters, such as length-scales, in Gaussian process regression. This approach consists of a novel variant of the variational framework that has been ...
- ArticleJune 2012
Manifold relevance determination
ICML'12: Proceedings of the 29th International Coference on International Conference on Machine LearningPages 531–538In this paper we present a fully Bayesian latent variable model which exploits conditional nonlinear (in)-dependence structures to learn an efficient latent representation. The latent space is factorized to represent shared and private information from ...