A Greedy Approximation for k-Determinantal Point Processes

Julia Grosse, Rahel Fischer, Roman Garnett, Phlipp Hennig

Research output: Contribution to journalConference articlepeer-review

Abstract

Determinantal point processes (DPPs) are an important concept in random matrix theory and combinatorics, and increasingly in machine learning. Samples from these processes exhibit a form of self-avoidance, so they are also helpful in guiding algorithms that explore to reduce uncertainty, such as in active learning, Bayesian optimization, reinforcement learning, and marginalization in graphical models. The best-known algorithms for sampling from DPPs exactly require significant computational expense, which can be unwelcome in machine learning applications when the cost of sampling is relatively low and capturing the precise repulsive nature of the DPP may not be critical. We suggest an inexpensive approximate strategy for sampling a fixed number of points (as would typically be desired in a machine learning setting) from a so-called k-DPP based on iterative inverse transform sampling. We prove that our algorithm satisfies a (1 − 1/e) approximation guarantee relative to exact sampling from the k-DPP, and provide an efficient implementation for many common kernels used in machine learning, including the Gaussian and Matérn class. Finally, we compare the empirical runtime of our method to exact and Markov-Chain-Monte-Carlo (MCMC) samplers and investigate the approximation quality in a Bayesian Quadrature (BQ) setting.

Original languageEnglish
Pages (from-to)3052-3060
Number of pages9
JournalProceedings of Machine Learning Research
Volume238
StatePublished - 2024
Event27th International Conference on Artificial Intelligence and Statistics, AISTATS 2024 - Valencia, Spain
Duration: May 2 2024May 4 2024

Fingerprint

Dive into the research topics of 'A Greedy Approximation for k-Determinantal Point Processes'. Together they form a unique fingerprint.

Cite this