To save content items to your account,
please confirm that you agree to abide by our usage policies.
If this is the first time you use this feature, you will be asked to authorise Cambridge Core to connect with your account.
Find out more about saving content to .
To save content items to your Kindle, first ensure no-reply@cambridge-org.demo.remotlog.com
is added to your Approved Personal Document E-mail List under your Personal Document Settings
on the Manage Your Content and Devices page of your Amazon account. Then enter the ‘name’ part
of your Kindle email address below.
Find out more about saving to your Kindle.
Note you can select to save to either the @free.kindle.com or @kindle.com variations.
‘@free.kindle.com’ emails are free but can only be saved to your device when it is connected to wi-fi.
‘@kindle.com’ emails can be delivered even when you are not connected to wi-fi, but note that service fees apply.
In this article, we give explicit bounds on the Wasserstein and Kolmogorov distances between random variables lying in the first chaos of the Poisson space and the standard normal distribution, using the results of Last et al. (Prob. Theory Relat. Fields165, 2016). Relying on the theory developed by Saulis and Statulevicius in Limit Theorems for Large Deviations (Kluwer, 1991) and on a fine control of the cumulants of the first chaoses, we also derive moderate deviation principles, Bernstein-type concentration inequalities, and normal approximation bounds with Cramér correction terms for the same variables. The aforementioned results are then applied to Poisson shot noise processes and, in particular, to the generalized compound Hawkes point processes (a class of stochastic models, introduced in this paper, which generalizes classical Hawkes processes). This extends the recent results of Hillairet et al. (ALEA19, 2022) and Khabou et al. (J. Theoret. Prob.37, 2024) regarding the normal approximation and those of Zhu (Statist. Prob. Lett.83, 2013) for moderate deviations.
We present and study a novel algorithm for the computation of 2-Wasserstein population barycenters of absolutely continuous probability measures on Euclidean space. The proposed method can be seen as a stochastic gradient descent procedure in the 2-Wasserstein space, as well as a manifestation of a law of large numbers therein. The algorithm aims to find a Karcher mean or critical point in this setting, and can be implemented ‘online’, sequentially using independent and identically distributed random measures sampled from the population law. We provide natural sufficient conditions for this algorithm to almost surely converge in the Wasserstein space towards the population barycenter, and we introduce a novel, general condition which ensures uniqueness of Karcher means and, moreover, allows us to obtain explicit, parametric convergence rates for the expected optimality gap. We also study the mini-batch version of this algorithm, and discuss examples of families of population laws to which our method and results can be applied. This work expands and deepens ideas and results introduced in an early version of Backhoff-Veraguas et al. (2022), in which a statistical application (and numerical implementation) of this method is developed in the context of Bayesian learning.
We study the poor-biased model for money exchange introduced in Cao & Motsch ((2023) Kinet. Relat. Models 16(5), 764–794.): agents are being randomly picked at a rate proportional to their current wealth, and then the selected agent gives a dollar to another agent picked uniformly at random. Simulations of a stochastic system of finitely many agents as well as a rigorous analysis carried out in Cao & Motsch ((2023) Kinet. Relat. Models 16(5), 764–794.), Lanchier ((2017) J. Stat. Phys. 167(1), 160–172.) suggest that, when both the number of agents and time become large enough, the distribution of money among the agents converges to a Poisson distribution. In this manuscript, we establish a uniform-in-time propagation of chaos result as the number of agents goes to infinity, which justifies the validity of the mean-field deterministic infinite system of ordinary differential equations as an approximation of the underlying stochastic agent-based dynamics.
Expectiles have received increasing attention as a risk measure in risk management because of their coherency and elicitability at the level $\alpha\geq1/2$. With a view to practical risk assessments, this paper delves into the worst-case expectile, where only partial information on the underlying distribution is available and there is no closed-form representation. We explore the asymptotic behavior of the worst-case expectile on two specified ambiguity sets: one is through the Wasserstein distance from a reference distribution and transforms this problem into a convex optimization problem via the well-known Kusuoka representation, and the other is induced by higher moment constraints. We obtain precise results in some special cases; nevertheless, there are no unified closed-form solutions. We aim to fully characterize the extreme behaviors; that is, we pursue an approximate solution as the level $\alpha $ tends to 1, which is aesthetically pleasing. As an application of our technique, we investigate the ambiguity set induced by higher moment conditions. Finally, we compare our worst-case expectile approach with a more conservative method based on stochastic order, which is referred to as ‘model aggregation’.
In this article, we propose an algorithm for aligning three-dimensional objects when represented as density maps, motivated by applications in cryogenic electron microscopy. The algorithm is based on minimizing the 1-Wasserstein distance between the density maps after a rigid transformation. The induced loss function enjoys a more benign landscape than its Euclidean counterpart and Bayesian optimization is employed for computation. Numerical experiments show improved accuracy and efficiency over existing algorithms on the alignment of real protein molecules. In the context of aligning heterogeneous pairs, we illustrate a potential need for new distance functions.
Omega ratio, a risk-return performance measure, is defined as the ratio of the expected upside deviation of return to the expected downside deviation of return from a predetermined threshold described by an investor. Motivated by finding a solution protected against sampling errors, in this paper, we focus on the worst-case Omega ratio under distributional uncertainty and its application to robust portfolio selection. The main idea is to deal with optimization problems with all uncertain parameters within an uncertainty set. The uncertainty set of the distribution of returns given characteristic information, including the first two orders of moments and the Wasserstein distance, can handle data problems with uncertainty while making the calculation feasible.
In this paper, we consider the convergence rate with respect to Wasserstein distance in the invariance principle for deterministic non-uniformly hyperbolic systems. Our results apply to uniformly hyperbolic systems and large classes of non-uniformly hyperbolic systems including intermittent maps, Viana maps, unimodal maps and others. Furthermore, as a non-trivial application to the homogenization problem, we investigate the Wasserstein convergence rate of a fast–slow discrete deterministic system to a stochastic differential equation.
Stein’s method is used to study discrete representations of multidimensional distributions that arise as approximations of states of quantum harmonic oscillators. These representations model how quantum effects result from the interaction of finitely many classical ‘worlds’, with the role of sample size played by the number of worlds. Each approximation arises as the ground state of a Hamiltonian involving a particular interworld potential function. Our approach, framed in terms of spherical coordinates, provides the rate of convergence of the discrete approximation to the ground state in terms of Wasserstein distance. Applying a novel Stein’s method technique to the radial component of the ground state solution, the fastest rate of convergence to the ground state is found to occur in three dimensions.
This article derives quantitative limit theorems for multivariate Poisson and Poisson process approximations. Employing the solution of the Stein equation for Poisson random variables, we obtain an explicit bound for the multivariate Poisson approximation of random vectors in the Wasserstein distance. The bound is then utilized in the context of point processes to provide a Poisson process approximation result in terms of a new metric called $d_\pi$, stronger than the total variation distance, defined as the supremum over all Wasserstein distances between random vectors obtained by evaluating the point processes on arbitrary collections of disjoint sets. As applications, the multivariate Poisson approximation of the sum of m-dependent Bernoulli random vectors, the Poisson process approximation of point processes of U-statistic structure, and the Poisson process approximation of point processes with Papangelou intensity are considered. Our bounds in $d_\pi$ are as good as those already available in the literature.
Gromov–Wasserstein distances were proposed a few years ago to compare distributions which do not lie in the same space. In particular, they offer an interesting alternative to the Wasserstein distances for comparing probability measures living on Euclidean spaces of different dimensions. We focus on the Gromov–Wasserstein distance with a ground cost defined as the squared Euclidean distance, and we study the form of the optimal plan between Gaussian distributions. We show that when the optimal plan is restricted to Gaussian distributions, the problem has a very simple linear solution, which is also a solution of the linear Gromov–Monge problem. We also study the problem without restriction on the optimal plan, and provide lower and upper bounds for the value of the Gromov–Wasserstein distance between Gaussian distributions.
Consider a Pólya urn with balls of several colours, where balls are drawn sequentially and each drawn ball is immediately replaced together with a fixed number of balls of the same colour. It is well known that the proportions of balls of the different colours converge in distribution to a Dirichlet distribution. We show that the rate of convergence is $\Theta(1/n)$ in the minimal $L_p$ metric for any $p\in[1,\infty]$, extending a result by Goldstein and Reinert; we further show the same rate for the Lévy distance, while the rate for the Kolmogorov distance depends on the parameters, i.e. on the initial composition of the urn. The method used here differs from the one used by Goldstein and Reinert, and uses direct calculations based on the known exact distributions.
We show that several machine learning estimators, including square-root least absolute shrinkage and selection and regularized logistic regression, can be represented as solutions to distributionally robust optimization problems. The associated uncertainty regions are based on suitably defined Wasserstein distances. Hence, our representations allow us to view regularization as a result of introducing an artificial adversary that perturbs the empirical distribution to account for out-of-sample effects in loss estimation. In addition, we introduce RWPI (robust Wasserstein profile inference), a novel inference methodology which extends the use of methods inspired by empirical likelihood to the setting of optimal transport costs (of which Wasserstein distances are a particular case). We use RWPI to show how to optimally select the size of uncertainty regions, and as a consequence we are able to choose regularization parameters for these machine learning estimators without the use of cross validation. Numerical experiments are also given to validate our theoretical findings.
The classical Monge–Kantorovich (MK) problem as originally posed is concerned with how best to move a pile of soil or rubble to an excavation or fill with the least amount of work relative to some cost function. When the cost is given by the square of the Euclidean distance, one can define a metric on densities called the Wasserstein distance. In this note, we formulate a natural matrix counterpart of the MK problem for positive-definite density matrices. We prove a number of results about this metric including showing that it can be formulated as a convex optimisation problem, strong duality, an analogue of the Poincaré–Wirtinger inequality and a Lax–Hopf–Oleinik–type result.
(1) For any finite metric space $M$ the Lipschitz-free space on $M$ contains a large well-complemented subspace that is close to $\ell _{1}^{n}$.
(2) Lipschitz-free spaces on large classes of recursively defined sequences of graphs are not uniformly isomorphic to $\ell _{1}^{n}$ of the corresponding dimensions. These classes contain well-known families of diamond graphs and Laakso graphs.
Interesting features of our approach are: (a) We consider averages over groups of cycle-preserving bijections of edge sets of graphs that are not necessarily graph automorphisms. (b) In the case of such recursive families of graphs as Laakso graphs, we use the well-known approach of Grünbaum (1960) and Rudin (1962) for estimating projection constants in the case where invariant projections are not unique.
The study of finite approximations of probability measures has a long history. In Xu and Berger (2017), the authors focused on constrained finite approximations and, in particular, uniform ones in dimension d=1. In the present paper we give an elementary construction of a uniform decomposition of probability measures in dimension d≥1. We then use this decomposition to obtain upper bounds on the rate of convergence of the optimal uniform approximation error. These bounds appear to be the generalization of the ones obtained by Xu and Berger (2017) and to be sharp for generic probability measures.
In this paper we consider linear functions constructed on two different weighted branching processes and provide explicit bounds for their Kantorovich–Rubinstein distance in terms of couplings of their corresponding generic branching vectors. Motivated by applications to the analysis of random graphs, we also consider a variation of the weighted branching process where the generic branching vector has a different dependence structure from the usual one. By applying the bounds to sequences of weighted branching processes, we derive sufficient conditions for the convergence in the Kantorovich–Rubinstein distance of linear functions. We focus on the case where the limits are endogenous fixed points of suitable smoothing transformations.
We show that the total number of collisions in the exchangeable coalescent process driven by the beta (1, b) measure converges in distribution to a 1-stable law, as the initial number of particles goes to ∞. The stable limit law is also shown for the total branch length of the coalescent tree. These results were known previously for the instance b = 1, which corresponds to the Bolthausen-Sznitman coalescent. The approach we take is based on estimating the quality of a renewal approximation to the coalescent in terms of a suitable Wasserstein distance. Application of the method to beta (a, b)-coalescents with 0 < a < 1 leads to a simplified derivation of the known (2 - a)-stable limit. We furthermore derive asymptotic expansions for the moments of the number of collisions and of the total branch length for the beta (1, b)-coalescent by exploiting the method of sequential approximations.
We present a multi-phase image segmentation method based on the histogram of the Gabor feature space, which consists of a set of Gabor-filter responses with various orientations, scales and frequencies. Our model replaces the error function term in the original fuzzy region competition model with squared 2-Wasserstein distance function, which is a metric to measure the distance of two histograms. The energy functional is minimized by alternative minimization method and the existence of closed-form solutions is guaranteed when the exponent of the fuzzy membership term being 1 or 2. We test our model on both simple synthetic texture images and complex natural images with two or more phases. Experimental results are shown and compared to other recent results.
We consider the problem of approximating a probability measure defined on a metric spaceby a measure supported on a finite number of points. More specifically we seek theasymptotic behavior of the minimal Wasserstein distance to an approximation when thenumber of points goes to infinity. The main result gives an equivalent when the space is aRiemannian manifold and the approximated measure is absolutely continuous and compactlysupported.
We consider the approximate Euler scheme for Lévy-drivenstochastic differential equations.We study the rate of convergence in law of the paths.We show that when approximating the small jumps by Gaussianvariables, the convergence is much faster than when simplyneglecting them.For example, when the Lévy measure of the driving processbehaves like |z|−1−αdz near 0, for some α∈ (1,2), we obtain an error of order 1/√n with a computational cost of order nα. For a similar error when neglecting the small jumps, see[S. Rubenthaler, Numerical simulation of the solution of a stochastic differential equation driven by a Lévy process. Stochastic Process. Appl.103 (2003) 311–349], the computational cost is of ordernα/(2−α), which is huge when α is close to 2. In the same spirit, we study the problem of the approximation of a Lévy-driven S.D.E.by a Brownian S.D.E. when the Lévy process has no large jumps. Our results rely on some results of [E. Rio, Upper bounds for minimal distances in the central limit theorem. Ann. Inst. Henri Poincaré Probab. Stat.45 (2009) 802–817] about the central limit theorem, in the spirit of the famous paper by Komlós-Major-Tsunády [J. Komlós, P. Major and G. Tusnády, An approximation of partial sums of independentrvs and the sample df I. Z. Wahrsch. verw. Gebiete32 (1975) 111–131].