Archive for CREST

Congrats to Arnak Dalalyan for his ERC advanced grant!

Posted in Statistics, University life with tags , , , , , , , , , on July 17, 2025 by xi'an

R[are]SS meeting

Posted in Statistics, Travel, University life with tags , , , , , , , , , , , , , , , , , , on September 29, 2024 by xi'an


Yesterday, I happened to be at the right time in the right place, as I was in Warwick for a RSS local section meeting on rare event simulation. (If missing the aurora borealis and the moon eclipse on previous nights!) And hence attended a seminar by Francesca Crucinio in six days!, as she talked about a turnkey approach to unbiased estimation of transforms of a moment, or wlog a mean μ, f(μ). A recent article with Nicolas Chopin (CREST) and Sumeet Singh, where they resort to Taylor expansions to achieve unbiasedness, using the Russian roulette trick to stop the summation from running to infinity. (As it happens, I heard Nicolas talk about this idea in the recent past namely at the ISBA-Fusion Sunday morn at Ca’Foscari.) Using a Taylor expansion is obviously natural and mathematically correct, albeit fraught with potential dangers [imho]:

  • the Taylor expansion involves central moments up to a random order R, which are harder & harder to estimate with increasing orders (i.e., more & more uncertain, with the possibility of infinite variance estimators after a certain order)
  • I did not spot a discussion on the moment estimators, that seems to rely on k iid replicas for the k-th moment
  • a lot of calibration ensues, from the choice of the centre x⁰ to the (artificial) distribution of the stopping value R, to the parameterisation of the random variable attached to the moment μ
  • the paper insists on recycling simulations to stabilise the moment estimators and ensure consistency, as a primary level of Rao-Blackwellisation, but this only applies to the smallest order moments and could be devised in many different ways, with varying computing costs
  • consistency of the estimate is not necessarily needed, as for instance for pseudo-marginal applications
  • as often with Russian roulette, positive quantities may receive negative estimations that are dominated by truncations to the positive real line (and alternating series offer the use of sandwiching estimators)
  • for the above reason, it is not always reasonable to tunnel vision on unbiasedness and alternative estimates like bridge sampling solutions could be integrating towards improving the quality of the estimator (especially since the conditions for finite variance involve unknown quantities)
  • while f-Taylored solutions like harmonic mean estimators for f(x)=1/x are not necessarily a panacea, they could be included in the comparison or as control variates

The first talk by Mathias Rousset was investigating adaptive multilevel sampling, a form of nested sampler, at the theoretical level, while the third talk by Tobias Grafke was a repetition of a talk he gave at the masterclass the interface between computational physics and computational statistics, last April.

PIPLA [mostly MCMC’nar]

Posted in Books, pictures, Statistics, University life with tags , , , , , , , , , , , , , on September 27, 2024 by xi'an

The first “mostly MCMC ” seminar (Season 2) had our new Ocean postdoc Tim Johnston, freshly graduated from the University of Edinburgh, involved in both talks, with proximal approximations for discontinuity! The first talk was given by Francesca Crucinio (formerly Warwick and formerly CREST, to point out potential COI!!), about the Proximal Interacting Particle Langevin algorithm (PIPLA) developed with her coauthors Paula Cordero Encinar, Deniz Akyildiz, Tim Johnston, and Mark Girolami, concerned with  maximising likelihoods with latent variables (i.e., an EM setting).  While offering one of many stochastic versions of EM, incl. simulated annealing, the solution they adopt very close to our SAME (2002) method, with duplicating latent variables N times to get near the marginal MAP (which as we noted differs from the join MAP). They start from interacting particle system with (unadjusted) Langevin dynamics, discretised over time, but the value of N does not move with iterations, which steps away from the simulated annealing motivation, thus requiring an evaluation of the error for a given N and possibly further runs with larger values of N. PIPLA is an extension of the above to non-differentiable targets, by using a proximity map, in continuation of MY-ULA [for Moreau-Yoshida] by Pereyra (2016), yet again fixing both N and the proximal parameter λ. With non-asymptotic convergence results requiring strong assumptions on the target.

In his talk, Tim started with interesting (and novel for me) arguments for proving strong convergence (Wasserstein, multilevel MC, unbiased MCMC), proceeding to establishing (again under favourable assumptions, and almost √n convergence speed for the proximal scheme with no regularity assumption on drift besides boundedness.

simulation as optimization [by kernel gradient descent]

Posted in Books, pictures, Statistics, University life with tags , , , , , , , , , , , , , , , , , , , , , , , , on April 13, 2024 by xi'an

Yesterday, which proved an unseasonal bright, warm, day, I biked (with a new wheel!) to the east of Paris—in the Gare de Lyon district where I lived for three years in the 1980’s—to attend a Mokaplan seminar at INRIA Paris, where Anna Korba (CREST, to which I am also affiliated) talked about sampling through optimization of discrepancies.
This proved a most formative hour as I had not seen this perspective earlier (or possibly had forgotten about it). Except through some of the talks at the Flatiron Institute on Transport, Diffusions, and Sampling last year. Incl. Marilou Gabrié’s and Arnaud Doucet’s.
The concept behind remains attractive to me, at least conceptually, since it consists in approximating the target distribution, known up to a constant (a setting I have always felt standard simulation techniques was not exploiting to the maximum) or through a sample (a setting less convincing since the sample from the target is already there), via a sequence of (particle approximated) distributions when using the discrepancy between the current distribution and the target or gradient thereof to move the particles. (With no randomness in the Kernel Stein Discrepancy Descent algorithm.)
Ana Korba spoke about practically running the algorithm, as well as about convexity properties and some convergence results (with mixed performances for the Stein kernel, as opposed to SVGD). I remain definitely curious about the method like the (ergodic) distribution of the endpoints, the actual gain against an MCMC sample when accounting for computing time, the improvement above the empirical distribution when using a sample from π and its ecdf as the substitute for π, and the meaning of an error estimation in this context.

“exponential convergence (of the KL) for the SVGD gradient flow does not hold whenever π has exponential tails and the derivatives of ∇ log π and k grow at most at a polynomial rate”

sequential meetings in Edinburgh

Posted in Books, Kids, Mountains, pictures, Running, Statistics, Travel, University life with tags , , , , , , , , , , , , , , , on October 24, 2023 by xi'an


There will be not one but two consecutive events in Edinburgh next May²⁴ on sequential Monte Carlo methods! Both hosted by the fantastic International Centre for Mathematical Sciences (ICMS) in Edinburgh Olde Town. Within the Bayes Centre. And running distance to Arthur’s Seat. (Reminding me of my first ICMS workshop in 2001 run with Mike Titterington. May have been my first week long visit to Edinburgh as well…)

First, a Summer School on Bayesian filtering: fundamental theory and numerical methods (SSBF 2024), Edinburgh (UK), May 6-10, 2024. This summer (in the Scottish sense!) school will cover topics related to fundamental theory, state-of-the-art methodologies, and real-world applications.

Second, a Sequential Monte Carlo workshop (SMC 2024), the week later, on May 13-17, 2024. The workshop will cover topics related to sequential Monte Carlo and nearby fields, from theory to applications, following earlier workshops in the series. Including the one at CREST in 2015.

Thanks to Víctor Elvira, Jana de Wiljes, and Dan Crisan for this double deal (and the opportunity to return to Scotland for the first time since the pandemic).