CNRS Défi Imag’In Cavalieri
CAlcul des Variations pour L'Imagerie, l'Edition et la Recherche
d'Images
Workshop
on Optimal Transport and Optimization in Imaging
Tuesday 11th
9h15-9h30: Welcoming
9h45-10h30: C.-B.
Schönlieb and Y.
Boink.
Combined modelling of optimal transport and segmentation.
Abstract
For studying vascular structures in 4D biomedical imaging, it is of
great importance to automatically determine the velocity of flow in
video sequences, for example blood flow in vessel networks. In this
thesis new optimal transport models focusing on direction and
segmentation are investigated to find an accurate displacement between
two density distributions. By incorporating fluid dynamics
constraints, one can obtain a realistic description of the
displacement. With an a-priori given segmentation of the network
structure, transport models can be improved. However, a segmentation
is not always known beforehand. Therefore, in this work a joint
segmentation-optimal transport model has been described. Other
contributions are the ability of the model to allow for inflow or
outflow and the incorporation of anisotropy in the displacement cost.
For the problem, a convex variational method has been used and
primal-dual proximal splitting algorithms have been implemented.
Existence of a solution of the model has been proved. The framework
has been applied to synthetic vascular structures and real data,
obtained from a collaboration with the applied mathematics and the
hospital in Cambridge.
This is joint work with Yoeri Boink and Christoph Brune.
Coffee
break
11h00-11h45: N.
Bonneel.
Wasserstein Barycentric Coordinates.
Abstract
This talk will expose a notion of barycentric coordinates for
histograms via optimal transport. This is formulated as the problem
of finding the Wasserstein barycenter of several basis histograms
that best fits an input histogram according to some arbitrary loss
function. A Wasserstein barycenter is a histogram in-between the
basis histograms according to the optimal transport metric ; we use
this geometry to project an input histogram onto the set of possible
Wasserstein barycenters. We developed an efficient and robust
algorithm from an automatic differentiation of the Sinkhorn
iterative procedure.
We illustrate our algorithm with applications in computer graphics,
such as color grading an input image using a database of
photographs, or filling-in missing values in captured reflectance
functions or geometries based on similar exemplars.
11h45-12h15: E.
Cazelles. Regularized
Wasserstein barycenter. Abstract
The concept of barycenter in the Wasserstein space allows the
dention of a notion of Frechet mean of a set of probability
measures. However, depending on the considered data, such
barycenters may be irregular. In this paper, we thus introduce a
convex regularization of Wasserstein barycenters for random measures
supported on ℝd. We prove
the existence and uniqueness of such barycenters for a large class
of regularizing functions. A stability result of regularized
barycenters in terms of Bregman divergence associated to the convex
regularization term is also given. Hence, we study the case of data
made of i.i.d. random probability measures. In particular, we prove
the convergence in Bregman divergence of the regularized empirical
barycenter of a set of n random probability measures towards its
population counterpart, and we discuss its rate of convergence. This
approach is shown to be appropriate for the study of discrete or
absolutely continuous random measures. We then focus on the analysis
of probability measures supported on the real line. In this setting,
we propose an ecient minimization algorithm based on accelerated
gradient descent for the computation of regularized Wasserstein
barycenters. This approach is nally applied to the statistical
analysis of simulated and real data sets.
Lunch
break
14h00-14h45: F. Iutzeler. Modified
fixed points iterations and applications to randomized and
accelerated optimization algorithms.
Abstract
Monotone operators have shown to be very interesting mathematical
objects and enabled the derivation of efficient algorithms. In this
talk, we will rely on simple fixed points iterations of an operator
with some contraction property enabling monotone convergence to some
fixed point. Then, we will see how deterministic and stochastic
variations of these iterations enable to derive randomized or
accelerated versions of popular algorithms such as the proximal
gradient or ADMM. More precisely, we will focus on i) randomization
to develop stochastic optimization algorithms; and ii) inertia to
accelerate the convergence.
14h45-15h30:
C. Boyer. Adapting
to unknown noise level in super-resolution.
Abstract
We study sparse spikes deconvolution over the space of
complex-valued measures when the input measure is a finite sum of
Dirac masses. We introduce a new procedure to handle the spike
deconvolution when the noise level is unknown. Prediction and
localization results will be presented for this approach. An insight
on the probabilistic tools used in the proofs could be briefly given
as well.
15h30-16h00:
C. Poon. Geometric
properties of solutions to the total variation denoising problem.
Abstract
Total variation (TV) denoising has been extensively studied in
imaging sciences since its introduction by Rudin, Osher and Fatemi
in 1992. However, the majority of theoretical works TV
regularization are actually concerned with solutions to total
variation denoising in the absence of noise. For instance, it is
known that the jump set of the regularized solutions are contained
in the jump set of the original data, but this knowledge is fairly
meaningless in the presence of noise since the noisy function can
introduce arbitrarily many new discontinuities. Furthermore, works
that consider the impact of noise on TV regularized solutions
typically derive integral estimates such as L^2 error bounds or
bounds on the total variation of the solution. However, such
estimates do not inform on the proximity of the gradient support of
the regularized solution to that of the clean function.
This talk is concerned with the impact of noise on the regularized
solutions. In particular, we consider stability in the gradient
support of solutions and address the following question: Given a
small neighbourhood around the support of Df, when will the gradient
support of the regularized solution be contained inside this
neighbourhood if the noise function has sufficiently small L^2 norm?
Coffee
break
16h30-17h15: B. Wirth.
Optimal design of transport networks.
Abstract
Several applications in biology and engineering are concerned with
the optimal transportation of substances from source locations to
sink locations. As opposed to classical optimal transport, models
for transport networks take into account that it is more efficient
to transport material in bulk. The resulting optimal transport
networks typically have a branching structure. We discuss
different model formulations and their equivalence as well as the
geometry of (almost) optimal networks, which can be analyzed by
proving energy scaling laws in the regime of small preference for
bulk transport.
17H15-17h45: L. Chizat. Scaling
algorithm for unbalanced optimal transport.
Abstract
This talk has three aims: (i) to describe the so-called unbalanced
optimal transport problem, (ii) to introduce a scaling algorithm
which generalizes Sinkorn's algorithm and solves this problem with
a linear convergence rate, and (iii) to showcase applications
which range from shape interpolation and colour transfer to the
simulation of very degenerate evolution PDEs.
This is joint work with Gabriel Peyré, Bernhard Schmitzer and
François-Xavier Vialard.
Wednesday 12th
9h45-10h30: B.
Levy.
Semi-discrete
Optimal Transport in 3D: A Numerical Algorithm and Some
Applications. Abstract
Semi-discrete algorithms seem to be a promising avenue for
developing efficient numerical methods for solving the
Monge-Ampère equation: Optimal Transport between a piecewise
linear density and a sum of Dirac masses can be directly
translated into a numerical algorithm, that maximizes a concave
function [Alexandrov, Aurenhammer, Mc Cann & Gangbo, Merigot].
The gradient and Hessian of this objective function involve
integrals over the intersection between a Laguerre diagram and the
tetrahedral mesh that supports the input source piecewise linear
density. I will present an efficient algorithm to compute these
integrals, as well as its use by a Newton-type solver
[Kitagawa,Merigot,Thibert], and some applications in fluid
dynamics [Gallouet Merigot].
Coffee break
11h00-11h45: A.
Gramfort. Optimal
transport for statistics problem in neuroimaging.
Abstract
Knowing how the Human brain is anatomically and functionally
organized at the level of a group of healthy individuals or
patients is the primary goal of neuroimaging research. Yet a
number of statistical challenges emerge when processing
functional brain imaging data such as functional MRI (fMRI),
electroencephalography (EEG) or magnetoencephalography (MEG).
One of them is computing an average of brain imaging data
defined over a voxel grid or a triangulation in order to
estimate a prototypical brain activation map for the
population. The problem is that data are large, the geometry
of the brain is complex and the between subjects variability
leads to spatially or temporally non-overlapping activation
foci. To address the problem of variability, data are commonly
smoothed before averaging, but this goes against the objective
of brain research to achieve better spatiotemporal resolution.
In this talk, I will show how using optimal transport and
Wasserstein barycenters one can average fMRI and MEG brain
activations without blurring the results. A specifity of the
approach is to be able to handle non-normalized data.
In a second part of the talk, I will present some work in
progress where we use optimal transport with entropic
smoothing and some capacity constraint to address the problem
of domain adaptation (DA) in machine learning. In biomedical
applications shifts in distributions between train data and
test data are quite common. For example test data could be
coming from a different machine or from a very different
cohort of patients. This is the problem that DA aims to solve
and we propose to use optimal transport for this.
11h45-12h15: K.
Modin. The
polar decomposition as the limit of a lifted entropy
gradient flow. Abstract
The work of Yann Brenier shows that optimal mass transport
(OMT) in the L2 sense is intrinsically tied to
polar decompositions. As a toy problem, one may consider OMT
in the category of multivariate Gaussian densities and the
associated polar decomposition A=PQ of real matrices. In this
talk I show that the P component can be obtained as the limit
of a gradient flow with respect to a lifted relative entropy
functional, thereby providing a new, geometric proof of the
polar decomposition of matrices. I shall also present a
corresponding infinite-dimensional lifted entropy gradient
flow on the space of convex functions. Here, the first
component in the polar decomposition of maps is an
equilibrium. Furthermore, I give a result hinting that the
infinite-dimensional flow has a unique limit in the smooth
category if the destination marginal is log-convex.
Lunch
break
14h00-14h45: E.
Chouzenoux. Convergence
analysis of a stochastic majorize-minimize memory gradient
algorithm. Abstract
Stochastic approximation techniques play a prominent role in
solving many large scale problems encountered in machine
learning or image/signal processing. In these contexts, the
statistics of the data are often unknown a priori or their
direct computation is too intensive, and they have thus to be
estimated online from the observations. For batch optimization
of an objective function being the sum of a data fidelity term
and a penalization (e.g. a sparsity promoting function),
Majorize-Minimize (MM) methods have recently attracted much
interest since they are fast, highly flexible, and effective
in ensuring convergence. The goal of this work is to show how
these methods can be successfully extended to the case when
the data fidelity term corresponds to a least squares
criterion and the cost function is replaced by a sequence of
stochastic approximations of it. In this context, we propose
an online version of an MM subspace algorithm and we establish
its convergence by using suitable probabilistic tools. We also
provide new results on the convergence rate of such kind of
algorithm. Numerical results illustrate the good practical
performance of the proposed algorithm associated with a memory
gradient subspace, when applied to both non-adaptive and
adaptive linear system identification scenarios.
14h45-15h15:
P. Tan. Accelerated
Alternating Descent Methods for Dystra-like Problems.
Abstract
In this talk, we present an alternating descent scheme for
problems involving nonsmooth objectives with a quadratic
coupling of the variables. Our algorithm performs in each
variable several descent steps, which improves the
performances compared to a single step. Thanks to a FISTA-like
trick, this scheme can also be accelerated, An application of
this work is the implementation of a fast parallel solver for
the proximity operator of the Total Variation for color
images. In this case, this method can be interpreted as an
inexact scheme.
15h15-15h45:
Q. Denoyelle. Support Recovery for Sparse Super-resolution of Positive Measures.
Abstract
We study sparse spikes super-resolution over the space of Radon measures on the real line or the torus when the input measure is a finite sum of positive Dirac masses using the BLASSO convex program. We focus on the recovery properties of the support and the amplitudes of the initial measure in the presence of noise as a function of the minimum separation of the input measure (the minimum distance between two spikes).
15H45-16h15: M.
Schmitz. Feature learning using Optimal Transport.
Abstract
Accounting for and correcting an instrument’s Point
Spread Function (PSF) is a key element in many astronomical
missions, including the upcoming Euclid survey where one of
the key objectives is to perform galaxy shape measurements for
weak lensing. Because of the non-linear way in which the PSF
varies across the field of view, and the way in which Optimal
Transport distances account for geometric warping, using them
for PSF-related problems is a very promising prospect. Using
recent numerical schemes to compute approximate Wasserstein
distances, and thus Wasserstein Barycentres (Sliced
Wasserstein Barycentres and entropic penalty), we perform
feature learning on images, and apply it in particular to PSFs
and galaxy images.
Coffee
break and conclusion of the Workshop