Lipschitz-free spaces
Abstract: Let M be a pointed metric space and Lip0 (M ) the space of Lipschitz functions vanishing at 0. Endowed with the Lipschitz norm this space is a Banach space. Denote F(M ) the closed subspace of Lip(M )* spanned by the evaluation points and call it the Lipschitz-free space over M. After an introduction explaining how one can use these spaces in the context of non linear classification of Banach spaces, we will more particularly be interested in dual Lipschitz-free spaces and shortly explain there link with optimal transportation.
Read MoreRecent advances on the acceleration of first-order methods in convex optimization
Abstract: Let f : H → R ∪ {+∞} be a proper lower-semicontinuous convex function defined on a Hilbert space H (think of RN), and let (xk) be a sequence in H generated by means of a “typical” first-order method, and intended to minimize f. If argmin(f) ̸= ∅, then (xk) will converge weakly, as k → +∞, to a minimizer of f, with a worst-case theoretical convergence rate of f(xk) − min(f) = O(1/k). In the 1980’s Y. Nesterov came up with a revolutionary – yet remarkably simple! – idea to modify the computation of the iterates with essentially the same computational cost, in order to...
Read MoreIncremental Proximal and Augmented Lagrangian Methods for Convex Optimization: A Survey
Abstract: Incremental methods deal effectively with an optimization problem of great importance in machine learning, signal processing, and large-scale and distributed optimization: the minimization of the sum of a large number of convex functions. We survey these methods and we propose incremental aggregated and nonaggregated versions of the proximal algorithm. Under cost function differentiability and strong convexity assumptions, we show linear convergence for a sufficiently small constant stepsize. This result also applies to distributed asynchronous variants of the method, involving...
Read MoreWhat is…the inverse function theorem
Abstract: After a gentle introduction to the paradigm of the inverse function theorem, advanced version of this theorem will be presented for nonsmooth functions and set-valued mappings.
Read More“Recovering a function from a subdifferential”
Abstract: The issue of recovering a function from its derivative or one of its directional derivatives is a central issue that dates back to the seminal work of Lebesgue (1904). The question of recovering a function from its subdifferential is more recent and has been the subject of intensive research in recent years, since the seminal work of Moreau and Rockafellar (1970) on convex functions, up to the many successive works of Thibault and his co-authors on increasingly broad classes of functions beyond the convex functions. The talk focuses on this issue. We divided...
Read MoreColoquio Optimización A mediodía de Optimización y análisis variacional”
EXPOSITORES – Pr. Dimitri Bertsekas, Departamento de Ingeniería Eléctrica e Informática en la Facultad de Ingeniería en el Instituto de Tecnología de Massachusetts (MIT), en Cambridge, Massachusetts. – Pr. Assen Dontchev, Mathematical Reviews y la Universidad de Michigan. – Pr. R. Tyrrell Rockafellar, Departamentos de matemáticas y matemáticas aplicadas en la Universidad de Washington, Seattle.
Read More



Noticias en español
