BEGIN:VCALENDAR
VERSION:2.0
PRODID:researchseminars.org
CALSCALE:GREGORIAN
X-WR-CALNAME:researchseminars.org
BEGIN:VEVENT
SUMMARY:Cláudia Soares (Instituto Superior Técnico and ISR)
DTSTART:20200514T163000Z
DTEND:20200514T173000Z
DTSTAMP:20260422T212558Z
UID:MPML/1
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/1/">The
  learning machine and beyond: a tour for the curious</a>\nby Cláudia Soar
 es (Instituto Superior Técnico and ISR) as part of Mathematics\, Physics 
 and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nThis talk will draw a f
 ew perspectives on the broad topic of Machine Learning\, with non-speciali
 sts in mind. We will go through major subfields like supervised\, unsuperv
 ised\, or active learning\, never forgetting the emergent reinforcement le
 arning. We will cover a few different trends over recent years\, like the 
 mathematically inclined Support Vector Machine\, or the empirical Deep Lea
 rning.\n
LOCATION:https://researchseminars.org/talk/MPML/1/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Afonso Bandeira (ETH Zurich)
DTSTART:20200604T163000Z
DTEND:20200604T173000Z
DTSTAMP:20260422T212558Z
UID:MPML/2
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/2/">Com
 putation\, statistics\, and optimization of random functions</a>\nby Afons
 o Bandeira (ETH Zurich) as part of Mathematics\, Physics and Machine Learn
 ing (IST\, Lisbon)\n\n\nAbstract\nWhen faced with a data analysis\, learni
 ng\, or statistical inference problem\, the amount and quality of data ava
 ilable fundamentally determines whether such tasks can be performed with c
 ertain levels of accuracy. Indeed\, many theoretical disciplines study lim
 its of such tasks by investigating whether a dataset effectively contains 
 the information of interest. With the growing size of datasets however\, i
 t is crucial not only that the underlying statistical task is possible\, b
 ut also that is doable by means of efficient algorithms. In this talk we w
 ill discuss methods aiming to establish limits of when statistical tasks a
 re possible with computationally efficient methods or when there is a fund
 amental Statistical-to-Computational gap in which an inference task is sta
 tistically possible but inherently computationally hard.\n\nThis is intima
 tely related to understanding the geometry of random functions\, with conn
 ections to statistical physics\, study of spin glasses\, random geometry\;
  and in an important example\, algebraic invariant theory.\n
LOCATION:https://researchseminars.org/talk/MPML/2/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Marcelo Pereyra (Heriot-Watt University)
DTSTART:20200611T163000Z
DTEND:20200611T173000Z
DTSTAMP:20260422T212558Z
UID:MPML/3
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/3/">Eff
 icient Bayesian computation by proximal Markov chain Monte Carlo: when Lan
 gevin meets Moreau</a>\nby Marcelo Pereyra (Heriot-Watt University) as par
 t of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstra
 ct\nThis talk summarises some new developments in Bayesian statistical met
 hodology for performing inference in high-dimensional inverse problems wit
 h an underlying convex geometry. We pay particular attention to problems r
 elated to imaging sciences and to new stochastic computation methods that 
 tightly combine proximal convex optimisation and Markov chain Monte Carlo 
 sampling techniques. The new computation methods are illustrated with a ra
 nge of imaging experiments\, where they are used to perform uncertainty qu
 antification analyses\, automatically adjust regularisation parameters\, a
 nd objectively compare alternative models in the absence of ground truth.\
 n
LOCATION:https://researchseminars.org/talk/MPML/3/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Csaba Szepesvári (University of Alberta and DeepMind)
DTSTART:20200625T163000Z
DTEND:20200625T173000Z
DTSTAMP:20260422T212558Z
UID:MPML/5
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/5/">Con
 fident Off-Policy Evaluation and Selection through Self-Normalized Importa
 nce Weighting</a>\nby Csaba Szepesvári (University of Alberta and DeepMin
 d) as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\
 n\nAbstract\nOff-policy evaluation is the problem of predicting the value 
 of a policy given some batch of data. In the language of statistics\, this
  is also called counterfactual estimation. Batch policy optimization refer
 s to the problem of finding a good policy\, again\, given some logged data
 .\nIn this talk\, I will consider the case of contextual bandits\, give a 
 brief (and incomplete) review of the approaches proposed in the literature
  and explain why this problem is difficult. Then\, I will describe a new a
 pproach based on self-normalized importance weighting. In this approach\, 
 a semi-empirical Efron-Stein concentration inequality is combined with Har
 ris' inequality to arrive at non-vacuous high-probability value lower boun
 ds\, which can then be used in a policy selection phase. On a number of sy
 nthetic and real datasets this new approach is found to be significantly s
 uperior than its main competitors\, both in terms of tightness of the conf
 idence intervals and the quality of the policies chosen. \n\nThe talk is b
 ased on joint work with Ilja Kuzborskij\, Claire Vernade and Andras Gyorgy
 .\n
LOCATION:https://researchseminars.org/talk/MPML/5/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Kyle Cranmer (NYU)
DTSTART:20200702T163000Z
DTEND:20200702T173000Z
DTSTAMP:20260422T212558Z
UID:MPML/6
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/6/">On 
 the Interplay between Physics and Deep Learning.</a>\nby Kyle Cranmer (NYU
 ) as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n
 \nAbstract\nThe interplay between physics and deep learning is typically d
 ivided into two themes.\nThe first is “physics for deep learning”\, wh
 ere techniques from physics are brought to bear on understanding dynamics 
 of learning. The second is “deep learning for physics\,” which focuses
  on application of deep learning techniques to physics problems. I will pr
 esent a more nuanced view of this interplay with examples of how the struc
 ture of physics problems have inspired advances in deep learning and how i
 t yields insights on topics such as inductive bias\, interpretability\, an
 d causality.\n
LOCATION:https://researchseminars.org/talk/MPML/6/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Hilbert Johan Kappen (Donder Institute\, Radboud University Nijmeg
 en\, the Netherlands)
DTSTART:20200528T163000Z
DTEND:20200528T173000Z
DTSTAMP:20260422T212558Z
UID:MPML/7
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/7/">Pat
 h integral control theory</a>\nby Hilbert Johan Kappen (Donder Institute\,
  Radboud University Nijmegen\, the Netherlands) as part of Mathematics\, P
 hysics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nStochastic optim
 al control theory deals with the problem to compute an optimal set of acti
 ons to attain some future goal. Examples are found in many contexts such a
 s motor control tasks for robotics\, planning and scheduling tasks or mana
 ging a financial portfolio. The computation of the optimal control is typi
 cally very difficult due to the size of the state space and the stochastic
  nature of the problem. Special cases for which the computation is tractab
 le are linear dynamical systems with quadratic cost and deterministic cont
 rol problems. For a special class of non-linear stochastic control problem
 s\, the solution can be mapped onto a statistical inference problem. For t
 hese so-called path integral control problems the optimal cost-to-go solut
 ion of the Bellman equation is given by the minimum of a free energy. I wi
 ll give a high level introduction to the underlying theory and illustrate 
 with some examples from robotics and other areas.\n
LOCATION:https://researchseminars.org/talk/MPML/7/
END:VEVENT
BEGIN:VEVENT
SUMMARY:André David Mendes (CERN)
DTSTART:20200521T163000Z
DTEND:20200521T173000Z
DTSTAMP:20260422T212558Z
UID:MPML/8
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/8/">How
  we discovered the Higgs ahead of schedule - ML's role in unveiling the ke
 ystone of elementary particle physics</a>\nby André David Mendes (CERN) a
 s part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nA
 bstract\nn 2010\, when the LHC started colliding proton pairs in earnest\,
  multi-variate analyses were newfangled methods starting to make inroads i
 n experimental particle physics. These methods faced widespread skepticism
  as to their performance and biases\, reflecting a winter of suspicion ove
 r overtrained neural networks that set in in the late 1990s. Thanks to mor
 e robust techniques\, like boosted decision trees\, it became possible to 
 make better and more extensive use of the full information recorded in par
 ticle collisions at the Tevatron and LHC colliders.\n\nThe Higgs boson dis
 covery by the CMS and ATLAS collaborations in 2012 was only possible becau
 se of the use of multi-variate techniques that enhanced the sensitivity by
  up to the equivalent of having 50% more collision data available for anal
 ysis.\n\nWe will review the use of classification and regression in the Hi
 ggs to diphoton search and subsequent discovery\, a concrete example of a 
 decade-old ML-based analysis in high-energy particle physics. Particular e
 mphasis will be placed in the modular design of the analysis and the inher
 ent explainability advantages\, used to great effect in assuaging concerns
  raised by hundreds of initially-skeptical colleagues in the CMS collabora
 tion.\nFinally\, we'll quickly highlight some particle physics challenges 
 that have contributed to\, and made use of\, the last decade of graph\, ad
 versarial\, and deep ML developments.\n
LOCATION:https://researchseminars.org/talk/MPML/8/
END:VEVENT
BEGIN:VEVENT
SUMMARY:João Miranda Lemos (Instituto Superior Técnico and INESC-ID)
DTSTART:20200716T163000Z
DTEND:20200716T173000Z
DTSTAMP:20260422T212558Z
UID:MPML/9
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/9/">Rei
 nforcement learning and adaptive control</a>\nby João Miranda Lemos (Inst
 ituto Superior Técnico and INESC-ID) as part of Mathematics\, Physics and
  Machine Learning (IST\, Lisbon)\n\n\nAbstract\nThe aim of this seminar is
  to explain\, to a wide audience\, how to combine optimal control techniqu
 es with reinforcement learning\, by using approximate dynamic programming\
 , and artificial neural networks\, to obtain adaptive optimal controllers.
  Although with roots since the end of the XX century\, this problem has be
 en the subject of an increasing attention. In addition to the promising to
 ols that it offers to tackle difficult nonlinear problems with major engin
 eering importance (ranging from robotics to biomedical engineering and bey
 hond)\, it has the charm of creating a meeting point between the control a
 nd machine learning research communities.\n
LOCATION:https://researchseminars.org/talk/MPML/9/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Francisco C. Santos (Instituto Superior Técnico and INESC-ID)
DTSTART:20200709T163000Z
DTEND:20200709T173000Z
DTSTAMP:20260422T212558Z
UID:MPML/10
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/10/">Cl
 imate action and cooperation dynamics under uncertainty</a>\nby Francisco 
 C. Santos (Instituto Superior Técnico and INESC-ID) as part of Mathematic
 s\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nWhen attemp
 ting to avoid global warming\, individuals often face a social dilemma in 
 which\, besides securing future benefits\, it is also necessary to reduce 
 the chances of future losses. In this talk\, I will resort to game theory 
 and populations of adaptive agents to offer a theoretical analysis of this
  type of dilemmas\, in which the risk of failure plays a central role in i
 ndividual decisions. I will discuss both deterministic dynamics in large p
 opulations\, and stochastic social learning dynamics in finite populations
 . This class of models can be shown to capture some of the essential featu
 res discovered in recent key experiments while allowing one to extend in n
 on-trivial ways the experimental conditions to regions of practical intere
 st. Moreover\, this approach leads us to identify useful parallels between
  ecological and socio-economic systems\, particularly in what concerns the
  evolution and self-organization of their institutions. Particularly\, our
  results suggest that global coordination for a common good should be atte
 mpted through a polycentric structure of multiple small-scale agreements\,
  in which perception of risk is high and uncertainty in collective goals i
 s minimized. Whenever the perception of risk is low\, our results indicate
  that sanctioning institutions may significantly enhance the chances of co
 ordinating to tame the planet's climate\, as long as they are implemented 
 in a bottom-up manner.  I will discuss the impact on public goods dilemmas
  of heterogeneous political networks and wealth inequality\, including dis
 tribution of wealth representative of existing inequalities among nations.
  Finally\, I will briefly discuss the impact of scientific uncertainty —
  both in what concerns the collective targets and the time window availabl
 e for action — on individuals' strategies and polarization of preference
 s.\n
LOCATION:https://researchseminars.org/talk/MPML/10/
END:VEVENT
BEGIN:VEVENT
SUMMARY:João Xavier (Instituto Superior Técnico and ISR)
DTSTART:20200618T163000Z
DTEND:20200618T173000Z
DTSTAMP:20260422T212558Z
UID:MPML/11
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/11/">Le
 arning from distributed datasets: an introduction with two examples</a>\nb
 y João Xavier (Instituto Superior Técnico and ISR) as part of Mathematic
 s\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nData are in
 creasingly measured\, in ever tinier minutiae\, by networks of spatially d
 istributed agents. Illustrative examples include a team of robots searchin
 g a large region\, a collection of sensors overseeing a critical infra-str
 ucture\, or a swarm of drones policing a wide area.\n\nHow to learn from t
 hese large\, spatially distributed datasets? In the centralized approach e
 ach agent forwards its dataset to a fusion center\, which then carries out
  the learning from the pile of amassed datasets. This approach\, however\,
  prevents the number of agents to scale up: as more and more agents ship d
 ata to the center\, not only the communication channels near the center qu
 ickly swell to congestion\, but also the computational power of the center
  is rapidly outpaced.\n\nIn this seminar\, I describe the alternative appr
 oach of distributed learning. Here\, no fusion center exists\, and the age
 nts themselves recreate the centralized computation by exchanging short me
 ssages (not data) between network neighbors. To illustrate\, I describe tw
 o learning algorithms: one solves convex learning problems via a token tha
 t randomly roams through the network\, and the other solves a classificati
 on problem via random meetings between agents (e.g.\, gossip)\, each agent
  measuring only its own stream of features.\n\nThis seminar is aimed at no
 n-specialists. Rather than trying to impart the latest developments of the
  field\, I hope to open a welcoming door to those wishing to have a peek a
 t this bubbling field of research\, where optimization\, control\, probabi
 lity\, and machine learning mingle happily.\n
LOCATION:https://researchseminars.org/talk/MPML/11/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Marylou Gabrié (Center for Data Science\, NYU and Flatiron Instit
 ute\, CCM)
DTSTART:20200723T163000Z
DTEND:20200723T173000Z
DTSTAMP:20260422T212558Z
UID:MPML/12
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/12/">Pr
 ogress and hurdles in the statistical mechanics of deep learning</a>\nby M
 arylou Gabrié (Center for Data Science\, NYU and Flatiron Institute\, CCM
 ) as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n
 \nAbstract\nUnderstanding the great performances of deep neural networks i
 s a very active direction of research with contributions coming from a wid
 e variety of fields. The statistical mechanics of learning is a theoretica
 l framework dating back to the 80s studying learning problems from a physi
 cist viewpoint and using tools from the physics of disordered systems. In 
 this talk\, I will first go over this traditional framework\, which relies
  on the teacher-student scenario\, bayesian analysis and mean-field approx
 imations. Then I will discuss some recent advances in the corresponding an
 alysis of modern deep neural network\, and highlight remaining challenges.
 \n
LOCATION:https://researchseminars.org/talk/MPML/12/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Masoud Mohseni (Google Quantum Artificial Intelligence Laboratory)
DTSTART:20200730T163000Z
DTEND:20200730T173000Z
DTSTAMP:20260422T212558Z
UID:MPML/13
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/13/">Te
 nsorFlow Quantum: An open source framework for hybrid quantum-classical ma
 chine learning.</a>\nby Masoud Mohseni (Google Quantum Artificial Intellig
 ence Laboratory) as part of Mathematics\, Physics and Machine Learning (IS
 T\, Lisbon)\n\n\nAbstract\nIn this talk\, I introduce TensorFlow Quantum (
 TFQ)\, an open source library that was launched by Google in March 2020\, 
 for the rapid prototyping of hybrid quantum-classical models for classical
  or quantum data.This framework offers high-level abstractions for the des
 ign\, training\, and testing of both discriminative and generative quantum
  models under TensorFlow and supports high-performance quantum circuit sim
 ulators. I provide an overview of the software architecture and building b
 locks through several examples and illustrate TFQ functionalities via cons
 tructing hybrid quantum-classical convolutional neural networks for quantu
 m state classification.\n
LOCATION:https://researchseminars.org/talk/MPML/13/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Gunnar Carlsson (Stanford University)
DTSTART:20200930T170000Z
DTEND:20200930T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/14
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/14/">To
 pological Data Analysis and Deep Learning</a>\nby Gunnar Carlsson (Stanfor
 d University) as part of Mathematics\, Physics and Machine Learning (IST\,
  Lisbon)\n\n\nAbstract\nDeep Learning is a powerful collection of techniqu
 es for statistical learning\, which has shown dramatic applications in man
 y different directions\, including including the study of data sets of ima
 ges\, text\, and time series. It uses neural networks\, specifically convo
 lutional neural networks (CNN's)\, to produce these results. What we have 
 observed recently is that methods of topology can contribute to this effor
 t\, in diagnosing behavior within the CNN's\, in the design of neural netw
 orks with excellent computational properties\, and in improving generaliza
 tion\, i.e. the transfer of results of one neural network from one data se
 t to another of similar type. We'll discuss topological methods in data sc
 ience\, as well as there application to this interesting set of techniques
 .\n
LOCATION:https://researchseminars.org/talk/MPML/14/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Lindsey Gray (Fermi National Accelerator Laboratory)
DTSTART:20201014T170000Z
DTEND:20201014T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/15
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/15/">Gr
 aph Neural Networks for Pattern Recognition in Particle Physics</a>\nby Li
 ndsey Gray (Fermi National Accelerator Laboratory) as part of Mathematics\
 , Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nModern partic
 le physics detectors generate copious amounts of data packed with meaning 
 that provides the means for high-quality measurements in demanding experim
 ental environments. To achieve these measurements there is a trend towards
  finer granularity in these detectors and that implies the data read out h
 as less intrinsic structure. Accurate pattern recognition is required to d
 efine the signatures of particles within those detectors and simultaneousl
 y extract physical parameters for the particles. Typically\, algorithms to
  achieve these goals are written using well known unsupervised algorithms\
 , but recent advances in machine learning on graph structures\, "Graph Neu
 ral Networks" (GNNs)\, provide powerful new methodologies for designing pa
 ttern recognition algorithms. In particular\, methodologies for predicting
  the link structure between pieces of data from detectors are well suited 
 to the particle physics pattern recognition task. Furthermore\, there are 
 interesting avenues for enforcing known symmetries of the data into the ou
 tput of such networks and there is ongoing research in this direction. Thi
 s talk will discuss the challenges of pattern recognition\, the advent of 
 GNNs and the connections to particle physics\, and the paths of research a
 head for fully utilizing this powerful new tool.\n
LOCATION:https://researchseminars.org/talk/MPML/15/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Carola-Bibiane Schönlieb (DAMTP\, University of Cambridge)
DTSTART:20201120T150000Z
DTEND:20201120T160000Z
DTSTAMP:20260422T212558Z
UID:MPML/16
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/16/">Co
 mbining knowledge and data driven methods for solving inverse imaging prob
 lems - getting the best from both worlds</a>\nby Carola-Bibiane Schönlieb
  (DAMTP\, University of Cambridge) as part of Mathematics\, Physics and Ma
 chine Learning (IST\, Lisbon)\n\n\nAbstract\nInverse problems in imaging r
 ange from tomographic reconstruction (CT\, MRI\, etc) to image deconvoluti
 on\, segmentation\, and classification\, just to name a few. In this talk 
 I will discuss\napproaches to inverse imaging problems which have both a m
 athematical modelling (knowledge driven) and a machine learning (data-driv
 en) component. Mathematical modelling is crucial in the presence of ill-po
 sedness\, making use of information about the imaging data\, for narrowing
  down the search space. Such an approach results in highly generalizable r
 econstruction and analysis methods which come with desirable solutions gua
 rantees. Machine learning on the other hand is a powerful tool for customi
 sing methods to individual data sets. Highly parametrised models such as d
 eep neural networks in particular\, are powerful tools for accurately mode
 lling prior information about solutions. The combination of these two para
 digms\, getting the best from both of these worlds\, is the topic of this 
 talk\, furnished with examples for image classification under minimal supe
 rvision and for tomographic image reconstruction.\n
LOCATION:https://researchseminars.org/talk/MPML/16/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Weinan E (Princeton University)
DTSTART:20201007T100000Z
DTEND:20201007T110000Z
DTSTAMP:20260422T212558Z
UID:MPML/17
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/17/">Ma
 chine Learning and Scientific Computing</a>\nby Weinan E (Princeton Univer
 sity) as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)
 \n\n\nAbstract\nNeural network-based deep learning is capable of approxima
 ting functions in very high dimension with unprecedented efficiency and ac
 curacy. This has opened up many exciting new possibilities\, not just in t
 raditional areas of artificial intelligence\, but also in scientific compu
 ting and computational science. At the same time\, deep learning has also 
 acquired the reputation of being a set of “black box” type of tricks\,
  without fundamental principles. This has been a real obstacle for making 
 further progress in machine learning.\n\nIn this talk\, I will try to addr
 ess the following two questions:\n\n1. How machine learning will impact co
 mputational mathematics and computational science?\n\n2. How computational
  mathematics\, particularly numerical analysis\, can impact machine learni
 ng? We describe some of the most important progresses that have been made 
 on these issues so far.\n\nOur hope is to put things into a perspective th
 at will help to integrate machine learning with computational science.\n
LOCATION:https://researchseminars.org/talk/MPML/17/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Gitta Kutyniok (Mathematical Institute of the University of Munich
 )
DTSTART:20201202T180000Z
DTEND:20201202T190000Z
DTSTAMP:20260422T212558Z
UID:MPML/18
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/18/">De
 ep Learning meets Physics: Taking the Best out of Both Worlds in Imaging S
 cience</a>\nby Gitta Kutyniok (Mathematical Institute of the University of
  Munich) as part of Mathematics\, Physics and Machine Learning (IST\, Lisb
 on)\n\n\nAbstract\nPure model-based approaches are today often insufficien
 t for solving complex inverse problems in imaging. At the same time\, we w
 itness the tremendous success of data-based methodologies\, in particular\
 , deep neural networks for such problems. However\, pure deep learning app
 roaches often neglect known and valuable information from physics.\n\nIn t
 his talk\, we will provide an introduction to this problem complex and the
 n discuss a general conceptual approach to inverse problems in imaging\, w
 hich combines deep learning and physics. This hybrid approach is based on 
 shearlet-based sparse regularization and deep learning and is guided by a 
 microlocal analysis viewpoint to pay particular attention to the singulari
 ty structures of the data. Finally\, we will present several applications 
 such as tomographic reconstruction and show that our approach outperforms 
 previous methodologies\, including methods entirely based on deep learning
 .\n
LOCATION:https://researchseminars.org/talk/MPML/18/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Tommaso Dorigo (Italian Institute for Nuclear Physics)
DTSTART:20201125T180000Z
DTEND:20201125T190000Z
DTSTAMP:20260422T212558Z
UID:MPML/19
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/19/">De
 aling with Systematic Uncertainties in HEP Analysis with Machine Learning 
 Methods</a>\nby Tommaso Dorigo (Italian Institute for Nuclear Physics) as 
 part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbs
 tract\nI will discuss the impact of nuisance parameters on the effectivene
 ss of supervised classification in high energy physics problems\, and tech
 niques that may mitigate or remove their effect in the search for optimal 
 selection criteria and variable transformations. The approaches discussed 
 include nuisance parametrized models\, modified or adversary losses\, semi
  supervised learning approaches and inference-aware techniques.\n
LOCATION:https://researchseminars.org/talk/MPML/19/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Florent Krzakala (EPFL)
DTSTART:20201028T180000Z
DTEND:20201028T190000Z
DTSTAMP:20260422T212558Z
UID:MPML/20
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/20/">So
 me exactly solvable models for statistical machine learning</a>\nby Floren
 t Krzakala (EPFL) as part of Mathematics\, Physics and Machine Learning (I
 ST\, Lisbon)\n\n\nAbstract\nThe increasing dimensionality of data in the m
 odern machine learning age presents new challenges and opportunities. The 
 high-dimensional settings allow one to use powerful asymptotic methods fro
 m probability theory and statistical physics to obtain precise characteriz
 ations and develop new algorithmic approaches. There is indeed a decades-l
 ong tradition in statistical physics with building and solving such simpli
 fied models of neural networks.\n\nI will give examples of recent works th
 at build on powerful methods of physics of disordered systems to analyze d
 ifferent problems in machine learning and neural networks\, including over
 parameterization\, kernel methods\, and the gradient descent algorithm in 
 a high dimensional non-convex setting.\n
LOCATION:https://researchseminars.org/talk/MPML/20/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Joan Bruna (Courant Institute and Center for Data Science\, NYU)
DTSTART:20201104T180000Z
DTEND:20201104T190000Z
DTSTAMP:20260422T212558Z
UID:MPML/21
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/21/">Ma
 thematical aspects of neural network learning through measure dynamics</a>
 \nby Joan Bruna (Courant Institute and Center for Data Science\, NYU) as p
 art of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbst
 ract\nHigh-dimensional learning remains an outstanding phenomena where exp
 erimental evidence outpaces our current mathematical understanding\, mostl
 y due to the recent empirical successes of Deep Learning algorithms. Neura
 l Networks provide a rich yet intricate class of functions with statistica
 l abilities to break the curse of dimensionality\, and where physical prio
 rs can be tightly integrated into the architecture to improve sample effic
 iency. Despite these advantages\, an outstanding theoretical challenge in 
 these models is computational\, ie providing an analysis that explains suc
 cessful optimization and generalization in the face of existing worst-case
  computational hardness results.\n\nIn this talk\, I will focus on the fra
 mework that lifts parameter optimization to an appropriate measure space. 
 I will cover existing results that guarantee global convergence of the res
 ulting Wasserstein gradient flows\, as well as recent results that study t
 ypical fluctuations of the dynamics around their mean field evolution. We 
 will also discuss extensions of this framework beyond vanilla supervised l
 earning\, to account for symmetries in the function\, as well as for compe
 titive optimization.\n
LOCATION:https://researchseminars.org/talk/MPML/21/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Bin Dong (BICMR\, Peking University)
DTSTART:20201111T110000Z
DTEND:20201111T120000Z
DTSTAMP:20260422T212558Z
UID:MPML/22
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/22/">Le
 arning and Learning to Solve PDEs</a>\nby Bin Dong (BICMR\, Peking Univers
 ity) as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\
 n\n\nAbstract\nDeep learning continues to dominate machine learning and ha
 s been successful in computer vision\, natural language processing\, etc. 
 Its impact has now expanded to many research areas in science and engineer
 ing. In this talk\, I will mainly focus on some recent impact of deep lear
 ning on computational mathematics. I will present our recent work on bridg
 ing deep neural networks with numerical differential equations. On the one
  hand\, I will show how to design transparent deep convolutional networks 
 to uncover hidden PDE models from observed dynamical data. On the other ha
 nd\, I will present our preliminary attempt to establish a deep reinforcem
 ent learning based framework to solve 1D scalar conservation laws\, and a 
 meta-learning approach for solving linear parameterized PDEs based on the 
 multigrid method.\n
LOCATION:https://researchseminars.org/talk/MPML/22/
END:VEVENT
BEGIN:VEVENT
SUMMARY:René Vidal (Mathematical Institute for Data Science\, Johns Hopki
 ns University)
DTSTART:20201216T180000Z
DTEND:20201216T190000Z
DTSTAMP:20260422T212558Z
UID:MPML/23
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/23/">Fr
 om Optimization Algorithms to Dynamical Systems and Back</a>\nby René Vid
 al (Mathematical Institute for Data Science\, Johns Hopkins University) as
  part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAb
 stract\nRecent work has shown that tools from dynamical systems can be use
 d to analyze accelerated optimization algorithms. For example\, it has bee
 n shown that the continuous limit of Nesterov’s accelerated gradient (NA
 G) gives an ODE whose convergence rate matches that of NAG for convex\, un
 constrained\, and smooth problems. Conversely\, it has been shown that NAG
  can be obtained as the discretization of an ODE\, however since different
  discretizations lead to different algorithms\, the choice of the discreti
 zation becomes important. The first part of this talk will extend this typ
 e of analysis to convex\, constrained and non-smooth problems by using Lya
 punov stability theory to analyze continuous limits of the Alternating Dir
 ection Method of Multipliers (ADMM). The second part of this talk will sho
 w that many existing and new optimization algorithms can be obtained by su
 itably discretizing a dissipative Hamiltonian. As an example\, we will pre
 sent a new method called Relativistic Gradient Descent (RGD)\, which empir
 ically outperforms momentum\, RMSprop\, Adam and AdaGrad on several non-co
 nvex\nproblems.\n\nThis is joint work with Guilherme Franca\, Daniel Robin
 son and Jeremias Sulam.\n
LOCATION:https://researchseminars.org/talk/MPML/23/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Mauro Maggioni (Johns Hopkins University)
DTSTART:20201021T170000Z
DTEND:20201021T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/24
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/24/">Le
 arning Interaction laws in particle- and agent-based systems</a>\nby Mauro
  Maggioni (Johns Hopkins University) as part of Mathematics\, Physics and 
 Machine Learning (IST\, Lisbon)\n\n\nAbstract\nInteracting agent-based sys
 tems are ubiquitous in science\, from modeling of particles in Physics to 
 prey-predator and colony models in Biology\, to opinion dynamics in econom
 ics and social sciences. Oftentimes the laws of interactions between the a
 gents are quite simple\, for example they depend only on pairwise interact
 ions\, and only on pairwise distance in each interaction. We consider the 
 following inference problem for a system of interacting particles or agent
 s: given only observed trajectories of the agents in the system\, can we l
 earn what the laws of interactions are? We would like to do this without a
 ssuming any particular form for the interaction laws\, i.e. they might be 
 "any" function of pairwise distances. We consider this problem both the me
 an-field limit (i.e. the number of particles going to infinity) and in the
  case of a finite number of agents\, with an increasing number of observat
 ions\, albeit in this talk we will mostly focus on the latter case. We cas
 t this as an inverse problem\, and study it in the case where the interact
 ion is governed by an (unknown) function of pairwise distances. We discuss
  when this problem is well-posed\, and we construct estimators for the int
 eraction kernels with provably good statistically and computational proper
 ties. We measure their performance on various examples\, that include exte
 nsions to agent systems with different types of agents\, second-order syst
 ems\, and families of systems with parametric interaction kernels. We also
  conduct numerical experiments to test the large time behavior of these sy
 stems\, especially in the cases where they exhibit emergent behavior.\n\nT
 his is joint work with F. Lu\, J.Miller\, S. Tang and M. Zhong.\n
LOCATION:https://researchseminars.org/talk/MPML/24/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Xavier Bresson (Nanyang Technological University)
DTSTART:20210127T110000Z
DTEND:20210127T120000Z
DTSTAMP:20260422T212558Z
UID:MPML/25
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/25/">Be
 nchmarking Graph Neural Networks</a>\nby Xavier Bresson (Nanyang Technolog
 ical University) as part of Mathematics\, Physics and Machine Learning (IS
 T\, Lisbon)\n\n\nAbstract\nGraph neural networks (GNNs) have become the st
 andard toolkit for analyzing and learning from data on graphs. As the fiel
 d grows\, it becomes critical to identify key architectures and validate n
 ew ideas that generalize to larger\, more complex datasets. Unfortunately\
 , it has been increasingly difficult to gauge the effectiveness of new mod
 els in the absence of a standardized benchmark with consistent experimenta
 l settings. In this work\, we introduce a reproducible GNN benchmarking fr
 amework\, with the facility for researchers to add new models conveniently
  for arbitrary datasets. We demonstrate the usefulness of our framework by
  presenting a principled investigation into the recent Weisfeiler-Lehman G
 NNs (WL-GNNs) compared to message passing-based graph convolutional networ
 ks (GCNs) for a variety of graph tasks with medium-scale datasets.\n
LOCATION:https://researchseminars.org/talk/MPML/25/
END:VEVENT
BEGIN:VEVENT
SUMMARY:James Halverson (Northeastern University)
DTSTART:20210120T180000Z
DTEND:20210120T190000Z
DTSTAMP:20260422T212558Z
UID:MPML/26
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/26/">Ne
 ural Networks and Quantum Field Theory</a>\nby James Halverson (Northeaste
 rn University) as part of Mathematics\, Physics and Machine Learning (IST\
 , Lisbon)\n\n\nAbstract\nIn this talk I will review essentials of quantum 
 field theory (QFT) and demonstrate how the function-space distribution of 
 many neural networks (NNs) shares similar properties. This allows\, for in
 stance\, computation of correlators of neural network outputs in terms of 
 Feynman diagrams and a direct analogy between non-Gaussian corrections in 
 NN distributions and particle interactions. Some cases yield divergences i
 n perturbation theory\, requiring the introduction of regularization and r
 enormalization. Potential advantages of this perspective will be discussed
 \, including a duality between function-space and parameter-space descript
 ions of neural networks.\n
LOCATION:https://researchseminars.org/talk/MPML/26/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Anna C. Gilbert (Yale University)
DTSTART:20210113T180000Z
DTEND:20210113T190000Z
DTSTAMP:20260422T212558Z
UID:MPML/27
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/27/">Me
 tric representations: Algorithms and Geometry</a>\nby Anna C. Gilbert (Yal
 e University) as part of Mathematics\, Physics and Machine Learning (IST\,
  Lisbon)\n\n\nAbstract\nGiven a set of distances amongst points\, determin
 ing what metric representation is most "consistent" with the input distanc
 es or the metric that best captures the relevant geometric features of the
  data is a key step in many machine learning algorithms. In this talk\, we
  focus on 3 specific metric constrained problems\, a class of optimization
  problems with metric constraints: metric nearness (Brickell et al. (2008)
 )\, weighted correlation clustering on general graphs (Bansal et al. (2004
 ))\, and metric learning (Bellet et al. (2013)\; Davis et al. (2007)).\n\n
 Because of the large number of constraints in these problems\, however\, t
 hese and other researchers have been forced to restrict either the kinds o
 f metrics learned or the size of the problem that can be solved. We provid
 e an algorithm\, PROJECT AND FORGET\, that uses Bregman projections with c
 utting planes\, to solve metric constrained problems with many (possibly e
 xponentially) inequality constraints. We also prove that our algorithm con
 verges to the global optimal solution. Additionally\, we show that the opt
 imality error decays asymptotically at an exponential rate. We show that u
 sing our method we can solve large problem instances of three types of met
 ric constrained problems\, out-performing all state of the art methods wit
 h respect to CPU times and problem sizes.\n\nFinally\, we discuss the adap
 tation of PROJECT AND FORGET to specific types of metric constraints\, nam
 ely tree and hyperbolic metrics.\n
LOCATION:https://researchseminars.org/talk/MPML/27/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Caroline Uhler (MIT and Institute for Data\, Systems and Society)
DTSTART:20210210T180000Z
DTEND:20210210T190000Z
DTSTAMP:20260422T212558Z
UID:MPML/28
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/28/">Ca
 usal Inference and Overparameterized Autoencoders in the Light of Drug Rep
 urposing for SARS-CoV-2</a>\nby Caroline Uhler (MIT and Institute for Data
 \, Systems and Society) as part of Mathematics\, Physics and Machine Learn
 ing (IST\, Lisbon)\n\n\nAbstract\nMassive data collection holds the promis
 e of a better understanding of complex phenomena and ultimately\, of bette
 r decisions. An exciting opportunity in this regard stems from the growing
  availability of perturbation / intervention data (drugs\, knockouts\, ove
 rexpression\,\netc.) in biology. In order to obtain mechanistic insights f
 rom such data\, a major challenge is the development of a framework that i
 ntegrates observational and interventional data and allows predicting the 
 effect of yet unseen interventions or transporting the effect of intervent
 ions observed in one context to another. I will present a framework for ca
 usal structure discovery based on such data and highlight the role of over
 parameterized autoencoders. We end by demonstrating how these ideas can be
  applied for drug repurposing in the current SARS-CoV-2 crisis.\n
LOCATION:https://researchseminars.org/talk/MPML/28/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Samantha Kleinberg (Stevens Institute of Technology)
DTSTART:20201209T180000Z
DTEND:20201209T190000Z
DTSTAMP:20260422T212558Z
UID:MPML/29
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/29/">Da
 ta\, Decisions\, and You: Making Causality Useful and Usable in a Complex 
 World</a>\nby Samantha Kleinberg (Stevens Institute of Technology) as part
  of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstrac
 t\nThe collection of massive observational datasets has led to unprecedent
 ed opportunities for causal inference\, such as using electronic health re
 cords to identify risk factors for disease. However\, our ability to under
 stand these complex data sets has not grown the same pace as our ability t
 o collect them. While causal inference has traditionally focused on pairwi
 se relationships between variables\, biological systems are highly complex
  and knowing when events may happen is often as important as knowing wheth
 er they will. In the first half of this talk I discuss new methods that al
 low causal relationships to be reliably inferred from complex observationa
 l data\, motivated by analysis of intensive care unit and other medical da
 ta. Causes are useful because they allow us to take action\, but how there
  is a gap between the output of machine learning and what helps people mak
 e decisions. In the second part of this talk I discuss our recent findings
  in testing just how people fare when using the output of machine learning
  and how we can go from data to knowledge to decisions.\n
LOCATION:https://researchseminars.org/talk/MPML/29/
END:VEVENT
BEGIN:VEVENT
SUMMARY:A. Pedro Aguiar (Faculdade de Engenharia\, Universidade do Porto)
DTSTART:20210303T180000Z
DTEND:20210303T190000Z
DTSTAMP:20260422T212558Z
UID:MPML/31
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/31/">Mo
 del based control design combining Lyapunov and optimization tools: Exampl
 es in the area of motion control of autonomous robotic vehicles</a>\nby A.
  Pedro Aguiar (Faculdade de Engenharia\, Universidade do Porto) as part of
  Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\n
 The past few decades have witnessed a significant research effort in the f
 ield of Lyapunov model based control design. In parallel\, optimal control
  and optimization model based design have also expanded their range of app
 lications\, and nowadays\, receding horizon approaches can be considered a
  mature field for particular classes of control systems.\nIn this talk\, I
  will argue that Lyapunov based techniques play an important role for anal
 ysis of model based optimization methodologies and moreover\, both approac
 hes can be combined for control design resulting in powerful frameworks wi
 th formal guarantees of robustness\, stability\, performance\, and safety.
  Illustrative examples in the area of motion control of autonomous robotic
  vehicles will be presented for Autonomous Underwater Vehicles (AUVs)\, Au
 tonomous Surface Vehicles (ASVs) and Unmanned Aerial Vehicles (UAVs).\n
LOCATION:https://researchseminars.org/talk/MPML/31/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Sanjeev Arora (Computer Science Department\, Princeton University)
DTSTART:20210106T180000Z
DTEND:20210106T190000Z
DTSTAMP:20260422T212558Z
UID:MPML/32
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/32/">Th
 e quest for mathematical understanding of deep learning</a>\nby Sanjeev Ar
 ora (Computer Science Department\, Princeton University) as part of Mathem
 atics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nDeep le
 arning has transformed Machine Learning and Artificial Intelligence in the
  past decade. It raises fundamental questions for mathematics and theory o
 f computer science\, since it relies upon solving large-scale nonconvex pr
 oblems via gradient descent and its variants. This talk will be an introdu
 ction to mathematical questions raised by deep learning\, and some partial
  understanding obtained in recent years with respect to optimization\, gen
 eralization\, self-supervised learning\, privacy etc.\n
LOCATION:https://researchseminars.org/talk/MPML/32/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Jan Peters (Technische Universitaet Darmstadt)
DTSTART:20210423T130000Z
DTEND:20210423T140000Z
DTSTAMP:20260422T212558Z
UID:MPML/33
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/33/">Ro
 bot Learning - Quo Vadis?</a>\nby Jan Peters (Technische Universitaet Darm
 stadt) as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon
 )\n\n\nAbstract\nAutonomous robots that can assist humans in situations of
  daily life have been a long standing vision of robotics\, artificial inte
 lligence\, and cognitive sciences. A first step towards this goal is to cr
 eate robots that can learn tasks triggered by environmental context or hig
 her level instruction. However\, learning techniques have yet to live up t
 o this promise as only few methods manage to scale to high-dimensional man
 ipulator or humanoid robots. In this talk\, we investigate a general frame
 work suitable for learning motor skills in robotics which is based on the 
 principles behind many analytical robotics approaches. It involves generat
 ing a representation of motor skills by parameterized motor primitive poli
 cies acting as building blocks of movement generation\, and a learned task
  module that transforms these movements into motor commands. We discuss le
 arning on three different levels of abstraction\, i.e.\, learning for accu
 rate control is needed to execute\, learning of motor primitives is needed
  to acquire simple movements\, and learning of the task-dependent „hyper
 parameters“ of these motor primitives allows learning complex tasks. We 
 discuss task-appropriate learning approaches for imitation learning\, mode
 l learning and reinforcement learning for robots with many degrees of free
 dom. Empirical evaluations on a several robot systems illustrate the effec
 tiveness and applicability to learning control on an anthropomorphic robot
  arm. These robot motor skills range from toy examples (e.g.\, paddling a 
 ball\, ball-in-a-cup\, juggling) to playing robot table tennis against a h
 uman being and manipulation of various objects.\n
LOCATION:https://researchseminars.org/talk/MPML/33/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Miguel Couceiro (Université de Lorraine)
DTSTART:20210203T180000Z
DTEND:20210203T190000Z
DTSTAMP:20260422T212558Z
UID:MPML/34
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/34/">Ma
 king ML Models fairer through explanations\, feature dropout\, and aggrega
 tion</a>\nby Miguel Couceiro (Université de Lorraine) as part of Mathemat
 ics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nAlgorithm
 ic decisions are now being used on a daily basis\, and based on Machine Le
 arning (ML) processes that may be complex and biased. This raises several 
 concerns given the critical impact that biased decisions may have on indiv
 iduals or on society as a whole. Not\nonly unfair outcomes affect human ri
 ghts\, they also undermine public trust in ML and AI. In this talk\, we wi
 ll address fairness issues of ML models based on decision outcomes\, and w
 e will show how the simple idea of "feature dropout" followed by an "ensem
 ble approach" can improve model fairness without compromising its accuracy
 . To illustrate we will present a general workflow that relies on explaine
 rs to tackle "process fairness"\, which essentially measures a model's rel
 iance on sensitive or discriminatory features. We will present different a
 pplications and empirical settings that show improvements not only with re
 spect to process fairness but also other fairness metrics.\n
LOCATION:https://researchseminars.org/talk/MPML/34/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Steve Brunton (University of Washington)
DTSTART:20210331T170000Z
DTEND:20210331T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/35
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/35/">Ma
 chine learning for Fluid Mechanics</a>\nby Steve Brunton (University of Wa
 shington) as part of Mathematics\, Physics and Machine Learning (IST\, Lis
 bon)\n\n\nAbstract\nMany tasks in fluid mechanics\, such as design optimiz
 ation and control\, are challenging because fluids are nonlinear and exhib
 it a large range of scales in both space and time. This range of scales ne
 cessitates exceedingly high-dimensional measurements and computational dis
 cretization to resolve all relevant features\, resulting in vast data sets
  and time-intensive computations. Indeed\, fluid dynamics is one of the or
 iginal big data fields\, and many high-performance computing architectures
 \, experimental measurement techniques\, and advanced data processing and 
 visualization algorithms were driven by decades of research in fluid mecha
 nics. Machine learning constitutes a growing set of powerful techniques to
  extract patterns and build models from this data\, complementing the exis
 ting theoretical\, numerical\, and experimental efforts in fluid mechanics
 . In this talk\, we will explore current goals and opportunities for machi
 ne learning in fluid mechanics\, and we will highlight a number of recent 
 technical advances. Because fluid dynamics is central to transportation\, 
 health\, and defense systems\, we will emphasize the importance of machine
  learning solutions that are interpretable\, explainable\, generalizable\,
  and that respect known physics.\n
LOCATION:https://researchseminars.org/talk/MPML/35/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Hsin Yuan Huang\, (Robert) (Caltech)
DTSTART:20210317T180000Z
DTEND:20210317T190000Z
DTSTAMP:20260422T212558Z
UID:MPML/36
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/36/">In
 formation-theoretic bounds on quantum advantage in machine learning</a>\nb
 y Hsin Yuan Huang\, (Robert) (Caltech) as part of Mathematics\, Physics an
 d Machine Learning (IST\, Lisbon)\n\n\nAbstract\nWe compare the complexity
  of training classical and quantum machine learning (ML) models for predic
 ting outcomes of physical experiments. The experiments depend on an input 
 parameter x and involve the execution of a (possibly unknown) quantum proc
 ess $E$. Our figure of merit is the number of runs of $E$ needed during tr
 aining\, disregarding other measures of complexity. A classical ML perform
 s a measurement and records the classical outcome after each run of $E$\, 
 while a quantum ML can access $E$ coherently to acquire quantum data\; the
  classical or quantum data is then used to predict outcomes of future expe
 riments. We prove that\, for any input distribution $D(x)$\, a classical M
 L can provide accurate predictions on average by accessing $E$ a number of
  times comparable to the optimal quantum ML. In contrast\, for achieving a
 ccurate prediction on all inputs\, we show that exponential quantum advant
 age exists in certain tasks. For example\, to predict expectation values o
 f all Pauli observables in an $n-$qubit system\, we present a quantum ML u
 sing only $O(n)$ data and prove that a classical ML requires $2^{\\Omega(n
 )}$ data.\n
LOCATION:https://researchseminars.org/talk/MPML/36/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Mário Figueiredo (Instituto Superior Técnico and IT)
DTSTART:20210217T180000Z
DTEND:20210217T190000Z
DTSTAMP:20260422T212558Z
UID:MPML/37
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/37/">De
 aling with Correlated Variables in Supervised Learning</a>\nby Mário Figu
 eiredo (Instituto Superior Técnico and IT) as part of Mathematics\, Physi
 cs and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nLinear (and generali
 zed linear) regression (LR) is an old\, but still essential\, statistical 
 tool: its goal is to learn to predict a (response) variable from a linear 
 combination of other (explanatory) variables. A central problem in LR is t
 he selection of relevant variables\, because using fewer variables tends t
 o yield better generalization and because this identification may be meani
 ngful (e.g.\, which genes are relevant to predict a certain disease). In t
 he past quarter-century\, variable selection (VS) based on sparsity-induci
 ng regularizers has been a central paradigm\, the most famous example bein
 g the LASSO\, which has been intensively studied\,\nextended\, and applied
 .\n\nIn many contexts\, it is natural to have highly-correlated variables 
 (e.g.\, several genes that are strongly co-regulated)\, thus simultaneousl
 y relevant as predictors. In this case\, sparsity-based VS may fail: it ma
 y select an arbitrary subset of these variables and it is unstable. Moreov
 er\, it is often desirable to identify all the relevant variables\, not ju
 st an arbitrary subset thereof\, a goal for which several approaches have 
 been proposed. This talk will be devoted to a recent class of such approac
 hes\, called ordered weighted l1 (OWL). The key feature of OWL is that it 
 is provably able to explicitly identify (i.e. cluster) sufficiently-correl
 ated features\, without having to compute these correlations. Several theo
 retical results characterizing OWL will be presented\, including connectio
 ns to the mathematics of economic inequality. Computational and optimizati
 on aspects will also be addressed\, as well as recent applications in subs
 pace clustering\, learning Gaussian graphical models\, and deep neural net
 works.\n
LOCATION:https://researchseminars.org/talk/MPML/37/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Maciej Koch-J8anusz (University of Zurich)
DTSTART:20210222T170000Z
DTEND:20210222T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/38
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/38/">St
 atistical physics through the lens of real-space mutual information</a>\nb
 y Maciej Koch-J8anusz (University of Zurich) as part of Mathematics\, Phys
 ics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nIdentifying the rel
 evant coarse-grained degrees of freedom in a complex physical system is a 
 key stage in developing effective theories. The renormalization group (RG)
  provides a framework for this task\, but its practical execution in unfam
 iliar systems is fraught with ad hoc choices. Machine learning approaches\
 , on the other hand\, though promising\, often lack formal interpretabilit
 y: it is unclear what relation\, if any\, the architecture- and training-d
 ependent learned "relevant" features bear to standard objects of physical 
 theory.\n\nI will present recent results addressing both issues. We develo
 p a fast algorithm\, the RSMI-NE\, employing state-of-art results in machi
 ne-learning-based estimation of information-theoretic quantities to constr
 uct the optimal coarse-graining. We use it to develop a new approach to id
 entifying the most relevant field theory operators describing a statistica
 l system\, which we validate on the example of interacting dimer model. I 
 will also discuss formal results underlying the method: we establish equiv
 alence between the information-theoretic notion of relevance defined in th
 e Information Bottleneck (IB) formalism of compression theory\, and the fi
 eld-theoretic relevance of the RG. We show analytically that for statistic
 al physical systems the "relevant" degrees of freedom found using IB compr
 ession indeed correspond to operators with the lowest scaling dimensions\,
  providing a dictionary connecting two distinct theoretical toolboxes.\n
LOCATION:https://researchseminars.org/talk/MPML/38/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Markus Heyl (Max-Planck Institute for the Physics of Complex Syste
 ms\, Dresden)
DTSTART:20210322T170000Z
DTEND:20210322T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/39
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/39/">Qu
 antum many-body dynamics in two dimensions with artificial neural networks
 </a>\nby Markus Heyl (Max-Planck Institute for the Physics of Complex Syst
 ems\, Dresden) as part of Mathematics\, Physics and Machine Learning (IST\
 , Lisbon)\n\n\nAbstract\nIn the last two decades the field of nonequilibri
 um quantum many-body physics has seen a rapid development driven\, in part
 icular\, by the remarkable progress in quantum simulators\, which today pr
 ovide access to dynamics in quantum matter with an unprecedented control. 
 However\, the efficient numerical simulation of nonequilibrium real-time e
 volution in isolated quantum matter still remains a key challenge for curr
 ent computational methods especially beyond one spatial dimension. In this
  talk I will present a versatile and efficient machine learning inspired a
 pproach. I will first introduce the general idea of encoding quantum many-
 body wave functions into artificial neural networks. I will then identify 
 and resolve key challenges for the simulation of real-time evolution\, whi
 ch previously imposed significant limitations on the accurate description 
 of large systems and long-time dynamics. As a concrete example\, I will co
 nsider the dynamics of the paradigmatic two-dimensional transverse field I
 sing model\, where we observe collapse and revival oscillations of ferroma
 gnetic order and demonstrate that the reached time scales are comparable t
 o or exceed the capabilities of state-of-the-art tensor network methods.\n
LOCATION:https://researchseminars.org/talk/MPML/39/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Pedro A. Santos (Instituto Superior Técnico and INESC-ID)
DTSTART:20210409T130000Z
DTEND:20210409T140000Z
DTSTAMP:20260422T212558Z
UID:MPML/40
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/40/">Tw
 o-time scale stochastic approximation for reinforcement learning with line
 ar function approximation</a>\nby Pedro A. Santos (Instituto Superior Téc
 nico and INESC-ID) as part of Mathematics\, Physics and Machine Learning (
 IST\, Lisbon)\n\n\nAbstract\nIn this presentation\, I will introduce some 
 traditional Reinforcement Learning problems and algorithms\, and analyze h
 ow some problems can be avoided and convergence results obtained using a t
 wo-time scale variation of the usual stochastic approximation approach.\n\
 nThis variation was inspired by the practical successes of Deep Q-Learning
  in attaining superhuman performance at some classical Atari games by Deep
 mind's research team in 2015. Machine Learning practical successes like th
 is often have no corresponding explaining theory. The work that will be pr
 esented intends to contribute to that goal.\n\nJoint work with Diogo Carva
 lho and Francisco Melo from INESC-ID.\n
LOCATION:https://researchseminars.org/talk/MPML/40/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Rebecca Willett (University of Chicago)
DTSTART:20210507T130000Z
DTEND:20210507T140000Z
DTSTAMP:20260422T212558Z
UID:MPML/41
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/41/">Ma
 chine Learning and Inverse Problems: Deeper and More Robust</a>\nby Rebecc
 a Willett (University of Chicago) as part of Mathematics\, Physics and Mac
 hine Learning (IST\, Lisbon)\n\n\nAbstract\nMany challenging image process
 ing tasks can be described by an ill-posed linear inverse problem: deblurr
 ing\, deconvolution\, inpainting\, compressed sensing\, and superresolutio
 n all lie in this framework. Recent advances in machine learning and image
  processing have illustrated that it is often possible to learn a regulari
 zer from training data that can outperform more traditional approaches by 
 large margins. In this talk\, I will describe the central prevailing theme
 s of this emerging area and a taxonomy that can be used to categorize diff
 erent problems and reconstruction methods. We will also explore mechanisms
  for model adaptation\; that is\, given a network trained to solve an init
 ial inverse problem with a known forward model\, we propose novel procedur
 es that adapt the network to a perturbed forward model\, even without full
  knowledge of the perturbation. Finally\, I will describe a new class of a
 pproaches based on "infinite-depth networks" that can yield up to a 4dB PS
 NR improvement in reconstruction accuracy above state-of-the-art alternati
 ves and where the computational budget can be selected at test time to opt
 imize context-dependent trade-offs between accuracy and computation.\n
LOCATION:https://researchseminars.org/talk/MPML/41/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Mikhail Belkin (Halicioğlu Data Science Institute\, University of
  California San Diego)
DTSTART:20210428T170000Z
DTEND:20210428T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/42
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/42/">Tw
 o mathematical lessons of deep learning</a>\nby Mikhail Belkin (Halicioğl
 u Data Science Institute\, University of California San Diego) as part of 
 Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nR
 ecent empirical successes of deep learning have exposed significant gaps i
 n our fundamental understanding of learning and optimization mechanisms. M
 odern best practices for model selection are in direct contradiction to th
 e methodologies suggested by classical analyses. Similarly\, the efficienc
 y of SGD-based local methods used in training modern models\, appeared at 
 odds with the standard intuitions on optimization.\n\nFirst\, I will prese
 nt evidence\, empirical and mathematical\, that necessitates revisiting cl
 assical notions\, such as over-fitting. I will continue to discuss the eme
 rging understanding of generalization\, and\, in particular\, the "double 
 descent" risk curve\, which extends the classical U-shaped generalization 
 curve beyond the point of interpolation.\n\nSecond\, I will discuss why th
 e landscapes of over-parameterized neural networks are generically never c
 onvex\, even locally. Instead\, as I will argue\, they satisfy the Polyak-
 Lojasiewicz condition across most of the parameter space instead\, which a
 llows SGD-type methods to converge to a global minimum.\n\nA key piece of 
 the puzzle remains - how does optimization align with statistics to form t
 he complete mathematical picture of modern ML?\n
LOCATION:https://researchseminars.org/talk/MPML/42/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Gabriel Peyré (École Normale Supérieure)
DTSTART:20210414T170000Z
DTEND:20210414T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/43
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/43/">Sc
 aling Optimal Transport for High dimensional Learning</a>\nby Gabriel Peyr
 é (École Normale Supérieure) as part of Mathematics\, Physics and Machi
 ne Learning (IST\, Lisbon)\n\n\nAbstract\nOptimal transport (OT) has recen
 tly gained lot of interest in machine learning. It is a natural tool to co
 mpare in a geometrically faithful way probability distributions. It finds 
 applications in both supervised learning (using geometric loss functions) 
 and unsupervised learning (to perform generative model fitting). OT is how
 ever plagued by the curse of dimensionality\, since it might require a num
 ber of samples which grows exponentially with the dimension. In this talk\
 , I will explain how to leverage entropic regularization methods to define
  computationally efficient loss functions\, approximating OT with a better
  sample complexity.\n\nMore information and references can be found on the
  website of our book\n"Computational Optimal Transport"\, https://optimalt
 ransport.github.io/\n
LOCATION:https://researchseminars.org/talk/MPML/43/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Kyriakos Vamvoudakis (Georgia Institute of Technology)
DTSTART:20210521T130000Z
DTEND:20210521T140000Z
DTSTAMP:20260422T212558Z
UID:MPML/44
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/44/">Le
 arning-Based Actuator Placement and Receding Horizon Control for Security 
 against Actuation Attacks</a>\nby Kyriakos Vamvoudakis (Georgia Institute 
 of Technology) as part of Mathematics\, Physics and Machine Learning (IST\
 , Lisbon)\n\n\nAbstract\nCyber-physical systems (CPS) comprise interacting
  digital\, analog\, physical\, and human components engineered for functio
 n through integrated physics and logic. Incorporating intelligence in CPS\
 , however\, makes their physical components more exposed to adversaries th
 at can potentially cause failure or malfunction through actuation attacks.
  As a result\, augmenting CPS with resilient control and design methods is
  of grave significance\, especially if an actuation attack is stealthy. To
 wards this end\, in the first part of the talk\, I will present a receding
  horizon controller\, which can deal with undetectable actuation attacks b
 y solving a game in a moving horizon fashion. In fact\, this controller ca
 n guarantee stability of the equilibrium point of the CPS\, even if the at
 tackers have an information advantage. The case where the attackers are no
 t aware of the decision-making mechanism of one another is also considered
 \, by exploiting the theory of bounded rationality. In the second part of 
 the talk\, and for CPS that have partially unknown dynamics\, I will prese
 nt an online actuator placement algorithm\, which chooses the actuators of
  the CPS that maximize an attack security metric. It can be proved that th
 e maximizing set of actuators is found in finite time\, despite the CPS ha
 ving uncertain dynamics.\n
LOCATION:https://researchseminars.org/talk/MPML/44/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Gustau Camps-Valls (Universitat de València)
DTSTART:20210528T130000Z
DTEND:20210528T140000Z
DTSTAMP:20260422T212558Z
UID:MPML/45
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/45/">Ph
 ysics Aware Machine Learning for the Earth Sciences</a>\nby Gustau Camps-V
 alls (Universitat de València) as part of Mathematics\, Physics and Machi
 ne Learning (IST\, Lisbon)\n\n\nAbstract\nMost problems in Earth sciences 
 aim to do inferences about the system\, where accurate predictions are jus
 t a tiny part of the whole problem. Inferences mean understanding variable
 s relations\, deriving models that are physically interpretable\, that are
  simple parsimonious\, and mathematically tractable. Machine learning mode
 ls alone are excellent approximators\, but very often do not respect the m
 ost elementary laws of physics\, like mass or energy conservation\, so con
 sistency and confidence are compromised. I will review the main challenges
  ahead in the field\, and introduce several ways to live in the Physics an
 d machine learning interplay that allows us (1) to encode differential equ
 ations from data\, (2) constrain data-driven models with physics-priors an
 d dependence constraints\, (3) improve parameterizations\, (4) emulate phy
 sical models\, and (5) blend data-driven and process-based models. This is
  a collective long-term AI agenda towards developing and applying algorith
 ms capable of discovering knowledge in the Earth system.\n
LOCATION:https://researchseminars.org/talk/MPML/45/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Ulugbek Kamilov (University of Washington)
DTSTART:20210611T130000Z
DTEND:20210611T140000Z
DTSTAMP:20260422T212558Z
UID:MPML/46
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/46/">Co
 mputational Imaging: Reconciling Physical and Learned Models</a>\nby Ulugb
 ek Kamilov (University of Washington) as part of Mathematics\, Physics and
  Machine Learning (IST\, Lisbon)\n\n\nAbstract\n<p class="western" style="
 text-align:justify"><span style="line-height:100%"><font face="Calibri\, s
 erif"><span style="font-size:11pt">Computational imaging is a rapidly grow
 ing area that seeks to enhance the capabilities of imaging instruments by 
 viewing imaging as an inverse problem. There are currently two distinct ap
 proaches for designing computational imaging methods: model-based and lear
 ning-based. Model-based methods leverage analytical signal properties and 
 often come with theoretical guarantees and insights. Learning-based method
 s leverage data-driven representations for best empirical performance thro
 ugh training on large datasets. This talk presents Regularization by Artif
 act Removal (RARE)\, as a framework for reconciling both viewpoints by pro
 viding a learning-based extension to the classical theory. RARE relies on 
 pre-trained “artifact-removing deep neural nets” for infusing learned 
 prior knowledge into an inverse problem\, while maintaining a clear separa
 tion between the prior and physics-based acquisition model. O</span></font
 ><font face="Calibri\, serif"><span style="font-size:11pt">ur results indi
 cate that RARE can achieve state-of-the-art performance in different compu
 tational imaging tasks\, while also being amenable to rigorous theoretical
  analysis. We will focus on the applications of RARE in biomedical imaging
 \, including magnetic resonance and tomographic imaging.</span></font></sp
 an></p>\n\n<p class="western" style="text-align:justify"><span style="line
 -height:100%"><font face="Calibri\, serif"><span style="font-size:11pt"><b
 >This talk will be based on the following references</b></span></font></sp
 an></p>\n\n<ol>\n	<li class="western"><span style="line-height:100%"><font
  face="Calibri\, serif"><span style="font-size:11pt">J. Liu\, Y. Sun\, C. 
 Eldeniz\, W. Gan\, H. An\, and U. S. Kamilov\, “<a href="https://arxiv.o
 rg/abs/1912.05854">RARE: Image Reconstruction using Deep Priors Learned wi
 thout Ground Truth\,</a>” IEEE J. Sel. Topics Signal Process.\, vol. 14\
 , no. 6\, pp. 1088-1099\, October 2020.</span></font></span></li>\n	<li cl
 ass="western" style="text-align: justify\;"><span style="line-height:100%"
 ><font face="Calibri\, serif"><span style="font-size:11pt">Z. Wu\, Y. Sun\
 , A. Matlock\, J. Liu\, L. Tian\, and U. S. Kamilov\, “<a href="https://
 arxiv.org/abs/1911.13241">SIMBA: Scalable Inversion in Optical Tomography 
 using Deep Denoising Priors</a>\,” IEEE J. Sel. Topics Signal Process.\,
  vol. 14\, no. 6\, pp. 1163-1175\, October 2020.</span></font></span></li>
 \n	<li class="western" style="text-align: justify\;"><span style="line-hei
 ght:100%"><font face="Calibri\, serif"><span style="font-size:11pt">J. Liu
 \, Y. Sun\, W. Gan\, X. Xu\, B. Wohlberg\, and U. S. Kamilov\, “<a href=
 "https://arxiv.org/abs/2101.09379">SGD-Net: Efficient Model-Based Deep Lea
 rning with Theoretical Guarantees</a>\,” IEEE Trans. Comput. Imag.\, in 
 press.</span></font></span></li>\n</ol>\n\n<p class="western" style="text-
 align:justify">&nbsp\;</p>\n
LOCATION:https://researchseminars.org/talk/MPML/46/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Ruth Misener (Imperial College London)
DTSTART:20210618T130000Z
DTEND:20210618T140000Z
DTSTAMP:20260422T212558Z
UID:MPML/47
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/47/">Pa
 rtition-based formulations for mixed-integer optimization of trained ReLU 
 neural networks</a>\nby Ruth Misener (Imperial College London) as part of 
 Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nT
 his work develops a class of relaxations in between the big-M and convex h
 ull formulations of disjunctions\, drawing advantages from both. We show t
 hat this class leads to mixed-integer formulations for trained ReLU neural
  networks. The approach balances model size and tightness by partitioning 
 node inputs into a number of groups and forming the convex hull over the p
 artitions via disjunctive programming. At one extreme\, one partition per 
 input recovers the convex hull of a node\, i.e.\, the tightest possible fo
 rmulation for each node. For fewer partitions\, we develop smaller relaxat
 ions that approximate the convex hull\, and show that they outperform exis
 ting formulations. Specifically\, we propose strategies for partitioning v
 ariables based on theoretical motivations and validate these strategies us
 ing extensive computational experiments. Furthermore\, the proposed scheme
  complements known algorithmic approaches\, e.g.\, optimization-based boun
 d tightening captures dependencies within a partition.\n\nThis joint work 
 with Calvin Tsay\, Jan Kronqvist\, Alexander Thebelt is based on two paper
 s: https://arxiv.org/abs/2102.04373  & https://arxiv.org/abs/2101.12708\n
LOCATION:https://researchseminars.org/talk/MPML/47/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Mathieu Blondel (Google Research\, Brain team\, Paris)
DTSTART:20210604T130000Z
DTEND:20210604T140000Z
DTSTAMP:20260422T212558Z
UID:MPML/48
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/48/">Ef
 ficient and Modular Implicit Differentiation</a>\nby Mathieu Blondel (Goog
 le Research\, Brain team\, Paris) as part of Mathematics\, Physics and Mac
 hine Learning (IST\, Lisbon)\n\n\nAbstract\nAutomatic differentiation (aut
 odiff) has revolutionized machine learning. It allows expressing complex c
 omputations by composing elementary ones in creative ways and removes the 
 burden of computing their derivatives by hand. More recently\, differentia
 tion of optimization problem solutions has attracted widespread attention 
 with applications such as optimization as a layer\, and in bi-level proble
 ms such as hyper-parameter optimization and meta-learning. However\, the f
 ormulas for these derivatives often involve case-by-case tedious mathemati
 cal derivations. In this work\, we propose a unified\, efficient and modul
 ar approach for implicit differentiation of optimization problems. In our 
 approach\, the user defines (in Python in the case of our implementation) 
 a function F capturing the optimality conditions of the problem to be diff
 erentiated. Once this is done\, we leverage autodiff of F and implicit dif
 ferentiation to automatically differentiate the optimization problem. Our 
 approach thus combines the benefits of implicit differentiation and autodi
 ff. It is efficient as it can be added on top of any state-of-the-art solv
 er and modular as the optimality condition specification is decoupled from
  the implicit differentiation mechanism. We show that seemingly simple pri
 nciples allow to recover many recently proposed implicit differentiation m
 ethods and create new ones easily. We demonstrate the ease of formulating 
 and solving bi-level optimization problems using our framework. We also sh
 owcase an application to the sensitivity analysis of molecular dynamics.\n
LOCATION:https://researchseminars.org/talk/MPML/48/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Yuejie Chi (Carnegie Mellon University)
DTSTART:20210625T130000Z
DTEND:20210625T140000Z
DTSTAMP:20260422T212558Z
UID:MPML/49
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/49/">Po
 licy Optimization in Reinforcement Learning: A Tale of Preconditioning and
  Regularization</a>\nby Yuejie Chi (Carnegie Mellon University) as part of
  Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\n
 Policy optimization\, which learns the policy of interest by maximizing th
 e value function via large-scale optimization techniques\, lies at the hea
 rt of modern reinforcement learning (RL). In addition to value maximizatio
 n\, other practical considerations arise commonly as well\, including the 
 need of encouraging exploration\, and that of ensuring certain structural 
 properties of the learned policy due to safety\, resource and operational 
 constraints. These considerations can often be accounted for by resorting 
 to regularized RL\, which augments the target value function with a struct
 ure-promoting regularization term\, such as Shannon entropy\, Tsallis entr
 opy\, and log-barrier functions. Focusing on an infinite-horizon discounte
 d Markov decision process\, this talk first shows that entropy-regularized
  natural policy gradient methods converge globally at a linear convergence
  that is near independent of the dimension of the state-action space. Next
 \, a generalized policy mirror descent algorithm is proposed to accommodat
 e a general class of convex regularizers beyond Shannon entropy. Encouragi
 ngly\, this general algorithm inherits similar convergence guarantees\, ev
 en when the regularizer lacks strong convexity and smoothness. Our results
  accommodate a wide range of learning rates\, and shed light upon the role
  of regularization in enabling fast convergence in RL.\n
LOCATION:https://researchseminars.org/talk/MPML/49/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Ard Louis (University of Oxford)
DTSTART:20210702T130000Z
DTEND:20210702T140000Z
DTSTAMP:20260422T212558Z
UID:MPML/50
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/50/">De
 ep neural networks have an inbuilt Occam's razor</a>\nby Ard Louis (Univer
 sity of Oxford) as part of Mathematics\, Physics and Machine Learning (IST
 \, Lisbon)\n\n\nAbstract\nOne of the most surprising properties of deep ne
 ural networks (DNNs) is that they perform best in the overparameterized re
 gime. We are taught early on that having more parameters than data points 
 is a terrible idea. So why do DNNs work so well in a regime where classica
 l learning theory predicts they should heavily overfit? By adapting the co
 ding theorem from algorithmic information theory (which every physicist sh
 ould learn about!) we show that DNNs are exponentially biased at initialis
 ation to functions that have low descriptional (Kolmogorov) complexity. In
  other words\, DNNs have an inbuilt Occam's razor\, a bias towards simple 
 functions. We next show that stochastic gradient descent (SGD)\, the most 
 popular optimisation method for DNNs\, follows the same bias\, and so does
  not itself explain the good generalisation of DNNs. Our approach naturall
 y leads to a marginal-likelihood PAC-Bayes generalisation bound which perf
 orms better than any other bounds on the market. Finally\, we discuss why 
 this bias towards simplicity allows DNNs to perform so well\, and speculat
 e on what this may tell us about the natural world.\n
LOCATION:https://researchseminars.org/talk/MPML/50/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Usman Khan (Tufts University)
DTSTART:20210709T130000Z
DTEND:20210709T140000Z
DTSTAMP:20260422T212558Z
UID:MPML/51
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/51/">Di
 stributed ML: Optimal algorithms for distributed stochastic non-convex opt
 imization</a>\nby Usman Khan (Tufts University) as part of Mathematics\, P
 hysics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nIn many emerging
  applications\, it is of paramount interest to learn hidden parameters fro
 m data. For example\, self-driving cars may use onboard cameras to identif
 y pedestrians\, highway lanes\, or traffic signs in various light and weat
 her conditions. Problems such as these can be framed as classification\, r
 egression\, or risk minimization in general\, at the heart of which lies s
 tochastic optimization and machine learning. In many practical scenarios\,
  distributed and decentralized learning methods are preferable as they ben
 efit from a divide-and-conquer approach towards data at the expense of loc
 al (short-range) communication. In this talk\, I will present our recent w
 ork that develops a novel algorithmic framework to address various aspects
  of decentralized stochastic first-order optimization methods for non-conv
 ex problems. A major focus will be to characterize regimes where decentral
 ized solutions outperform their centralized counterparts and lead to optim
 al convergence guarantees. Moreover\, I will characterize certain desirabl
 e attributes of decentralized methods in the context of linear speedup and
  networkindependent convergence rates. Throughout the talk\, I will demons
 trate such key aspects of the proposed methods with the help of provable t
 heoretical results and numerical experiments on real data.\n
LOCATION:https://researchseminars.org/talk/MPML/51/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Simon Du (University of Washington)
DTSTART:20210728T160000Z
DTEND:20210728T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/52
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/52/">Pr
 ovable Representation Learning</a>\nby Simon Du (University of Washington)
  as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\
 nAbstract\nRepresentation learning has been widely used in many applicatio
 ns. In this talk\, I will present our work\, which uncovers when and why r
 epresentation learning provably improves the sample efficiency\, from a st
 atistical learning point of view. I will show 1) the existence of a good r
 epresentation among all tasks\, and 2) the diversity of tasks are key cond
 itions that permit improved statistical efficiency via multi-task represen
 tation learning. These conditions provably improve the sample efficiency f
 or functions with certain complexity measures as the representation. If ti
 me permits\, I will also talk about leveraging the theoretical insights to
  improve practical performance.\n
LOCATION:https://researchseminars.org/talk/MPML/52/
END:VEVENT
BEGIN:VEVENT
SUMMARY:J. Nathan Kutz (University of Washington)
DTSTART:20210916T160000Z
DTEND:20210916T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/53
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/53/">De
 ep learning for the discovery of parsimonious physics models</a>\nby J. Na
 than Kutz (University of Washington) as part of Mathematics\, Physics and 
 Machine Learning (IST\, Lisbon)\n\n\nAbstract\nA major challenge in the st
 udy of dynamical systems is that of model discovery: turning data into red
 uced order models that are not just predictive\, but provide insight into 
 the nature of the underlying dynamical system that generated the data. We 
 introduce a number of data-driven strategies for discovering nonlinear mul
 tiscale dynamical systems and their embeddings from data. We consider two 
 canonical cases: (i) systems for which we have full measurements of the go
 verning variables\, and (ii) systems for which we have incomplete measurem
 ents. For systems with full state measurements\, we show that the recent s
 parse identification of nonlinear dynamical systems (SINDy) method can dis
 cover governing equations with relatively little data and introduce a samp
 ling method that allows SINDy to scale efficiently to problems with multip
 le time scales\, noise and parametric dependencies.   For systems with inc
 omplete observations\, we show that the Hankel alternative view of Koopman
  (HAVOK) method\, based on time-delay embedding coordinates and the dynami
 c mode decomposition\, can be used to obtain a linear models and Koopman i
 nvariant measurement systems that nearly perfectly captures the dynamics o
 f nonlinear quasiperiodic systems. Neural networks are used in targeted wa
 ys to aid in the model reduction process. Together\, these approaches prov
 ide a suite of mathematical strategies for reducing the data required to d
 iscover and model nonlinear multiscale systems.\n
LOCATION:https://researchseminars.org/talk/MPML/53/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Leong Chuan Kwek (Nanyang Technological University\, Singapore)
DTSTART:20210923T090000Z
DTEND:20210923T100000Z
DTSTAMP:20260422T212558Z
UID:MPML/54
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/54/">Ma
 chine Learning and Quantum Technology</a>\nby Leong Chuan Kwek (Nanyang Te
 chnological University\, Singapore) as part of Mathematics\, Physics and M
 achine Learning (IST\, Lisbon)\n\n\nAbstract\nThe rise of machine learning
  in recent times has remarkably transformed science and society. The goal 
 of machine learning is to get computers to act without being explicitly pr
 ogrammed. Machine learning with deep reinforcement learning (RL) was recen
 tly recognized as a powerful tool to engineer dynamics in quantum system. 
 Also\, recently there has been some interest to exploit and leverage the l
 imited available quantum resources for performing classically challenging 
 tasks with noisy intermediate-scale quantum (NISQ) computers. Here\, we di
 scuss some of our recent work on the applications of machine learning to q
 uantum systems.\n
LOCATION:https://researchseminars.org/talk/MPML/54/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Constantino Tsallis (Group of Statistical Physics\, CBPF and Santa
  Fe Institute)
DTSTART:20211021T160000Z
DTEND:20211021T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/55
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/55/">St
 atistical mechanics for complex systems</a>\nby Constantino Tsallis (Group
  of Statistical Physics\, CBPF and Santa Fe Institute) as part of Mathemat
 ics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nTogether 
 with Newtonian mechanics\, Maxwell electromagnetism\, Einstein relativity 
 and quantum mechanics\, Boltzmann-Gibbs (BG) statistical mechanics constit
 utes one of the pillars of contemporary theoretical physics\, with uncount
 able applications in science and technology. This theory applies formidabl
 y well to a plethora of physical systems. Still\, it fails in the realm of
  complex systems\, characterized by generically strong space-time entangle
 ment of their elements. On the basis of a nonadditive entropy (defined by 
 an index q\, which recovers\, for q=1\, the celebrated Boltzmann-Gibbs-von
  Neumann-Shannon entropy)\, it is possible to generalize the BG theory. We
  will briefly review the foundations and applications in natural\, artific
 ial and social systems.\n\nA Bibliography is available at http://tsallis.c
 at.cbpf.br/biblio.htm\n
LOCATION:https://researchseminars.org/talk/MPML/55/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Volkan Cevher (Laboratory for Information and Inference Systems 
 – LIONS\, EPFL)
DTSTART:20210930T160000Z
DTEND:20210930T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/56
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/56/">Op
 timization Challenges in Adversarial Machine Learning</a>\nby Volkan Cevhe
 r (Laboratory for Information and Inference Systems – LIONS\, EPFL) as p
 art of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbst
 ract\nThanks to neural networks (NNs)\, faster computation\, and massive d
 atasets\, machine learning (ML) is under increasing pressure to provide au
 tomated solutions to even harder real-world tasks beyond human performance
  with ever faster response times due to potentially huge technological and
  societal benefits. Unsurprisingly\, the NN learning formulations present 
 a fundamental challenge to the back-end learning algorithms despite their 
 scalability\, in particular due to the existence of traps in the non-conve
 x optimization landscape\, such as saddle points\, that can prevent algori
 thms from obtaining “good” solutions.\n\nIn this talk\, we describe ou
 r recent research that has demonstrated that the non-convex optimization d
 ogma is false by showing that scalable stochastic optimization algorithms 
 can avoid traps and rapidly obtain locally optimal solutions. Coupled with
  the progress in representation learning\, such as over-parameterized neur
 al networks\, such local solutions can be globally optimal.\n\nUnfortunate
 ly\, this talk will also demonstrate that the central min-max optimization
  problems in ML\, such as generative adversarial networks (GANs)\, robust 
 reinforcement learning (RL)\, and\ndistributionally robust ML\, contain sp
 urious attractors that do not include any stationary points of the origina
 l learning formulation. Indeed\, we will describe how algorithms are subje
 ct to a grander challenge\, including unavoidable convergence failures\, w
 hich could explain the stagnation in their progress despite the impressive
  earlier demonstrations.\n
LOCATION:https://researchseminars.org/talk/MPML/56/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Clément Hongler (EPFL)
DTSTART:20211014T160000Z
DTEND:20211014T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/57
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/57/">Ne
 ural Tangent Kernel</a>\nby Clément Hongler (EPFL) as part of Mathematics
 \, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nThe Neural T
 angent Kernel is a new way to understand the gradient descent in deep neur
 al networks\, connecting them with kernel methods. In this talk\, I'll int
 roduce this formalism and give a number of results on the Neural Tangent K
 ernel and explain how they give us insight into the dynamics of neural net
 works during training and into their generalization features.\n\nBased off
  joint works with Arthur Jacot and Franck Gabriel.\n
LOCATION:https://researchseminars.org/talk/MPML/57/
END:VEVENT
BEGIN:VEVENT
SUMMARY:George Em Karniadakis (Brown University)
DTSTART:20211104T170000Z
DTEND:20211104T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/58
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/58/">Op
 erator regression via DeepOnet: Theory\, Algorithms and Applications</a>\n
 by George Em Karniadakis (Brown University) as part of Mathematics\, Physi
 cs and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nWe will review physi
 cs-informed neural network and summarize available theoretical results. We
  will also introduce new NNs that learn functionals and nonlinear operator
 s from functions and corresponding responses for system identification. Th
 e universal approximation theorem of operators is suggestive of the potent
 ial of NNs in learning from scattered data any continuous operator or comp
 lex system. We first generalize the theorem to deep neural networks\, and 
 subsequently we apply it to design a new composite NN with small generaliz
 ation error\, the deep operator network (DeepONet)\, consisting of a NN fo
 r encoding the discrete input function space (branch net) and another NN f
 or encoding the domain of the output functions (trunk net). We demonstrate
  that DeepONet can learn various explicit operators\, e.g.\, integrals\, L
 aplace transforms and fractional Laplacians\, as well as implicit operator
 s that represent deterministic and stochastic differential equations. More
  generally\, DeepOnet can learn multiscale operators spanning across many 
 scales and trained by diverse sources of data simultaneously.\n
LOCATION:https://researchseminars.org/talk/MPML/58/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Michael Arbel (INRIA Grenoble Rhône-Alpes)
DTSTART:20211111T170000Z
DTEND:20211111T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/59
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/59/">An
 nealed Flow Transport Monte Carlo</a>\nby Michael Arbel (INRIA Grenoble Rh
 ône-Alpes) as part of Mathematics\, Physics and Machine Learning (IST\, L
 isbon)\n\n\nAbstract\nAnnealed Importance Sampling (AIS) and its Sequentia
 l Monte Carlo (SMC) extensions are state-of-the-art methods for estimating
  normalizing constants of probability distributions. We propose here a nov
 el Monte Carlo algorithm\, Annealed Flow Transport (AFT)\, that builds upo
 n AIS and SMC and combines them with normalizing flows (NF) for improved p
 erformance. This method transports a set of particles using not only impor
 tance sampling (IS)\, Markov chain Monte Carlo (MCMC) and resampling steps
  - as in SMC\, but also relies on NF which are learned sequentially to pus
 h particles towards the successive annealed targets. We provide limit theo
 rems for the resulting Monte Carlo estimates of the normalizing constant a
 nd expectations with respect to the target distribution. Additionally\, we
  show that a continuous-time scaling limit of the population version of AF
 T is given by a Feynman--Kac measure which simplifies to the law of a cont
 rolled diffusion for expressive NF. We demonstrate experimentally the bene
 fits and limitations of our methodology on a variety of applications.\n
LOCATION:https://researchseminars.org/talk/MPML/59/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Soledad Villar (Mathematical Institute for Data Science at Johns H
 opkins University)
DTSTART:20211202T170000Z
DTEND:20211202T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/60
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/60/">Eq
 uivariant machine learning structure like classical physics</a>\nby Soleda
 d Villar (Mathematical Institute for Data Science at Johns Hopkins Univers
 ity) as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\
 n\n\nAbstract\nThere has been enormous progress in the last few years in d
 esigning neural networks that respect the fundamental symmetries and coord
 inate freedoms of physical law. Some of these frameworks make use of irred
 ucible representations\, some make use of high-order tensor objects\, and 
 some apply symmetry-enforcing constraints. Different physical laws obey di
 fferent combinations of fundamental symmetries\, but a large fraction (pos
 sibly all) of classical physics is equivariant to translation\, rotation\,
  reflection (parity)\, boost (relativity)\, and permutations. Here we show
  that it is simple to parameterize universally approximating polynomial fu
 nctions that are equivariant under these symmetries\, or under the Euclide
 an\, Lorentz\, and Poincare groups\, at any dimensionality d. The key obse
 rvation is that nonlinear O(d)-equivariant (and related-group-equivariant)
  functions can be expressed in terms of a lightweight collection of scalar
 s---scalar products and scalar contractions of the scalar\, vector\, and t
 ensor inputs. These results demonstrate theoretically that gauge-invariant
  deep learning models for classical physics with good scaling for large pr
 oblems are feasible right now.\n
LOCATION:https://researchseminars.org/talk/MPML/60/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Pier Luigi Dragotti (Department of Electrical and Electronic Engin
 eering\, Imperial College\, London)
DTSTART:20211209T170000Z
DTEND:20211209T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/61
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/61/">Co
 mputational Imaging for Art investigation and for Neuroscience</a>\nby Pie
 r Luigi Dragotti (Department of Electrical and Electronic Engineering\, Im
 perial College\, London) as part of Mathematics\, Physics and Machine Lear
 ning (IST\, Lisbon)\n\n\nAbstract\nThe revolution in sensing\, with the em
 ergence of many new imagingtechniques\, offers the possibility of gaining 
 unprecedented access tothe physical world\, but this revolution can only b
 ear fruit through the skilful interplay between the physical and computati
 onal worlds. This is the domain of computational imaging which advocates t
 hat\, to develop effective imaging systems\, it will be necessary to go be
 yond the traditional decoupled imaging pipeline where device physics\, ima
 ge processing and the end-user application are considered separately. Inst
 ead\, we need to rethink imaging as an integrated sensing and inference mo
 del. In this talk we cover two research areas where computational imaging 
 is likely to have an impact.\n\nWe first focus on the heritage sector whic
 h is experiencing a digital revolution driven in part by the increasing us
 e of non-invasive\, non-destructive imaging techniques. These new imaging 
 methods provide a way to capture information about an entire painting and 
 can give us information about features at or below the surface of the pain
 ting. We focus on Macro X-Ray Fluorescence (XRF) scanning which is a techn
 ique for the mapping of chemical elements in paintings. After describing i
 n broad terms the working of this device\, a method that can process XRF s
 canning data from paintings is introduced. The method is based on connecti
 ng the problem of extracting elemental maps in XRF data to Prony's method\
 , a technique broadly used in engineering to estimate frequencies of a sum
  of sinusoids. The results presented show the ability of our method to det
 ect and separate weak signals related to hidden chemical elements in the p
 aintings. We then discuss results on the Leonardo's "The Virgin of the Roc
 ks" and show that our algorithm is able to reveal\, more clearly than ever
  before\, the hidden drawings of a previous composition that Leonardo then
  abandoned for the painting that we can now see.\n\nIn the second part of 
 the talk\, we focus on two-photon microscopy and neuroscience. To understa
 nd how networks of neurons process information\, it is essential to monito
 r their activity in living tissue. Multi-photon microscopy is unparalleled
  in its ability to image cellular activity and neural circuits\, deep in l
 iving tissue\, at single-cell resolution. However\, in order to achieve st
 ep changes in our understanding of brain function\, large-scale imaging st
 udies of neural populations are needed and this can be achieved only by de
 veloping computational tools that can enhance the quality of the data acqu
 ired and can scan 3-D volumes quickly. In this talk we introduce light-fie
 ld microscopy and present a method to localize neurons in 3-D. The method 
 is based on the use of proper sparsity priors\, novel optimization strateg
 ies and machine learning.\n\n\nThis is joint work with A. Foust\, P. Song\
 , C. Howe\, H. Verinaz\, J. Huang and Y.Su from Imperial College London\, 
 and C. Higgitt and N. Daly from The National Gallery in London\n
LOCATION:https://researchseminars.org/talk/MPML/61/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Suman Ravuri (DeepMind)
DTSTART:20211125T170000Z
DTEND:20211125T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/62
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/62/">Sk
 ilful precipitation nowcasting using deep generative models of radar</a>\n
 by Suman Ravuri (DeepMind) as part of Mathematics\, Physics and Machine Le
 arning (IST\, Lisbon)\n\n\nAbstract\nPrecipitation nowcasting\, the high-r
 esolution forecasting of precipitation up to two hours ahead\, supports th
 e real-world socioeconomic needs of many sectors reliant on weather-depend
 ent decision-making. State-of-the-art operational nowcasting methods typic
 ally advect precipitation fields with radar-based wind estimates\, and str
 uggle to capture important non-linear events such as convective initiation
 s. Recently introduced deep learning methods use radar to directly predict
  future rain rates\, free of physical constraints. While they accurately p
 redict low-intensity rainfall\, their operational utility is limited becau
 se their lack of constraints produces blurry nowcasts at longer lead times
 \, yielding poor performance on rarer medium-to-heavy rain events. Here we
  present a deep generative model for the probabilistic nowcasting of preci
 pitation from radar that addresses these challenges. Using statistical\, e
 conomic and cognitive measures\, we show that our method provides improved
  forecast quality\, forecast consistency and forecast value. Our model pro
 duces realistic and spatiotemporally consistent predictions over regions u
 p to 1\,536 km × 1\,280 km and with lead times from 5–90 min 
 ahead. Using a systematic evaluation by more than 50 expert meteorologists
 \, we show that our generative model ranked first for its accuracy and use
 fulness in 89% of cases against two competitive methods. When verified qua
 ntitatively\, these nowcasts are skillful without resorting to blurring. W
 e show that generative nowcasting can provide probabilistic predictions th
 at improve forecast value and support operational utility\, and at resolut
 ions and lead times where alternative methods struggle.\n
LOCATION:https://researchseminars.org/talk/MPML/62/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Dan Roberts (MIT\, Center for Theoretical Physics)
DTSTART:20220113T170000Z
DTEND:20220113T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/63
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/63/">Th
 e Principles of Deep Learning Theory</a>\nby Dan Roberts (MIT\, Center for
  Theoretical Physics) as part of Mathematics\, Physics and Machine Learnin
 g (IST\, Lisbon)\n\n\nAbstract\nDeep learning is an exciting approach to m
 odern artificial intelligence based on artificial neural networks. The goa
 l of this talk is to provide a blueprint — using tools from physics — 
 for theoretically analyzing deep neural networks of practical relevance. T
 his task will encompass both understanding the statistics of initialized d
 eep networks and determining the training dynamics of such an ensemble whe
 n learning from data.\n\nThis talk is based on a book\, <a href="https://a
 rxiv.org/pdf/2106.10165.pdf">"The Principles of Deep Learning Theory\,"</a
 > co-authored with Sho Yaida and based on research also in collaboration w
 ith Boris Hanin. It will be published next year by Cambridge University Pr
 ess.\n
LOCATION:https://researchseminars.org/talk/MPML/63/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Anders Hansen (Faculty of Mathematics and Department of Applied Ma
 thematics and Theoretical Physics\, University of Cambridge)
DTSTART:20220120T170000Z
DTEND:20220120T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/64
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/64/">Wh
 y things don’t work — On the extended Smale's 9th and 18th problems (t
 he limits of AI) and methodological barriers</a>\nby Anders Hansen (Facult
 y of Mathematics and Department of Applied Mathematics and Theoretical Phy
 sics\, University of Cambridge) as part of Mathematics\, Physics and Machi
 ne Learning (IST\, Lisbon)\n\n\nAbstract\nThe alchemists wanted to create 
 gold\, Hilbert wanted an algorithm to solve Diophantine equations\, resear
 chers want to make deep learning robust in AI\, MATLAB wants (but fails) t
 o detect when it provides wrong solutions to linear programs etc. Why does
  one not succeed in so many of these fundamental cases? The reason is typi
 cally methodological barriers. The history of science is full of methodolo
 gical barriers — reasons for why we never succeed in reaching certain go
 als. In many cases\, this is due to the foundations of mathematics. We wil
 l present a new program on methodological barriers and foundations of math
 ematics\, where — in this talk — we will focus on two basic problems: 
 (1) The instability problem in deep learning: Why do researchers fail to p
 roduce stable neural networks in basic classification and computer vision 
 problems that can easily be handled by humans — when one can prove that 
 there exist stable and accurate neural networks? Moreover\, AI algorithms 
 can typically not detect when they are wrong\, which becomes a serious iss
 ue when striving to create trustworthy AI. The problem is more general\, a
 s for example MATLAB's linprog routine is incapable of certifying correct 
 solutions of basic linear programs. Thus\, we’ll address the following q
 uestion: (2) Why are algorithms (in AI and computations in general) incapa
 ble of determining when they are wrong? These questions are deeply connect
 ed to the extended Smale’s 9th and 18th problems on the list of mathemat
 ical problems for the 21st century.\n
LOCATION:https://researchseminars.org/talk/MPML/64/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Joosep Pata (National Institute of Chemical Physics and Biophysics
 \, Estonia)
DTSTART:20220203T170000Z
DTEND:20220203T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/65
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/65/">Ma
 chine learning for data reconstruction at the LHC</a>\nby Joosep Pata (Nat
 ional Institute of Chemical Physics and Biophysics\, Estonia) as part of M
 athematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nPh
 ysics analyses at the CERN experiments rely on detector hits being interpr
 eted or reconstructed as particle candidates. The data reconstruction syst
 ems are built on decades of physics and detector knowledge and must operat
 e reliably on petabytes of data in diverse computing centers spread around
  the world. In the recent years\, machine learning (ML) is playing an incr
 easingly important role at the LHC experiments for reconstructing and inte
 rpreting the data\, from calibrating the detector readouts to the final in
 terpretation for complex signal processes. We will discuss the various asp
 ects of ML at the LHC experiments\, focusing on data reconstruction and pa
 rticle identification approaches using modern machine learning methods suc
 h as graph neural networks. We will bring a concrete detailed example from
  machine learned particle flow (MLPF)\, an R&D effort to develop a fully o
 ptimizable particle flow reconstruction across detector subsystems in CMS.
 \n
LOCATION:https://researchseminars.org/talk/MPML/65/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Jan Kieseler (European Organization for Nuclear Research (CERN))
DTSTART:20220303T170000Z
DTEND:20220303T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/66
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/66/">Th
 e MODE project</a>\nby Jan Kieseler (European Organization for Nuclear Res
 earch (CERN)) as part of Mathematics\, Physics and Machine Learning (IST\,
  Lisbon)\n\n\nAbstract\nThe effective design of instruments that rely on t
 he interaction of radiation with matter for their operation is a complex t
 ask. Furthermore\, the underlying physics processes are intrinsically stoc
 hastic in nature and open a vast space of possible choices for the physica
 l characteristics of the instrument. While even large scale detectors such
  as e.g. at the LHC are built using surrogates for the ultimate physics ob
 jective\, the MODE Collaboration (an acronym for Machine-learning Optimize
 d Design of Experiments) aims at developing tools also based on deep learn
 ing techniques to achieve end-to-end optimization of the design of instrum
 ents via a fully differentiable pipeline capable of exploring the Pareto-o
 ptimal frontier of the utility function for future particle collider exper
 iments and related detectors. The construction of such a differentiable mo
 del requires inclusion of information-extraction procedures\, including da
 ta collection\, detector response\, pattern recognition\, and other existi
 ng constraints such as cost. This talk will give an introduction to the go
 als of the newly founded MODE collaboration and highlight some of the alre
 ady existing ingredients.\n
LOCATION:https://researchseminars.org/talk/MPML/66/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Fernando E. Rosas (Faculty of Medicine\, Department of Brain Scien
 ces\, Imperial College)
DTSTART:20220324T170000Z
DTEND:20220324T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/67
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/67/">To
 wards a deeper understanding of high-order interdependencies in complex sy
 stems</a>\nby Fernando E. Rosas (Faculty of Medicine\, Department of Brain
  Sciences\, Imperial College) as part of Mathematics\, Physics and Machine
  Learning (IST\, Lisbon)\n\n\nAbstract\nWe live in an increasingly interco
 nnected world and\, unfortunately\, our understanding of interdependency i
 s still limited. As a matter of fact\, while bivariated relationships are 
 at the core of most of our data analysis methods\, there is still no princ
 ipled theory to account for the different types of interactions that can o
 ccur between three or more variables. This talk explores the vast and larg
 ely unexplored territory of multivariate complexity\, and discusses inform
 ation-theoretic approaches that have been introduced to fill this importan
 t knowledge gap.\n\nThe first part of the talk is devoted to synergistic p
 henomena\, which correspond to statistical regularities that affect the wh
 ole but not the parts. We explain how synergy can be effectively captured 
 by information-theoretic measures inspired in the nature of high brain fun
 ctions\, and how these measures allow us to map complex interdependencies 
 into hypergraphs. The second part of the talk focuses on a new theory of w
 hat constitutes causal emergence\, and how it can be measured from time se
 ries data. This theory enables a formal\, quantitative account of downward
  causation\, and introduces “causal decoupling” as a complementary mod
 ality of emergence. Importantly\, this not only establishes conceptual too
 ls to frame conjectures about emergence rigorously\, but also provides pra
 ctical procedures to test them on data. We illustrate the considered analy
 sis tools on different case studies\, including cellular automata\, baroqu
 e music\, flocking models\, and neuroimaging datasets.\n
LOCATION:https://researchseminars.org/talk/MPML/67/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Josef Urban (Czech Institute of of Informatics\, Robotics and Cybe
 rnetics (CIIRC))
DTSTART:20220331T160000Z
DTEND:20220331T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/68
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/68/">Ma
 chine Learning and Theorem Proving</a>\nby Josef Urban (Czech Institute of
  of Informatics\, Robotics and Cybernetics (CIIRC)) as part of Mathematics
 \, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nThe talk wil
 l describe several ways in which machine learning is combined with theorem
  proving today over large corpora of formal proof. If time permits\, I wil
 l also show some demos of the systems and mention related topics such as M
 L-guided conjecturing and autoformalization.\n
LOCATION:https://researchseminars.org/talk/MPML/68/
END:VEVENT
BEGIN:VEVENT
SUMMARY:André F. T. Martins (Instituto Superior Técnico)
DTSTART:20220224T163000Z
DTEND:20220224T173000Z
DTSTAMP:20260422T212558Z
UID:MPML/69
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/69/">Fr
 om Sparse Modeling to Sparse Communication</a>\nby André F. T. Martins (I
 nstituto Superior Técnico) as part of Mathematics\, Physics and Machine L
 earning (IST\, Lisbon)\n\n\nAbstract\nNeural networks and other machine le
 arning models compute continuous representations\, while humans communicat
 e mostly through discrete symbols. Reconciling these two forms of communic
 ation is desirable for generating human-readable interpretations or learni
 ng discrete latent variable models\, while maintaining end-to-end differen
 tiability.\n\nIn the first part of the talk\, I will describe how sparse m
 odeling techniques can be extended and adapted for facilitating sparse com
 munication in neural models. The building block is a family of sparse tran
 sformations called alpha-entmax\, a drop-in replacement for softmax\, whic
 h contains sparsemax as a particular case. Entmax transformations are diff
 erentiable and (unlike softmax) they can return sparse probability distrib
 utions\, useful to build interpretable attention mechanisms. Variants of t
 hese sparse transformations have been applied with success to machine tran
 slation\, natural language inference\, visual question answering\, and oth
 er tasks.\n\nIn the second part\, I will introduce mixed random variables\
 , which are in-between the discrete and continuous worlds. We build rigoro
 us theoretical foundations for these hybrids\, via a new “direct sum” 
 base measure defined on the face lattice of the probability simplex. From 
 this measure\, we introduce new entropy and Kullback-Leibler divergence fu
 nctions that subsume the discrete and differential cases and have interpre
 tations in terms of code optimality. Our framework suggests two strategies
  for representing and sampling mixed random variables\, an extrinsic (“s
 ample-and-project”) and an intrinsic one (based on face stratification).
 \n\nIn the third part\, I will show how sparse transformations can also be
  used to design new loss functions\, replacing the cross-entropy loss. To 
 this end\, I will introduce the family of Fenchel-Young losses\, revealing
  connections between generalized entropy regularizers and separation margi
 n. I will illustrate with applications in natural language generation\, mo
 rphology\, and machine translation.\n\nThis work was funded by the DeepSPI
 N ERC project - https://deep-spin.github.io\n
LOCATION:https://researchseminars.org/talk/MPML/69/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Dmitry Krotov (Watson AI Lab and IBM Research in Cambridge)
DTSTART:20220414T160000Z
DTEND:20220414T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/70
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/70/">Mo
 dern Hopfield Networks in AI and Neurobiology</a>\nby Dmitry Krotov (Watso
 n AI Lab and IBM Research in Cambridge) as part of Mathematics\, Physics a
 nd Machine Learning (IST\, Lisbon)\n\n\nAbstract\n<p>Modern Hopfield Netwo
 rks or Dense Associative Memories are recurrent neural networks with fixed
  point attractor states that are described by an energy function. In contr
 ast to conventional Hopfield Networks\, their modern versions have a very 
 large memory storage capacity\, which makes them appealing tools for many 
 problems in machine learning and cognitive and neuro-sciences. In this tal
 k I will introduce an intuition and a mathematical formulation of this cla
 ss of models\, and will give examples of problems in AI that can be tackle
 d using these new ideas. I will also explain how different individual mode
 ls of this class (e.g. hierarchical memories\, attention mechanism in tran
 sformers\, etc.) arise from their general mathematical formulation with th
 e Lagrangian functions.</p>\n\n<p><strong>References:</strong></p>\n\n<ol>
 \n	<li><a href="https://arxiv.org/abs/1606.01164">D.Krotov\, J.Hopfield\, 
 "Dense associative memory for pattern recognition"</a></li>\n	<li><a href=
 "https://arxiv.org/abs/2008.06996">D.Krotov\, J.Hopfield\, "Large Associat
 ive Memory Problem in Neurobiology and Machine Learning</a>"</li>\n	<li><a
  href="https://arxiv.org/abs/1702.01929">M.Demircigil\, et al.\, "On a mod
 el of associative memory with huge storage capacity"</a></li>\n	<li><a hre
 f="https://arxiv.org/abs/2008.02217">H.Ramsauer\, et al.\, "Hopfield Netwo
 rks is All You Need"</a></li>\n	<li><a href="https://arxiv.org/abs/2107.06
 446">D.Krotov\, "Hierarchical Associative Memory"</a></li>\n</ol>\n\n<p>&n
 bsp\;</p>\n
LOCATION:https://researchseminars.org/talk/MPML/70/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Emtiyaz Khan (RIKEN-AIP\, Tokyo and OIST\, Okinawa\, Japan)
DTSTART:20220428T090000Z
DTEND:20220428T100000Z
DTSTAMP:20260422T212558Z
UID:MPML/71
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/71/">Th
 e Bayesian Learning Rule for Adaptive AI</a>\nby Emtiyaz Khan (RIKEN-AIP\,
  Tokyo and OIST\, Okinawa\, Japan) as part of Mathematics\, Physics and Ma
 chine Learning (IST\, Lisbon)\n\n\nAbstract\nHumans and animals have a nat
 ural ability to autonomously learn and quickly adapt to their surroundings
 . How can we design AI systems that do the same? In this talk\, I will pre
 sent Bayesian principles to bridge such gaps between humans and AI. I will
  show that a wide variety of machine-learning algorithms are instances of 
 a single learning-rule called the Bayesian learning rule. The rule unravel
 s a dual perspective yielding new adaptive mechanisms for machine-learning
  based AI systems. My hope is to convince the audience that Bayesian princ
 iples are indispensable for an AI that learns as efficiently as we do.\n\n
 <p><strong>Reference: </strong>M.E. Khan\, H. Rue\, The Bayesian Learning 
 Rule [<a href="https://arxiv.org/abs/2107.04562" rel="noreferrer" target="
 _blank">arXiv</a>] [<a href="https://twitter.com/EmtiyazKhan/status/141449
 8922584711171?s=20" rel="noreferrer" target="_blank">Tweet</a>]</p>\n
LOCATION:https://researchseminars.org/talk/MPML/71/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Rianne van den Berg (Microsoft Research Amsterdam)
DTSTART:20220421T160000Z
DTEND:20220421T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/72
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/72/">Ge
 nerative models for discrete random variables</a>\nby Rianne van den Berg 
 (Microsoft Research Amsterdam) as part of Mathematics\, Physics and Machin
 e Learning (IST\, Lisbon)\n\n\nAbstract\nn this talk I will discuss how di
 fferent classes of generative models can be adapted to handle discrete ran
 dom variables\, and how this can be used to connect generative models to d
 ownstream tasks such as lossless compression. I will start by discussing n
 ormalizing flow models\, and the challenges that arise when converting the
 se models that are typically designed for real-valued random variables to 
 discrete random variables. Next\, I will demonstrate how denoising diffusi
 on models with discrete state spaces have a rich design space in terms of 
 the noising process\, and how this influences the performance of the learn
 ed denoising model. Finally\, I will show how denoising diffusion models c
 an be connected to autoregressive models\, and introduce an autoregressive
  model with a random generation order.\n
LOCATION:https://researchseminars.org/talk/MPML/72/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Andrea L. Bertozzi (University of California Los Angeles)
DTSTART:20220505T160000Z
DTEND:20220505T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/73
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/73/">Gr
 aph based models in semi-supervised and unsupervised learning</a>\nby Andr
 ea L. Bertozzi (University of California Los Angeles) as part of Mathemati
 cs\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nSimilarity
  graphs provide a structure for analyzing high dimensional data. 
 These undirected weighted graphs provide structure for identifying inheren
 t clusters in datasets and many methods exist to sort through such data bu
 ilding on the graph laplacian matrix.  One way to think about such proble
 ms is in terms of penalized cut problems.  These can be expressed in term
 s of the graph total variation which has a well-known analogue in Euclidea
 n space.  We show how to use ideas from geometric methods for PDEs to dev
 elop efficient and high performing methods for semi-supervised and unsuper
 vised learning.  These methods also extend to active learning and to modu
 larity optimization for community detection on networks.\n
LOCATION:https://researchseminars.org/talk/MPML/73/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Stanley Osher (Department of Mathematics\, University of Californi
 a\, Los Angeles)
DTSTART:20220519T160000Z
DTEND:20220519T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/74
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/74/">Co
 nservation laws and generalized optimal transport</a>\nby Stanley Osher (D
 epartment of Mathematics\, University of California\, Los Angeles) as part
  of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstrac
 t\nIn this talk\, we connect Lax’s entropy-entropy flux in conservation 
 laws with optimal transport type metric spaces. Following this connection\
 , we further design variational discretizations for conservation laws and 
 mean field control of conservation laws. In particular\, we design uncondi
 tionally stable time discretization methods that are easy to implement.\n\
 nOn joint work with Siting Liu\, UCLA and Wuchen Li\, University of South 
 Carolina.\n
LOCATION:https://researchseminars.org/talk/MPML/74/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Anja Butter (ITP\, University of Heidelberg)
DTSTART:20220602T160000Z
DTEND:20220602T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/75
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/75/">Ma
 chine Learning and LHC Event Generation</a>\nby Anja Butter (ITP\, Univers
 ity of Heidelberg) as part of Mathematics\, Physics and Machine Learning (
 IST\, Lisbon)\n\n\nAbstract\nFirst-principle simulations are at the heart 
 of the high-energy physics research program. They link the vast data outpu
 t of multi-purpose detectors with fundamental theory predictions and inter
 pretation. In the coming LHC runs\, these simulations will face unpreceden
 ted precision requirements to match the experimental accuracy. New ideas a
 nd tools based on neural networks have been developed at the interface of 
 particle physics and machine learning. They can improve the speed and prec
 ision of forward simulations and handle the complexity of collision data. 
 Such networks can be employed within established simulation tools or as pa
 rt of a new framework. Since neural networks can be inverted\, they open n
 ew avenues in LHC analyses.\n
LOCATION:https://researchseminars.org/talk/MPML/75/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Paulo Tabuada (UCLA)
DTSTART:20220609T160000Z
DTEND:20220609T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/77
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/77/">De
 ep neural networks\, universal approximation\, and geometric control</a>\n
 by Paulo Tabuada (UCLA) as part of Mathematics\, Physics and Machine Learn
 ing (IST\, Lisbon)\n\n\nAbstract\nDeep neural networks have drastically ch
 anged the landscape of several engineering areas such as computer vision a
 nd natural language processing. Notwithstanding the widespread success of 
 deep networks in these\, and many other areas\, it is still not well under
 stood why deep neural networks work so well. In particular\, the question 
 of which functions can be learned by deep neural networks has remained una
 nswered.\nIn this talk we give an answer to this question for deep residua
 l neural networks\, a class of deep networks that can be interpreted as th
 e time discretization of nonlinear control systems. We will show that the 
 ability of these networks to memorize training data can be expressed throu
 gh the control theoretic notion of controllability which can be proved usi
 ng geometric control techniques. We then add an additional ingredient\, mo
 notonicity\, to conclude that deep residual networks can approximate\, to 
 arbitrary accuracy with respect to the uniform norm\, any continuous funct
 ion on a compact subset of n-dimensional Euclidean space by using at most 
 n+1 neurons per layer. We will conclude the talk by showing how these resu
 lts pave the way for the use of deep networks in the perception pipeline o
 f autonomous systems while providing formal (and probability free) guarant
 ees of stability and robustness.\n
LOCATION:https://researchseminars.org/talk/MPML/77/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Petar Veličković (DeepMind and University of Cambridge)
DTSTART:20220929T160000Z
DTEND:20220929T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/78
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/78/">Ge
 ometric Deep Learning: Grids\, Graphs\, Groups\, Geodesics and Gauges</a>\
 nby Petar Veličković (DeepMind and University of Cambridge) as part of M
 athematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nTh
 e last decade has witnessed an experimental revolution in data science and
  machine learning\, epitomised by deep learning methods. Indeed\, many hig
 h-dimensional learning tasks previously thought to be beyond reach –such
  as computer vision\, playing Go\, or protein folding – are in fact feas
 ible with appropriate computational scale. Remarkably\, the essence of dee
 p learning is built from two simple algorithmic principles: first\, the no
 tion of representation or feature learning\, whereby adapted\, often hiera
 rchical\, features capture the appropriate notion of regularity for each t
 ask\, and second\, learning by local gradient-descent type methods\, typic
 ally implemented as backpropagation.\n\nWhile learning generic functions i
 n high dimensions is a cursed estimation problem\, most tasks of interest 
 are not generic\, and come with essential pre-defined regularities arising
  from the underlying low-dimensionality and structure of the physical worl
 d. This talk is concerned with exposing these regularities through unified
  geometric principles that can be applied throughout a wide spectrum of ap
 plications.\n\nSuch a 'geometric unification' endeavour in the spirit of F
 elix Klein's Erlangen Program serves a dual purpose: on one hand\, it prov
 ides a common mathematical framework to study the most successful neural n
 etwork architectures\, such as CNNs\, RNNs\, GNNs\, and Transformers. On t
 he other hand\, it gives a constructive procedure to incorporate prior phy
 sical knowledge into neural architectures and provide principled way to bu
 ild future architectures yet to be invented.\n
LOCATION:https://researchseminars.org/talk/MPML/78/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Yongji Wang (Department of Geosciences\, Princeton University)
DTSTART:20220526T160000Z
DTEND:20220526T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/79
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/79/">Ph
 ysics-informed neural networks for solving 3-D Euler equation</a>\nby Yong
 ji Wang (Department of Geosciences\, Princeton University) as part of Math
 ematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nOne o
 f the most challenging open questions in mathematical fluid dynamics is wh
 ether an inviscid incompressible fluid\, described by the 3-dimensional Eu
 ler equations\, with initially smooth velocity and finite energy can devel
 op singularities in finite time. This long-standing open problem is closel
 y related to one of the seven Millennium Prize Problems which considers th
 e problem the viscous analogue to the Euler equations (the Navier-Stokes e
 quations). In this talk\, I will describe how we leverage the power of dee
 p learning\, using deep neural networks with equation constraints\, namely
  physics-informed neural networks (PINNs)\, to find a smooth self-similar 
 blow-up solution for the 3-dimensional Euler equations in the presence of 
 a cylindrical boundary. To the best of our knowledge\, the solution repres
 ents the first example of a truly 2-D or higher dimensional backwards self
 -similar solution. This new numerical framework based on PINNs is shown to
  be robust and readily adaptable to other fluid equations\, which sheds ne
 w light to the century-old mystery of capital importance in the field of m
 athematical fluid dynamics.\n
LOCATION:https://researchseminars.org/talk/MPML/79/
END:VEVENT
BEGIN:VEVENT
SUMMARY:John Baez (U.C. Riverside)
DTSTART:20220616T170000Z
DTEND:20220616T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/80
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/80/">Sh
 annon Entropy from Category Theory</a>\nby John Baez (U.C. Riverside) as p
 art of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbst
 ract\nShannon entropy is a powerful concept. But what properties single ou
 t Shannon entropy as special? Instead of focusing on the entropy of a prob
 ability measure on a finite set\, it can help to focus on the "information
  loss"\, or change in entropy\, associated with a measure-preserving funct
 ion. Shannon entropy then gives the only concept of information loss that 
 is functorial\, convex-linear and continuous.\n\nThis is joint work with T
 om Leinster and Tobias Fritz.\n
LOCATION:https://researchseminars.org/talk/MPML/80/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Dario Izzo (European Space Agency)
DTSTART:20220630T160000Z
DTEND:20220630T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/81
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/81/">Ge
 odesy of irregular small bodies via neural density fields: geodesyNets</a>
 \nby Dario Izzo (European Space Agency) as part of Mathematics\, Physics a
 nd Machine Learning (IST\, Lisbon)\n\n\nAbstract\nThe problem of determini
 ng the density distribution of celestial bodies from the induced gravitati
 onal pull is of great importance in astrophysics as well as space engineer
 ing (thinking of situations where spacecraft need to perform orbital and s
 urface proximity operations). Knowledge of a body density distribution pro
 vides also great insights on the body's origin and composition. In practic
 e\, the state-of-the-art approaches for modelling the gravity field of ext
 ended bodies are spherical harmonics models\, mascon models and polyhedral
  gravity models. All of these\, however\, while being widely studied and d
 eveloped since the early works from Laplace\, introduce requirements such 
 as knowledge of a shape model\, assumption of a homogeneous internal densi
 ty\, being outside the\nBrillouin sphere\, etc...\n\n\nIn this talk\, we i
 ntroduce and explain Neural Density Fields\, a new approach to represent t
 he density of extended bodies and learn its accurate form inverting data f
 rom gravitational accelerations\, orbits or the gravity potential. The res
 ulting deep learning model\, called  geodesyNets is able to compete with c
 lassical approaches while solving most of their limitations. We also intro
 duce eclipseNets\, a deep learning model based on related ideas and able t
 o learn the eclipse shadow cones of irregular bodies\, thus allowing highl
 y precise propagation and stability studies.\n
LOCATION:https://researchseminars.org/talk/MPML/81/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Audrey Durand (IID\, Université Laval\, Canada)
DTSTART:20220707T160000Z
DTEND:20220707T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/82
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/82/">In
 teractive learning for Neurosciences - Between Simulation and Reality</a>\
 nby Audrey Durand (IID\, Université Laval\, Canada) as part of Mathematic
 s\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nLearning a 
 behaviour to conduct a given task can be achieved by interacting with the 
 the environment. This is the crux of reinforcement learning (RL)\, where a
 n (automated) agent learns to solve a problem through an iterative trial-a
 nd-error process. More specifically\, an RL agent can interact with the en
 vironment and learn from these interactions by observing a feedback on the
  goal task. Therefore\, these methods typically require to be able to inte
 rvene on the environment and make (possibly a very large number of) mistak
 es. Although this can be a limiting factor in some applications\, simple R
 L settings\, such as bandit settings\, can still host a variety of problem
 s for interactively learning behaviours. In other situations\, simulation 
 might be the key.\n\nIn this talk\, we will show that RL can be used to fo
 rmulate and tackle data acquisition (imaging) problems in neurosciences. W
 e will see how bandit methods can be used to optimize super-resolution ima
 ging by learning on real devices through an actual empirical process. We w
 ill also see how simulation can be leveraged to learn more sequential deci
 sion making strategies. These applications highlight the potential of RL t
 o support expert users on difficult task and enable new discoveries.\n
LOCATION:https://researchseminars.org/talk/MPML/82/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Joseph Bakarji (University of Washington)
DTSTART:20220714T160000Z
DTEND:20220714T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/83
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/83/">Di
 mensionally Consistent Learning with Buckingham Pi</a>\nby Joseph Bakarji 
 (University of Washington) as part of Mathematics\, Physics and Machine Le
 arning (IST\, Lisbon)\n\n\nAbstract\nDimensional analysis is a robust tech
 nique for extracting insights and finding symmetries in physical systems\,
  especially when the governing equations are not known. The Buckingham Pi 
 theorem provides a procedure for finding a set of dimensionless groups fro
 m given measurements\, although this set is not unique. We propose an auto
 mated approach using the symmetric and self-similar structure of available
  measurement data to discover the dimensionless groups that best collapse 
 this data to a lower dimensional space according to an optimal fit. We dev
 elop three data-driven techniques that use the Buckingham Pi theorem as a 
 constraint: (i) a constrained optimization problem with a nonparametric fu
 nction\, (ii) a deep learning algorithm (BuckiNet) that projects the input
  parameter space to a lower dimension in the first layer\, and (iii) a spa
 rse identification of nonlinear dynamics (SINDy) to discover dimensionless
  equations whose coefficients parameterize the dynamics. I discuss the acc
 uracy and robustness of these methods when applied to known nonlinear syst
 ems.\n
LOCATION:https://researchseminars.org/talk/MPML/83/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Inês Hipólito (Humboldt-Universität)
DTSTART:20220908T160000Z
DTEND:20220908T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/84
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/84/">Th
 e Free Energy Principle in the Edge of Chaos</a>\nby Inês Hipólito (Humb
 oldt-Universität) as part of Mathematics\, Physics and Machine Learning (
 IST\, Lisbon)\n\n\nAbstract\nLiving beings do an extraordinary thing. By b
 eing alive they are resisting the second law of thermodynamics. This law s
 tipulates that open\, living systems tend to dissipation by the increase o
 f entropy or chaos. From minimal cognitive organisms like plants to more c
 omplex organisms equipped with nervous systems\, all living systems adjust
  and adapt to their environments\, thereby resisting the second law. Impre
 ssively\, while all animals cognitively enact and survive their local envi
 ronments\, more complex systems do so also by actively constructing their 
 local environments\, thereby not only defying the second law\, but also (e
 volution) selective properties. Because all living beings defy the second 
 law by adjusting and engaging with the environment\, a prominent question 
 is how do living organisms persist while engaging in adaptive exchanges wi
 th their complex environments? In this talk I will offer an overview of ho
 w the Free Energy Principle (FEP) offers a principled solution to this pro
 blem. The FEP prescribes that living systems maintain themselves by remain
 ing in non-equilibrium steady states by restricting themselves to a limite
 d number of states\; it has been widely applied to explain neurocognitive 
 function and embodied action\, develop artificial intelligence and inspire
  psychopathology models.\n
LOCATION:https://researchseminars.org/talk/MPML/84/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Robert Nowak (University of Wisconsin-Madison)
DTSTART:20221027T160000Z
DTEND:20221027T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/85
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/85/">Th
 e Neural Balance Theorem and its Consequences</a>\nby Robert Nowak (Univer
 sity of Wisconsin-Madison) as part of Mathematics\, Physics and Machine Le
 arning (IST\, Lisbon)\n\nAbstract: TBA\n
LOCATION:https://researchseminars.org/talk/MPML/85/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Frederico Fiuza (SLAC)
DTSTART:20221103T170000Z
DTEND:20221103T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/86
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/86/">Ac
 celerating the understanding of nonlinear dynamical systems using machine 
 learning</a>\nby Frederico Fiuza (SLAC) as part of Mathematics\, Physics a
 nd Machine Learning (IST\, Lisbon)\n\n\nAbstract\nThe description of nonli
 near\, multi-scale dynamics is a common challenge in a wide range of physi
 cal systems and research fields — from weather forecast to controlled nu
 clear fusion. The development of reduced models that balance between accur
 acy and complexity is critical to advancing theoretical comprehension and 
 enabling holistic computational descriptions of these problems. I will dis
 cuss how techniques from statistical and machine learning are offering new
  ways of inferring reduced physics models from the increasingly abundant d
 ata of nonlinear dynamics produced by experiments\, observations\, and sim
 ulations. In particular\, I will focus on how sparse regression techniques
  can be used to infer interpretable plasma physics models (in the form of 
 nonlinear partial differential equations) directly from the data of first-
 principles fully-kinetic simulations. The potential of this approach is de
 monstrated by recovering the fundamental hierarchy of plasma physics model
 s based solely on particle-based simulation data of complex plasma dynamic
 s. I will discuss how this data-driven methodology provides a promising to
 ol to accelerate the development of reduced theoretical models of nonlinea
 r dynamical systems and to design computationally efficient algorithms for
  multi-scale simulations.\n
LOCATION:https://researchseminars.org/talk/MPML/86/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Markus Reichstein (MPI for Biogeochemistry)
DTSTART:20221124T170000Z
DTEND:20221124T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/87
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/87/">In
 tegrating Machine Learning with System Modelling and Observations for a be
 tter understanding of the Earth System</a>\nby Markus Reichstein (MPI for 
 Biogeochemistry) as part of Mathematics\, Physics and Machine Learning (IS
 T\, Lisbon)\n\n\nAbstract\nThe Earth is a complex dynamic networked system
 . Machine learning\, i.e. derivation of computational models from data\, h
 as already made important contributions to predict and understand componen
 ts of the Earth system\, specifically in climate\, remote sensing and envi
 ronmental sciences. For instance\, classifications of land cover types\, p
 rediction of land-atmosphere and ocean-atmosphere exchange\, or detection 
 of extreme events have greatly benefited from these approaches. Such data-
 driven information has already changed how Earth system models are evaluat
 ed and further developed. However\, many studies have not yet sufficiently
  addressed and exploited dynamic aspects of systems\, such as memory effec
 ts for prediction and effects of spatial context\, e.g. for classification
  and change detection. In particular new developments in deep learning off
 er great potential to overcome these limitations. Yet\, a key challenge an
 d opportunity is to integrate (physical-biological) system modeling approa
 ches with machine learning into hybrid modeling approaches\, which combine
 s physical consistency and machine learning versatility. A couple of examp
 les are given with focus on the terrestrial biosphere\, where the combinat
 ion of system-based and machine-learning-based modelling helps our underst
 anding of aspects of the Earth system.\n
LOCATION:https://researchseminars.org/talk/MPML/87/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Bruno Loureiro (École Polytechnique Fédérale de Lausanne (EPFL)
 )
DTSTART:20221215T170000Z
DTEND:20221215T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/88
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/88/">Ph
 ase diagram of Stochastic Gradient Descent in high-dimensional two-layer n
 eural networks</a>\nby Bruno Loureiro (École Polytechnique Fédérale de 
 Lausanne (EPFL)) as part of Mathematics\, Physics and Machine Learning (IS
 T\, Lisbon)\n\n\nAbstract\nDespite the non-convex optimization landscape\,
  over-parametrized shallow networks are able to achieve global convergence
  under gradient descent. The picture can be radically different for narrow
  networks\, which tend to get stuck in badly-generalizing local minima. He
 re we investigate the cross-over between these two regimes in the high-dim
 ensional setting\, and in particular investigate the connection between th
 e so-called mean-field/hydrodynamic regime and the seminal approach of Saa
 d & Solla. Focusing on the case of Gaussian data\, we study the interplay 
 between the learning rate\, the time scale\, and the number of hidden unit
 s in the high-dimensional dynamics of stochastic gradient descent (SGD). O
 ur work builds on a deterministic description of SGD in high-dimensions fr
 om statistical physics\, which we extend and for which we provide rigorous
  convergence rates.\n
LOCATION:https://researchseminars.org/talk/MPML/88/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Diogo Gomes (KAUST)
DTSTART:20221014T083000Z
DTEND:20221014T110000Z
DTSTAMP:20260422T212558Z
UID:MPML/89
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/89/">Fr
 om Calculus of Variations to Reinforcement Learning (Lectures 1 & 2)</a>\n
 by Diogo Gomes (KAUST) as part of Mathematics\, Physics and Machine Learni
 ng (IST\, Lisbon)\n\n\nAbstract\nThis course begins with a brief introduct
 ion to classical calculus of variations and its applications to classical 
 problems such as geodesic trajectories and the brachistochrone problem. Th
 en\, we examine Hamilton-Jacobi equations\, the role of convexity and the 
 classical verification theorem. Next\, we illustrate the lack of classical
  solutions and motivate the definition of viscosity solutions. The course 
 ends with a brief description of the reinforcement learning problem and it
 s connection with Hamilton-Jacobi equations.\n
LOCATION:https://researchseminars.org/talk/MPML/89/
END:VEVENT
BEGIN:VEVENT
SUMMARY:José Miguel Urbano (KAUST)
DTSTART:20221014T133000Z
DTEND:20221014T160000Z
DTSTAMP:20260422T212558Z
UID:MPML/90
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/90/">Se
 mi-Supervised Learning and the infinite-Laplacian (Lectures 1 & 2)</a>\nby
  José Miguel Urbano (KAUST) as part of Mathematics\, Physics and Machine 
 Learning (IST\, Lisbon)\n\n\nAbstract\nMotivated by a recent application i
 n Semi-Supervised Learning (SSL)\, the minicourse is a brief introduction 
 to the analysis of infinity-harmonic functions. We will discuss the Lipsch
 itz extension problem\, its solution via MacShane-Whitney extensions and i
 ts several drawbacks\, leading to the notion of AMLE (Absolutely Minimisin
 g Lipschitz Extension). We then explore the equivalence between being abso
 lutely minimising Lipschitz\, enjoying comparison with cones and solving t
 he infinity-Laplace equation in the viscosity sense.\n
LOCATION:https://researchseminars.org/talk/MPML/90/
END:VEVENT
BEGIN:VEVENT
SUMMARY:João Sacramento (ETH Zürich)
DTSTART:20221110T170000Z
DTEND:20221110T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/91
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/91/">Th
 e least-control principle for learning at equilibrium</a>\nby João Sacram
 ento (ETH Zürich) as part of Mathematics\, Physics and Machine Learning (
 IST\, Lisbon)\n\n\nAbstract\nA large number of models of interest in both 
 neuroscience and machine learning can be expressed as dynamical systems at
  equilibrium. This class of systems includes deep neural networks\, equili
 brium recurrent neural networks\, and meta-learning. In this talk I will p
 resent a new principle for learning equilibria with a temporally - and spa
 tially - local rule. Our principle casts learning as a least-control probl
 em\, where we first introduce an optimal controller to lead the system tow
 ards a solution state\, and then define learning as reducing the amount of
  control needed to reach such a state. We show that incorporating learning
  signals within a dynamics as an optimal control enables transmitting acti
 vity-dependent credit assignment information\, avoids storing intermediate
  states in memory\, and does not rely on infinitesimal learning signals. I
 n practice\, our principle leads to strong performance matching that of le
 ading gradient-based learning methods when applied to an array of benchmar
 king experiments. Our results shed light on how the brain might learn and 
 offer new ways of approaching a broad class of machine learning problems.\
 n
LOCATION:https://researchseminars.org/talk/MPML/91/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Tom Goldstein (University of Maryland)
DTSTART:20221117T170000Z
DTEND:20221117T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/92
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/92/">Bu
 ilding (and breaking) neural networks that think fast and slow</a>\nby Tom
  Goldstein (University of Maryland) as part of Mathematics\, Physics and M
 achine Learning (IST\, Lisbon)\n\n\nAbstract\nMost neural networks are bui
 lt to solve simple patternmatching tasks\, a process that is often known a
 s “fast” thinking. In this talk\, I’ll use adversarial methods to ex
 plore the robustness of neural networks. I’ll also discuss whether vulne
 rabilities of AI systems that have been observed in academic labs can pose
  real security threats to industrial systems. Then\, I’ll present method
 s for constructing neural networks that exhibit “slow” thinking abilit
 ies akin to human logical reasoning. Rather than learning simple pattern m
 atching rules\, these networks have the ability to synthesize algorithmic 
 reasoning processes and solve difficult discrete search and planning probl
 ems that cannot be solved by conventional AI systems. Interestingly\, thes
 e reasoning systems naturally exhibit error correction and robustness prop
 erties that make them more difficult to break than their fast thinking cou
 nterparts.\n
LOCATION:https://researchseminars.org/talk/MPML/92/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Yang-Hui He (London Institute for Mathematical Sciences & Merton C
 ollege\, Oxford University)
DTSTART:20230202T170000Z
DTEND:20230202T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/93
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/93/">CO
 LLOQUIUM: Universes as Bigdata: Physics\, Geometry and Machine-Learning</a
 >\nby Yang-Hui He (London Institute for Mathematical Sciences & Merton Col
 lege\, Oxford University) as part of Mathematics\, Physics and Machine Lea
 rning (IST\, Lisbon)\n\n\nAbstract\nThe search for the Theory of Everythin
 g has led to superstring theory\, which then led physics\, first to algebr
 aic/differential geometry/topology\, and then to computational geometry\, 
 and now to data science. With a concrete playground of the geometric lands
 cape\, accumulated by the collaboration of physicists\, mathematicians and
  computer scientists over the last 4 decades\, we show how the latest tech
 niques in machine-learning can help explore problems of interest to theore
 tical physics and to pure mathematics. At the core of our programme is the
  question: how can AI help us with mathematics?\n
LOCATION:https://researchseminars.org/talk/MPML/93/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Sebastian Engelke (University of Geneva)
DTSTART:20230112T170000Z
DTEND:20230112T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/94
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/94/">Ma
 chine learning beyond the data range: extreme quantile regression</a>\nby 
 Sebastian Engelke (University of Geneva) as part of Mathematics\, Physics 
 and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nMachine learning method
 s perform well in prediction tasks within the range of the training data. 
 When interest is in quantiles of the response that go beyond the observed 
 records\, these methods typically break down. Extreme value theory provide
 s the mathematical foundation for estimation of such extreme quantiles. A 
 common approach is to approximate the exceedances over a high threshold by
  the generalized Pareto distribution. For conditional extreme quantiles\, 
 one may model the parameters of this distribution as functions of the pred
 ictors. Up to now\, the existing methods are either not flexible enough or
  do not generalize well in higher dimensions. We develop new approaches fo
 r extreme quantile regression that estimate the parameters of the generali
 zed Pareto distribution with tree-based methods and recurrent neural netwo
 rks. Our estimators outperform classical machine learning methods and meth
 ods from extreme value theory in simulations studies. We illustrate how th
 e recurrent neural network model can be used for effective forecasting of 
 flood risk.\n
LOCATION:https://researchseminars.org/talk/MPML/94/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Alhussein Fawzi (DeepMind)
DTSTART:20230119T170000Z
DTEND:20230119T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/95
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/95/">Di
 scovering faster matrix multiplication algorithms with deep reinforcement 
 learning</a>\nby Alhussein Fawzi (DeepMind) as part of Mathematics\, Physi
 cs and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nImproving the effici
 ency of algorithms for fundamental computational tasks such as matrix mult
 iplication can have widespread impact\, as it affects the overall speed of
  a large amount of computations. The automatic discovery of algorithms usi
 ng machine learning offers the prospect of reaching beyond human intuition
  and outperforming the current best human-designed algorithms. In this tal
 k I'll present AlphaTensor\, our reinforcement learning agent based on Alp
 haZero for discovering efficient and provably correct algorithms for the m
 ultiplication of arbitrary matrices. AlphaTensor discovered algorithms tha
 t outperform the state-of-the-art complexity for many matrix sizes. Partic
 ularly relevant is the case of 4 × 4 matrices in a finite field\, where A
 lphaTensor's algorithm improves on Strassen's two-level algorithm for the 
 first time since its discovery 50 years ago. I'll present our problem form
 ulation as a single-player game\, the key ingredients that enable tackling
  such difficult mathematical problems using reinforcement learning\, and t
 he flexibility of the AlphaTensor framework.\n
LOCATION:https://researchseminars.org/talk/MPML/95/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Sara A. Solla (Northwestern University)
DTSTART:20230302T170000Z
DTEND:20230302T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/96
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/96/">Lo
 w Dimensional Manifolds for Neural Dynamics</a>\nby Sara A. Solla (Northwe
 stern University) as part of Mathematics\, Physics and Machine Learning (I
 ST\, Lisbon)\n\n\nAbstract\nThe ability to simultaneously record the activ
 ity from tens to hundreds to thousands of neurons has allowed us to analyz
 e the computational role of population activity as opposed to single neuro
 n activity. Recent work on a variety of cortical areas suggests that neura
 l function may be built on the activation of population-wide activity patt
 erns\, the neural modes\, rather than on the independent modulation of ind
 ividual neural activity. These neural modes\, the dominant covariation pat
 terns within the neural population\, define a low dimensional neural manif
 old that captures most of the variance in the recorded neural activity. We
  refer to the time-dependent activation of the neural modes as their laten
 t dynamics and argue that latent cortical dynamics within the manifold are
  the fundamental and stable building blocks of neural population activity.
 \n
LOCATION:https://researchseminars.org/talk/MPML/96/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Andreas Döpp (Ludwig-Maximilians-Universität München)
DTSTART:20230601T160000Z
DTEND:20230601T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/97
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/97/">Ma
 chine-learning strategies in laser-plasma physics</a>\nby Andreas Döpp (L
 udwig-Maximilians-Universität München) as part of Mathematics\, Physics 
 and Machine Learning (IST\, Lisbon)\n\n\nAbstract\n<p>The field of laser-p
 lasma physics has experienced significant advancements in the past few dec
 ades\, owing to the increasing power and accessibility of high-power laser
 s. Initially\, research in this area was limited to single-shot experiment
 s with minimal exploration of parameters. However\, recent technological a
 dvancements have enabled the collection of a wealth of data through both e
 xperimental and simulation-based approaches.</p>\n\n<p>In this seminar tal
 k\, I will present a range of machine learning techniques that we have dev
 eloped for applications in laser-plasma physics [1]. The first part of my 
 talk will focus on Bayesian optimization\, where I will showcase our lates
 t findings on multi-objective and multi-fidelity optimization of laser-pla
 sma accelerators and neural networks [2-4].</p>\n\n<p>In the second part o
 f the talk\, I will discuss machine learning solutions for tackling comple
 x inverse problems\, such as image deblurring or extracting 3D information
  from 2D sensors [5-6]. Specifically\, I will discuss various adaptations 
 of established convolutional network architectures\, such as the U-Net\, a
 s well as novel physics-informed retrieval methods like deep algorithm unr
 olling. These techniques have shown promising results in overcoming the ch
 allenges posed by these intricate inverse problems.</p>\n\n<p><strong>Refe
 rences:</strong></p>\n\n<p>[1] Data-driven Science and Machine Learning Me
 thods in Laser-Plasma Physics<br />\n<a href="https://arxiv.org/abs/2212.0
 0026">https://arxiv.org/abs/2212.00026</a></p>\n\n<p>[2] Expected hypervol
 ume improvement for simultaneous multi-objective and multi-fidelity optimi
 zation<br />\n<a href="https://arxiv.org/abs/2112.13901">https://arxiv.org
 /abs/2112.13901</a></p>\n\n<p>[3] Multi-objective and multi-fidelity Bayes
 ian optimization of laser-plasma acceleration<br />\n<a href="https://arxi
 v.org/abs/2210.03484">https://arxiv.org/abs/2210.03484</a></p>\n\n<p>[4] P
 areto Optimization of a Laser Wakefield Accelerator<br />\n<a href="https:
 //arxiv.org/abs/2303.15825">https://arxiv.org/abs/2303.15825</a></p>\n\n<p
 >[5] Measuring spatio-temporal couplings using modal spatio-spectral wavef
 ront retrieval<br />\n<a href="https://arxiv.org/abs/2303.01360">https://a
 rxiv.org/abs/2303.01360</a></p>\n\n<p>[6] Hyperspectral Compressive Wavefr
 ont Sensing<br />\n<a href="https://arxiv.org/abs/2303.03555">https://arxi
 v.org/abs/2303.03555</a></p>\n\n<p>&nbsp\;</p>\n
LOCATION:https://researchseminars.org/talk/MPML/97/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Ben Edelman (Harvard University)
DTSTART:20230209T170000Z
DTEND:20230209T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/98
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/98/">St
 udies in feature learning through the lens of sparse boolean functions</a>
 \nby Ben Edelman (Harvard University) as part of Mathematics\, Physics and
  Machine Learning (IST\, Lisbon)\n\n\nAbstract\nHow do deep neural network
 s learn to construct useful features? Why do self-attention-based networks
  such as transformers perform so well on combinatorial tasks such as langu
 age learning? Why do some capabilities of networks emerge "discontinuously
 " as the computational resources used for training are scaled up? We will 
 present perspectives on these questions through the lens of a particular c
 lass of simple synthetic tasks: learning sparse boolean functions. In part
  one\, we will show that the hypothesis class of one-layer transformers ca
 n learn these functions in a statistically efficient manner. This leads to
  a view of each layer of a transformer as creating new "variables" out of 
 sparse combinations of the previous layer's outputs. In part two\, we will
  focus on the classic task of learning sparse parities\, which is statisti
 cally easy but computationally difficult. We will demonstrate that SGD on 
 various neural networks (transformers\, MLPs\, etc.) successfully learns s
 parse parities\, with computational efficiency that is close to known lowe
 r bounds. Moreover\, the training curves display no apparent progress for 
 a long time\, and then quickly drop late in training. We show that despite
  this apparent delayed breakthrough in performance\, hidden progress is ac
 tually being made throughout the course of training.\n\nBased on joint wor
 k with Surbhi Goel\, Sham Kakade\, Cyril Zhang\, Boaz Barak\, and Eran Mal
 ach:\n\nhttps://arxiv.org/abs/2110.10090\n\nhttps://arxiv.org/abs/2207.087
 99\n
LOCATION:https://researchseminars.org/talk/MPML/98/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Valentin De Bortoli (Center for Sciences of Data\, ENS Ulm\, Paris
 )
DTSTART:20230316T170000Z
DTEND:20230316T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/99
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/99/">Di
 ffusion models\, theory and methodology</a>\nby Valentin De Bortoli (Cente
 r for Sciences of Data\, ENS Ulm\, Paris) as part of Mathematics\, Physics
  and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nGenerative modeling is
  the task of drawing new samples from an underlying distribution known onl
 y via an empirical measure. There exists a myriad of models to tackle this
  problem with applications in image and speech processing\, medical imagin
 g\, forecasting and protein modeling to cite a few. Among these methods di
 ffusion models are a new powerful class of generative models that exhibit 
 remarkable empirical performance. They consist of a ``noising'' stage\, wh
 ereby a diffusion is used to gradually add Gaussian noise to data\, and a 
 generative model\, which entails a ``denoising'' process defined by approx
 imating the time-reversal of the diffusion. In this talk we discuss three 
 aspects of diffusion models. First\, we will dive into the methodology beh
 ind diffusion models. Second\, we will present some of their theoretical g
 uarantees with an emphasis on their behavior under the so-called manifold 
 hypothesis. Such theoretical guarantees are non-vacuous and provide insigh
 t on the empirical behavior of these models. Finally\, I will present an e
 xtension of diffusion models to the Optimal Transport setting and introduc
 e Diffusion Schrodinger Bridges.\n
LOCATION:https://researchseminars.org/talk/MPML/99/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Memming Park (Champalimaud Foundation)
DTSTART:20230323T170000Z
DTEND:20230323T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/100
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/100/">O
 n learning signals in recurrent networks</a>\nby Memming Park (Champalimau
 d Foundation) as part of Mathematics\, Physics and Machine Learning (IST\,
  Lisbon)\n\n\nAbstract\nNeural dynamical systems with stable attractor str
 uctures such as point attractors and continuous attractors are widely hypo
 thesized to underlie meaningful temporal behavior that requires working me
 mory. However\, perhaps counterintuitively\, having good working memory is
  not sufficient for supporting useful learning signals that are necessary 
 to adapt to changes in the temporal structure of the environment. We show 
 that in addition to the well-known continuous attractors\, the periodic an
 d quasi-periodic attractors are also fundamentally capable of supporting l
 earning arbitrarily long temporal relationships. Due to the fine tuning pr
 oblem of the continuous attractors and the lack of\ntemporal fluctuations\
 , we believe the less explored quasi-periodic attractors are uniquely qual
 ified for learning to produce temporally structured behavior. Our theory h
 as wide implications for the design of artificial learning systems\, and m
 akes predictions on the observable signatures of biological neural dynamic
 s that can support temporal dependence learning. Based on our theory\, we 
 developed a new initialization scheme for artificial recurrent neural netw
 orks which outperforms standard methods for tasks that require learning te
 mporal dynamics. Finally\, we speculate on their biological implementation
 s and make predictions on neuronal dynamics.\n
LOCATION:https://researchseminars.org/talk/MPML/100/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Rongjie Lai (Rensselaer Polytechnic Institute)
DTSTART:20230420T160000Z
DTEND:20230420T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/101
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/101/">L
 earning Manifold-Structured Data using Deep Neural Networks: Theory and Ap
 plications</a>\nby Rongjie Lai (Rensselaer Polytechnic Institute) as part 
 of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract
 \nDeep artificial neural networks have made great success in many problems
  in science and engineering. In this talk\, I will discuss our recent effo
 rts to develop DNNs capable of learning non-trivial geometry information h
 idden in data. In the first part\, I will discuss our work on advocating t
 he use of a multi-chart latent space for better data representation. Inspi
 red by differential geometry\, we propose a Chart Auto-Encoder (CAE) and p
 rove a universal approximation theorem on its representation capability. C
 AE admits desirable manifold properties that conventional auto-encoders wi
 th a flat latent space fail to obey. We further establish statistical guar
 antees on the generalization error for trained CAE models and show their r
 obustness to noise. Our numerical experiments also demonstrate satisfactor
 y performance on data with complicated geometry and topology. If time perm
 its\, I will discuss our work on defining convolution on manifolds via par
 allel transport. This geometric way of defining parallel transport convolu
 tion (PTC) provides a natural combination of modeling and learning on mani
 folds. PTC allows for the construction of compactly supported filters and 
 is also robust to manifold deformations. I will demonstrate its applicatio
 ns to shape analysis and point clouds processing using PTC-nets. This talk
  is based on a series of joint work with my students and collaborators.\n
LOCATION:https://researchseminars.org/talk/MPML/101/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Gonçalo Correia (IST and Priberam Labs)
DTSTART:20230309T170000Z
DTEND:20230309T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/102
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/102/">L
 earnable Sparsity and Weak Supervision for Data-Efficient\, Transparent\, 
 and Compact Neural Models</a>\nby Gonçalo Correia (IST and Priberam Labs)
  as part of Mathematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\
 nAbstract\nNeural network models have become ubiquitous in Machine Learnin
 g literature. These models are compositions of differentiable building blo
 cks that result in dense representations of the underlying data. To obtain
  good representations\, conventional neural models require many training d
 ata points. Moreover\, those representations\, albeit capable of obtaining
  a high performance on many tasks\, are largely uninterpretable. These mod
 els are often overparameterized and give out representations that do not c
 ompactly represent the data. To address these issues\, we find solutions i
 n sparsity and various forms of weak supervision. For data-efficiency\, we
  leverage transfer learning as a form of weak supervision. The proposed mo
 del can perform similarly to models trained on millions of data points on 
 a sequence-to-sequence generation task\, even though we only train it on a
  few thousand. For transparency\, we propose a probability normalizing fun
 ction that can learn its sparsity. The model learns the sparsity it needs 
 differentiably and thus adapts it to the data according to the neural comp
 onent's role in the overall structure. We show that the proposed model imp
 roves the interpretability of a popular neural machine translation archite
 cture when compared to conventional probability normalizing functions. Fin
 ally\, for compactness\, we uncover a way to obtain exact gradients of dis
 crete and structured latent variable models efficiently. The discrete node
 s in these models can compactly represent implicit clusters and structures
  in the data\, but training them was often complex and prone to failure si
 nce it required approximations that rely on sampling or relaxations. We pr
 opose to train these models with exact gradients by parameterizing discret
 e distributions with sparse functions\, both unstructured and structured. 
 We obtain good performance on three latent variable model applications whi
 le still achieving the practicality of the approximations mentioned above.
  Through these novel contributions\, we challenge the conventional wisdom 
 that neural models cannot exhibit data-efficiency\, transparency\, or comp
 actness.\n
LOCATION:https://researchseminars.org/talk/MPML/102/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Diogo Gomes (KAUST)
DTSTART:20230504T160000Z
DTEND:20230504T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/103
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/103/">M
 athematics for data science and AI - curriculum design\, experiences\, and
  lessons learned</a>\nby Diogo Gomes (KAUST) as part of Mathematics\, Phys
 ics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nIn this talk\, we w
 ill explore the importance of mathematical foundations for AI and data sci
 ence and the design of an academic curriculum for graduate students. While
  traditional mathematics for AI and data science has focused on core techn
 iques like linear algebra\, basic probability\, and optimization methods (
 e.g.\, gradient and stochastic gradient descent)\, several advanced mathem
 atical techniques are now essential to understanding modern data science. 
 These include ideas from the calculus of variations in spaces of random va
 riables\, functional analytic methods\, ergodic theory\, control theory me
 thods in reinforcement learning\, and metrics in spaces of probability mea
 sures. We will discuss the author's experience designing an applied mathem
 atics curriculum on data science and draw on the author's experience and l
 essons learned in teaching an advanced course on the mathematical foundati
 ons of data science. This talk aims to promote discussion and exchange of 
 ideas on how mathematicians can play an important role in AI and data scie
 nce and better equip our students to excel in this field.\n
LOCATION:https://researchseminars.org/talk/MPML/103/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Harry Desmond (University of Portsmouth)
DTSTART:20230511T160000Z
DTEND:20230511T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/104
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/104/">E
 xhaustive Symbolic Regression (or how to find the best function for your d
 ata)</a>\nby Harry Desmond (University of Portsmouth) as part of Mathemati
 cs\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nSymbolic r
 egression aims to find optimal functional representation of datasets\, wit
 h broad applications across science. This is traditionally done using a "g
 enetic algorithm" which stochastically searches function space using an ev
 olution-inspired method for generating new trial functions. Motivated by t
 he uncertainties inherent in this approach -- and its failure on seemingly
  simple test cases -- I will describe a new method which exhaustively sear
 ches and evaluates function space. Coupled to a model selection principle 
 based on minimum description length\, Exhaustive Symbolic Regression is gu
 aranteed to find the simple equations that optimally balance simplicity wi
 th accuracy on any dataset. I will describe how the method works and showc
 ase it on Hubble rate measurements and dynamical galaxy data.\n\nBased on 
 work with Deaglan Bartlett and Pedro G. Ferreira: <br>\nhttps://arxiv.org/
 abs/2211.11461 <br>\nhttps://arxiv.org/abs/2301.04368\n
LOCATION:https://researchseminars.org/talk/MPML/104/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Paulo Rosa (Deimos)
DTSTART:20230427T160000Z
DTEND:20230427T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/105
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/105/">D
 eep Reinforcement Learning based Integrated Guidance and Control for a Lau
 ncher Landing Problem</a>\nby Paulo Rosa (Deimos) as part of Mathematics\,
  Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nDeep Reinforce
 ment Learning (Deep-RL) has received considerable attention in recent year
 s due to its ability to make an agent learn how to take optimal control ac
 tions\, given rich observation data via the maximization of a reward funct
 ion. Future space missions will need new on-board autonomy capabilities wi
 th increasingly complex requirements at the limits of the vehicle performa
 nce. This justifies the use of machine learning based techniques\, in part
 icular reinforcement learning in order to allow exploring the edge of the 
 performance trade-off space. The guidance and control systems development 
 for Reusable Launch Vehicles (RLV) can take advantage of reinforcement lea
 rning techniques for optimal adaption in the face of multi-objective requi
 rements and uncertain scenarios.\n\nIn AI4GNC - a project funded by the Eu
 ropean Space Agency (ESA)\, led by DEIMOS and participated by INESC-ID\, t
 he University of Lund\, and TASC - a Deep-RL algorithm was used to train a
 n actor-critic agent to simultaneously control the engine thrust magnitude
  and the two TVC gimbal angles to land a RLV in 6-DoF simulation. The desi
 gn followed an incremental approach\, progressively augmenting the number 
 of degrees of freedom and introducing more complexity factors such as nonl
 inearity in models. Ultimately\, the full 6-DoF problem was addressed usin
 g a high fidelity simulator that includes a nonlinear actuator model and a
  realistic vehicle aerodynamic model. Starting from an initial vehicle sta
 te along a reentry trajectory\, the problem consists of precisely land the
  RLV while ensuring system requirements satisfaction\, such as saturation 
 and rate limits in the actuation\, and aiming at fuel consumption optimali
 ty. The Deep Deterministic Policy Gradient (DDPG) algorithm was adopted as
  candidate strategy to allow the design of an integrated guidance and cont
 rol algorithm in continuous action and observation spaces.\n\nThe results 
 obtained are very satisfactory in terms of landing accuracy and fuel consu
 mption. These results were also compared to a more classical and industria
 lly used solution\, due to its capability to yield satisfactory landing ac
 curacy and fuel consumption\, composed of a successive convexification gui
 dance and a PID controller tuned independently for the non-disturbed nomin
 al scenario. A reachability analysis was also performed to assess the stab
 ility and robustness of the closed-loop system composed by the integrated 
 guidance and control NN\, trained for the 1-DoF scenario\, and the RLV dyn
 amics.\n\nTaking into account the fidelity of the benchmark adopted and th
 e results obtained\, this approach is deemed to have a significant potenti
 al for further developments and ultimately space industry applications\, s
 uch as In-Orbit Servicing (IOS) and Active Debris Removal (ADR)\, that als
 o require a high level of autonomy.\n
LOCATION:https://researchseminars.org/talk/MPML/105/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Artemy Kolchinsky (Universal Biology Institute\, University of Tok
 yo)
DTSTART:20230622T160000Z
DTEND:20230622T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/106
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/106/">I
 nformation geometry for nonequilibrium processes</a>\nby Artemy Kolchinsky
  (Universal Biology Institute\, University of Tokyo) as part of Mathematic
 s\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nRecently\, 
 there has been dramatic progress in nonequilibrium thermodynamics\, with d
 iverse applications in biological and chemical systems. The central quanti
 ty of interest in the field is “entropy production” (EP)\, which refle
 cts the increase of the entropy of a system and its environment. Major que
 stions of interest include (1) quantitative tradeoffs between EP and perfo
 rmance measures like speed and precision\, (2) inference of EP from data\,
  and (3) decomposition of EP into contributions from different sources of 
 dissipation. In this work\, we study the thermodynamics of nonequilibrium 
 processes by considering the information geometry of fluxes. Our approach 
 can be seen as a dynamical generalization of existing work on the informat
 ion geometry of probability distributions considered at a given instant in
  time. It is applicable to a broad range of nonequilibrium processes\, inc
 luding nonlinear ones that exhibit oscillations and/or chaos\, and it has 
 implications for thermodynamic tradeoffs\, thermodynamic inference\, and d
 ecompositions of EP. As one application\, we derive a universal decomposit
 ion of EP into “excess” and “housekeeping” contributions\, represe
 nting contributions from nonstationarity and cyclic fluxes respectively.\n
 \n(joint work with Andreas Dechant\, Kohei Yoshimura\, Sosuke Ito. arXiv:2
 206.14599)\n
LOCATION:https://researchseminars.org/talk/MPML/106/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Rui Castro (Mathematics Department\, TU Eindhoven)
DTSTART:20230518T160000Z
DTEND:20230518T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/107
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/107/">A
 nomaly detection for a large number of streams: a permutation/rank-based h
 igher criticism approach</a>\nby Rui Castro (Mathematics Department\, TU E
 indhoven) as part of Mathematics\, Physics and Machine Learning (IST\, Lis
 bon)\n\n\nAbstract\nAnomaly detection when observing a large number of dat
 a streams is essential in a variety of applications\, ranging from epidemi
 ological studies to monitoring of complex systems. High-dimensional scenar
 ios are usually tackled with scan-statistics and related methods\, requiri
 ng stringent distributional assumptions for proper test calibration. In th
 is talk we take a non-parametric stance\, and introduce two variants of th
 e higher criticism test that do not require knowledge of the null distribu
 tion for proper calibration. In the first variant we calibrate the test by
  permutation\, while in the second variant we use a rank-based approach. B
 oth methodologies result in exact tests in finite samples. Our permutation
  methodology is applicable when observations within null streams are indep
 endent and identically distributed\, and we show this methodology is asymp
 totically optimal in the wide class of exponential models. Our rank-based 
 methodology is more flexible\, and only requires observations within null 
 streams to be independent. We provide an asymptotic characterization of th
 e power of the test in terms of the probability of mis-ranking null observ
 ations\, showing that the asymptotic power loss (relative to an oracle tes
 t) is minimal for many common models. As the proposed statistics do not re
 ly on asymptotic approximations\, they typically perform better than popul
 ar variants of higher criticism relying on such approximations. Finally\, 
 we demonstrate the use of these methodologies when monitoring the content 
 uniformity of an active ingredient for a batch-produced drug product\, and
  monitoring the daily number of COVID-19 cases in the Netherlands.\n\nBase
 d on joint work with Ivo Stoepker\, Ery Arias-Castro and Edwin van de den 
 Heuvel:\nhttps://arxiv.org/abs/2009.03117\n
LOCATION:https://researchseminars.org/talk/MPML/107/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Sara Magliacane (University of Amsterdam and MIT-IBM Watson AI Lab
 )
DTSTART:20230608T160000Z
DTEND:20230608T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/108
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/108/">C
 ausal vs causality-inspired representation learning</a>\nby Sara Magliacan
 e (University of Amsterdam and MIT-IBM Watson AI Lab) as part of Mathemati
 cs\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\n<p>Causal 
 representation learning (CRL) aims at learning causal factors and their ca
 usal relations from high-dimensional observations\, e.g. images. In genera
 l\, this is an ill-posed problem\, but under certain assumptions or with t
 he help of additional information or interventions\, we are able to guaran
 tee that the representations we learn are corresponding to some true under
 lying causal factors up to some equivalence class.<br />\nIn this talk I w
 ill first present CITRIS (<a href="https://proceedings.mlr.press/v162/lipp
 e22a/lippe22a.pdf" rel="noreferrer" target="_blank">https://proceedings.ml
 r.press/v162/lippe22a/lippe22a.pdf</a>)\, a variational autoencoder framew
 ork for causal representation learning from temporal sequences of images\,
  in systems in which we can perform interventions. CITRIS exploits tempora
 lity and observing intervention targets to identify scalar and multidimens
 ional causal factors\, such as 3D rotation angles. In experiments on 3D re
 ndered image sequences\, CITRIS outperforms previous methods on recovering
  the underlying causal variables. Moreover\, using pretrained autoencoders
 \, CITRIS can even generalize to unseen instantiations of causal factors.<
 br />\n<br />\nWhile CRL is an exciting and promising new field of researc
 h\, the assumptions required by CITRIS and other current CRL methods can b
 e difficult to satisfy in many settings. Moreover\, in many practical case
 s learning representations that are not guaranteed to be fully causal\, bu
 t exploit some ideas from causality\, can still be extremely useful. As ex
 amples\, I will describe some of our work on exploiting these "causality-i
 nspired" representations for adapting policies across domains in RL (<a hr
 ef="https://openreview.net/forum?id=8H5bpVwvt5" rel="noreferrer" target="_
 blank">https://openreview.net/forum?id=8H5bpVwvt5</a>) and to nonstationar
 y environments (<a href="https://openreview.net/forum?id=VQ9fogN1q6e" rel=
 "noreferrer" target="_blank">https://openreview.net/forum?id=VQ9fogN1q6e</
 a>)\, and how learning a factored graphical representations (even if not n
 ecessarily causal) can be beneficial in these (and possibly other) setting
 s.</p>\n
LOCATION:https://researchseminars.org/talk/MPML/108/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Mário Figueiredo (Instituto Superior Técnico and IT)
DTSTART:20230615T160000Z
DTEND:20230615T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/109
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/109/">C
 ausal Discovery from Observations: Introduction and Some Recent Advances</
 a>\nby Mário Figueiredo (Instituto Superior Técnico and IT) as part of M
 athematics\, Physics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nCa
 usal discovery is an active research field that aims to uncover the underl
 ying causal mechanisms that drive the relationship between a collection of
  variables and which has applications in many areas\, including medicine\,
  biology\, economics\, and social sciences. In principle\, identifying cau
 sal relationships requires interventions. However\, intervening is often i
 mpossible\, impractical\, or unethical\, which has stimulated much researc
 h on causal discovery from purely observational data or mixed observationa
 l-interventional data. In this talk\, after overviewing the causal discove
 ry field\, I will discuss some recent advances\, namely on causal discover
 y from data with latent interventions and on what is the quintessential ca
 usal discovery problem: distinguishing the cause from the effect on a pair
  of dependent variables.\n
LOCATION:https://researchseminars.org/talk/MPML/109/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Francisco Förster Burón (Universidad de Chile)
DTSTART:20240111T170000Z
DTEND:20240111T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/110
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/110/">T
 he ALeRCE astronomical alert broker</a>\nby Francisco Förster Burón (Uni
 versidad de Chile) as part of Mathematics\, Physics and Machine Learning (
 IST\, Lisbon)\n\n\nAbstract\nA new generation of large aperture and large 
 field of view telescopes is allowing the exploration of large volumes of t
 he Universe in an unprecedented fashion. In order to take advantage of the
 se new telescopes\, notably the Vera C. Rubin Observatory\, a new time dom
 ain ecosystem is developing. Among the tools required are fast machine lea
 rning aided discovery and classification algorithms\, interoperable tools 
 to allow for an effective communication with the community and follow-up t
 elescopes\, and new models and tools to extract the most physical knowledg
 e from these observations. In this talk I will review the challenges and p
 rogress of building one of these systems: the Automatic Learning for the R
 apid Classification of Events (ALeRCE) astronomical alert broker. ALeRCE (
 http://alerce.science/) is an alert annotation and classification system l
 ed by an interdisciplinary and interinstitutional group of scientists from
  Chile since 2019. ALeRCE is focused around three scientific cases: transi
 ents\, variable stars and active galactic nuclei. Thanks to its state-of-t
 he-art machine learning models\, ALeRCE has become the 3rd group to report
  most transient candidates to the Transient Name Server\, and it is enabli
 ng new science with different astrophysical objects\, e.g. AGN science. I 
 will discuss some of the challenges associated with the problem of alert c
 lassification\, including the ingestion of multiple alert streams\, annota
 tion\, database management\, training set building\, feature computation a
 nd distributed processing\, machine learning classification and visualizat
 ion\, or the challenges of working in large interdisciplinary teams. I wil
 l also show some results based on the real‐time ingestion and classifica
 tion using the Zwicky Transient Facility (ZTF) alert stream as input\, as 
 well as some of the tools available.\n
LOCATION:https://researchseminars.org/talk/MPML/110/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Olga Mula (TU Eindhoven)
DTSTART:20230922T130000Z
DTEND:20230922T140000Z
DTSTAMP:20260422T212558Z
UID:MPML/111
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/111/">O
 ptimal State and Parameter Estimation Algorithms and Applications to Biome
 dical Problems</a>\nby Olga Mula (TU Eindhoven) as part of Mathematics\, P
 hysics and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nIn this talk\, I
  will present an overview of recent works aiming at solving inverse proble
 ms (state and parameter estimation) by combining optimally measurement obs
 ervations and parametrized PDE models. After defining a notion of optimal 
 performance in terms of the smallest possible reconstruction error that an
 y reconstruction algorithm can achieve\, I will present practical numerica
 l algorithms based on nonlinear reduced models for which we can prove that
  they can deliver a performance close to optimal. The proposed concepts ma
 y be viewed as exploring alternatives to Bayesian inversion in favor of mo
 re deterministic notions of accuracy quantification. I will illustrate the
  performance of the approach on simple benchmark examples and we will also
  discuss applications of the methodology to biomedical problems which are 
 challenging due to shape variability.\n\nhttps://arxiv.org/pdf/2203.07769.
 pdf\nhttps://arxiv.org/pdf/2009.02687.pdf\n
LOCATION:https://researchseminars.org/talk/MPML/111/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Pedro Domingos (University of Washington)
DTSTART:20240215T170000Z
DTEND:20240215T180000Z
DTSTAMP:20260422T212558Z
UID:MPML/112
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/112/">D
 eep Networks Are Kernel Machines</a>\nby Pedro Domingos (University of Was
 hington) as part of Mathematics\, Physics and Machine Learning (IST\, Lisb
 on)\n\n\nAbstract\nDeep learning's successes are often attributed to its a
 bility to automatically discover new representations of the data\, rather 
 than relying on handcrafted features like other learning methods. In this 
 talk\, however\, I will show that deep networks learned by the standard gr
 adient descent algorithm are in fact mathematically approximately equivale
 nt to kernel machines\, a learning method that simply memorizes the data a
 nd uses it directly for prediction via a similarity function (the kernel).
  This greatly enhances the interpretability of deep network weights\, by e
 lucidating that they are effectively a superposition of the training examp
 les. The network architecture incorporates knowledge of the target functio
 n into the kernel. The talk will include a discussion of both the main ide
 as behind this result and some of its more startling consequences for deep
  learning\, kernel machines\, and machine learning at large.\n
LOCATION:https://researchseminars.org/talk/MPML/112/
END:VEVENT
BEGIN:VEVENT
SUMMARY:Kathryn Hess (EPFL)
DTSTART:20240606T160000Z
DTEND:20240606T170000Z
DTSTAMP:20260422T212558Z
UID:MPML/113
DESCRIPTION:Title: <a href="https://researchseminars.org/talk/MPML/113/">O
 f mice and men</a>\nby Kathryn Hess (EPFL) as part of Mathematics\, Physic
 s and Machine Learning (IST\, Lisbon)\n\n\nAbstract\nMotivated by the desi
 re to automate classification of neuron morphologies\, we designed a topol
 ogical signature\, the Topological Morphology Descriptor (TMD)\, that assi
 gns a "barcode" to any any finite binary tree embedded in ${\\mathbb R}^3$
 . Using the TMD we performed an objective\, stable classification of pyram
 idal cells in the rat neocortex\, based only on the shape of their dendrit
 es.\n\nIn this talk\, I will introduce the TMD\, then focus on a very rece
 nt application to comparing mouse and human cortical neurons and character
 izing the differences between them. I'll also briefly discuss the role of 
 machine learning in our work.\n\nThis talk is based on collaborations led 
 by Lida Kanari of the Blue Brain Project.\n
LOCATION:https://researchseminars.org/talk/MPML/113/
END:VEVENT
END:VCALENDAR
