21 - Deeply learned regularisation for inverse problems [ID:35278]
50 von 490 angezeigt

Okay, thank you very much, Leon. Thanks a lot, Daniel, for the really nice introduction.

And thank you everyone for switching who switched on the cameras. It's nice to see you. And

actually, Josh, thanks a lot for advertising it in Cambridge. I can see lots of Cambridge

people here. Okay. Well, okay. So let me actually just start by thanking my group, my research

group, without whom nothing that I'm going to show you today would have been possible.

So they are doing some tremendous work, and I would like to acknowledge them. They're

great. And I would also like to acknowledge all the funders from government, from private

trusts and from industry who are supporting our research. And then, and I only do this

advertisement because it's topic is connected to the topic of this talk. I'd like to call

to anyone who is looking for a postdoc that we are looking for postdocs in mathematics

to join a relatively large research program. We have, or we are starting with the University

of Bath and with University College London on the mathematical foundations of deep learning.

So please do get in touch or I'll try to later paste this link in the chat if you're interested.

Have a look at this website where you can also find a link to the postdoc positions.

Okay. So what I would like to do in this presentation is to convey a couple of messages. First one

is I would like to convey really my fascination for deep neural networks and the capabilities

of deep neural networks that I have gathered, I would say in the last couple of years, and

in particular their promise and their potential for helping us to solve inverse problems for

two reasons. One is that they are very, very good in capturing structure and capturing

pattern information and data. And we can use that to our advantage to produce, to help

us produce highly accurate solutions to inverse problems. Accurate actually both qualitatively

as well as quantitatively as I will show you later. And the second reason is that for all

of us who have been solving inverse problems in the past with kind of maybe traditional

or statistical mathematical approaches, all of those will appreciate that mostly these

type of solutions are very, very costly to computationally achieve because they boil

down to very large scale non-convex, variational problems or non-linear PDEs that we need to

solve. And that is definitely in contrast to this an advantage of deep neural networks,

which once they're trained, it's an explicit application of linear and non-linear operations.

So once they're trained, they are cheap. I mean, apart from always a kind of memory bottleneck

that you have depending on how large your inverse problem is. So that's really the main

point I would like to convey. The other one is that deep learning doesn't work as a hammer

for inverse problem in the sense that you can't use off the shelf deep learning solutions

and just make them work for inverse problems if you really are interested in solving them

for practical problems where you need mathematical guarantees to get robust solutions and to

get solutions which generalize to the problem you're interested in solving to not just the

tiniest small data set that you're training with, but that actually have the potential

to solve your problem in general. So you need to combine, therefore, deep neural networks

with mathematical and statistical modeling. And in contrast to that, and this is where

all of us are called to change things, most of the approaches that are out there are actually

lacking mathematical scrutiny. And so this is something we need to change. And one attempt

to go, let's say, in the right direction, and I emphasize it's really one attempt, is

to combine deep neural networks with so-called variational models. And I will show you what

variational models are in a moment. And all of what I'm going to tell you about draws

a lot of the information and the knowledge that we have gained in writing this review

paper a couple of years ago with Simon Erich, Peter Maas, and Uso Nugtem.

So what is an inverse problem? And I'm mainly interested in inverse problems where the solution

is an image, so inverse imaging problems. So an inverse problem always appears when

what you're interested in is not directly what you're measuring, but the data is an

indirect measurement of the image, in our case, that we want to reconstruct. So just

putting this in one equation, y are my measurements, u is the image that I want to reconstruct.

Teil einer Videoserie :

Zugänglich über

Offener Zugang

Dauer

00:58:36 Min

Aufnahmedatum

2021-06-29

Hochgeladen am

2021-07-03 00:47:56

Sprache

en-US

Carola-Bibiane Schönlieb (Uni Cambridge) on "Deeply learned regularisation for inverse problems"

 Inverse problems are about the reconstruction of an unknown physical quantity from indirect measurements. In imaging, they appear in a variety of places, from medical imaging, for instance MRI or CT, to remote sensing, for instance Radar, to material sciences and molecular biology, for instance electron microscopy. Here, imaging is a tool for looking inside specimen, resolving structures beyond the scale visible to the naked eye, and to quantify them. It is a mean for diagnosis, prediction and discovery.
Most inverse problems of interest are ill-posed and require appropriate mathematical treatment for recovering meaningful solutions. Classically, inversion approaches are derived almost conclusively in a knowledge driven manner, constituting handcrafted mathematical models. Examples include variational regularization methods with Tikhonov regularisation, the total variation and several sparsity-promoting regularizers such as the L1 norm of Wavelet coefficients of the solution. While such handcrafted approaches deliver mathematically rigorous and computationally robust solutions to inverse problems, they are also limited by our ability to model solution properties accurately and to realise these approaches in a computationally efficient manner.
Recently, a new paradigm has been introduced to the regularisation of inverse problems, which derives regularised solutions to inverse problems in a data driven way. Here, the inversion approach is not mathematically modelled in the classical sense, but modelled by highly over-parametrised models, typically deep neural networks, that are adapted to the inverse problems at hand by appropriately selected (and usually plenty of) training data. Current approaches that follow this new paradigm distinguish themselves through solution accuracies paired with computational efficieny that were previously unconceivable.
In this talk I will provide a glimpse into such deep learning approaches and some of their mathematical properties. I will finish with open problems and future research perspectives.  

Tags

functional minimization methods framework approximation control distance reconstruction energy deep search basic weights models measure layer activation problem example propagation
Einbetten
Wordpress FAU Plugin
iFrame
Teilen