•     EPSRC Postdoctoral Fellowship

    I was awarded an EPSRC Fellowship! The title is “Regularisation theory in the data driven setting”. My goal is to extend regularisation theory to the setting where there is no direct access to the forward operator at the time of solving the inverse problem and only input-output training pairs are available. Such pairs can be either collected experimentally or obtained from a computationally expensive model prior to solving the inverse problem. The latter scenario is relevant for time-sensitive applications where near real-time reconstructions are required.
    Furthermore, the model free setting is the natural habitat of neural networks, and my long-term goal is to better understand their regularisation properties in the context of ill-posed inverse problems in infinite dimensions.
    The fellowship is due to start in April 2021.

  •     Data driven regularisation

    Our paper with Andrea Aspri and Otmar Scherzer on Data Driven Regularization by Projection has appeared in Inverse Problems! We show that regularisation can be defined and rigorously studied in the setting when there is no numerical access to the forward operator and the operator is given only via input-output training pairs. Such pairs can be either collected experimentally or obtained from a computationally expensive model prior to solving the inverse problem (which may be useful in time-sensitive applications). We show connections to classical regularisation methods such as regularisation by projection and variational regularisation, study stability of the reconstructions and present numerical experiments on data driven inversion of the Radon transform.
    The paper is in open acces and can be found here: DOI:10.1088/1361-6420/abb61b

  •     SIAM Imaging Science 2020

    I gave a talk about our paper on variational regularisation in Banach lattices at the SIAM Imaging Science conference. Thank you Nicolai Riis for inviting me to your mini-sympsium! And also many thanks to the organisers Michael Elad and Stacey Levine for organising the conference under current difficult circumstances.
    Slides are available here.

  •     Deep learning and non-linear spectral analysis

    Our preprint with Tamara Grossmann, Guy Gilboa and Carola Schönlieb on Deeply Learned Spectral Total Variation Decomposition is online! We compute a non-linear Total Variation decomposition of an image 10000 faster then classical methods.
    Check it out here: arXiv:2006.10004

  •     Inverse problems, Banach lattices and general fidelity functions

    Our preprint with Leon Bungert, Martin Burger and Carola Schönlieb on Variational regularisation for inverse problems with imperfect forward operators and general noise models is now online! We analyse inverse problems where the operator contains errors that can be described by an interval in a Banach lattice (e.g., kernel of an integral operator with pointwise errors) and the data are corrupted by noise that can be described using some data fidelity function. Our results apply to, e.g., Gaussian, salt-and-pepper, Poisson and mixed noise.
    Check it out here: arXiv:2005.14131

  •     Computing distance functions using gradient flows

    Our preprint with Leon Bungert and Martin Burger on L-infinity variational problems and relations to distance functions is now online! We show that the distance function is the unique ground state of a certain L-infinity variational problem and devise an efficient numerical algorithm for computing distance functions on graphs.
    Check it out here: arXiv:2001.07411

subscribe via RSS