Humboldt-Universität zu Berlin - Mathematisch-Naturwissenschaftliche Fakultät - Institut für Mathematik

Forschungsseminar Algorithmische Optimierung (AGs Hante/Walther)

Ort: Rudower Chaussee 25, Raum 2.417

Zeit: Donnerstag, 15:15 Uhr

Studierende und Gäste sind herzlich willkommen.

 

Vorträge im Wintersemester 2025/26

 

 

 

   
06.11.2025 Nicolas Gauger, University of Kaiserslautern-Landau (RPTU)
  Beginn: 11:00 Uhr, Achtung: Abweichender Termin
 
Solving Distributionally Robust Shape Design Problems with Applications in Aerospacey 🖉

We formulate and solve data-driven aerodynamic shape design problems with distributionally robust optimization (DRO) approaches. We study the connections between a class of DRO and robust design optimization,which is classically based on the mean-variance (standard deviation) optimization formulation introduced by Taguchi. Our findings provide a new perspective on robust design by enabling statistically

principled and data-driven approaches to quantify the trade-offs between robustness and performance. Furthermore, we introduce a new method to solve DRO problems applied to computationally expensive

PDE-constrained optimization problems by leveraging surrogate models. We demonstrate the method through design applications in aerospace.

   
24.11.25 Marvin Pförtner, Universität Tübingen
 
Physics-Informed Gaussian Process Regression Generalizes Linear PDE Solvers 🖉

Linear partial differential equations (PDEs) are an important, widely applied class of mechanistic models, describing physical processes such as heat transfer, electromagnetism, and wave propagation. In practice, specialized numerical methods based on discretization are used to solve PDEs. They generally use an estimate of the unknown model parameters and, if available, physical measurements for initialization. Such solvers are often embedded into larger scientific models with a downstream application and thus error quantification plays a key role. However, by ignoring parameter and measurement uncertainty, classical PDE solvers may fail to produce consistent estimates of their inherent approximation error. In this work, we approach this problem in a principled fashion by interpreting solving linear PDEs as physics-informed Gaussian process (GP) regression. Our framework is based on a key generalization of the Gaussian process inference theorem to observations made via an arbitrary bounded linear operator. Crucially, this probabilistic viewpoint allows to (1) quantify the inherent discretization error; (2) propagate uncertainty about the model parameters to the solution; and (3) condition on noisy measurements. Demonstrating the strength of this formulation, we prove that it strictly generalizes methods of weighted residuals, a central class of PDE solvers including collocation, finite volume, pseudospectral, and (generalized) Galerkin methods such as finite element and spectral methods. This class can thus be directly equipped with a structured error estimate. In summary, our results enable the seamless integration of mechanistic models as modular building blocks into probabilistic models by blurring the boundaries between numerical analysis and Bayesian inference.

  Beginn: 15:00 Achtung: Abweichender Termin
   
04.12.25 Sebastian Pokutta, ZIB und TU Berlin
 
A gentle introduction to Frank-Wolfe Algorithms 🖉

This talk focuses on constrained optimization problems that can be efficiently solved using first-order methods, particularly Frank-Wolfe methods (also known as Conditional Gradients). These algorithms have emerged as a crucial class of methods for minimizing smooth convex functions over polytopes, and their applicability extends beyond this domain. Recently, they have garnered significant attention due to their ability to facilitate structured optimization, a key aspect in some machine learning applications. I will provide a broad overview of these methods, highlighting their applications and presenting recent advances in both traditional optimization and machine learning. Time permitting, I will also discuss an extension of these methods to the mixed-integer convex optimization setting.

   
18.12.25 Martin Skutella, TU Berlin
 
Integer Multiflows and Cut Conditions 🖉

For directed graphs with arc capacities, Nagamochi and Ibaraki (1989) showed that if the cut condition guarantees the existence of a fractional multiflow, and this implication holds in a certain hereditary way, then the cut condition also guarantees the existence of an integer multiflow. Motivated by our recent results with Mohammed Majthoub Almoghrabi and Philipp Warode on integer and unsplittable multiflows in series-parallel digraphs, we discuss a hierarchy of cut conditions and a somewhat refined version of the Nagamochi-Ibaraki Theorem. 

   
22.01.26 Raphael Kuess, HU Berlin
  TBA
   
   
  weitere Termine folgen
   
   

Vorträge im Sommersemester 2025

 

   
24.04.2025 Robert Luce, Gurobi Optimization
 
Solving Nonlinear Problems to Global Optimality 🖉

In this talk, we provide an overview of Gurobi's algorithmic
components for solving nonlinear optimization problems to global optimality. In essence, we extend our existing mixed-integer programming (MIP) framework to handle such problems. This includes our presolve algorithms, an extension of the branch-and-bound method utilizing spatial relaxations, and an interior point algorithm for nonlinear problems, which serves as a primal heuristic to find high-quality solutions. As a result, we can compute solutions to nonlinear optimization  problems along with certificates for global optimality. Finally, we have extended gurobipy to facilitate the easy formulation of expression-based nonlinear optimization problems in Python.
Gurobi's nonlinear solver applies to explicit expression-based
constraints and does not require the supply of derivative data.

   
03.06.25 Tim Siebert, Humboldt-Universität zu Berlin
  Collapsing Taylor Mode Automatic Differentiation
  Beginn: 15:00 Uhr, Achtung: Abweichender Termin
   
17.06.25 Sri Tadinada, Humboldt-Universität zu Berlin
 
Abs-Smooth Frank-Wolfe method for convex functions 🖉

The Abs-Smooth Frank-Wolfe algorithm (ASFW) is a non-smooth variant of the popular Frank-Wolfe algorithms. In this talk we sketch and analyze the "vanilla" and the "heavy-ball" variants of the ASFW algorithm. We provide stronger and more general primal-dual convergence results for ASFW when applied in the convex setting. We derive a convergence rate for our algorithm which is identical to the smooth case. So far, there is limited understanding of accelerated convergence regimes in the context of ASFW. We also provide some answers in this context by looking into some special cases.

  Beginn: 15:00 Uhr, Achtung: Abweichender Termin
   
19.06.25 Rowan Turner, University of Edinburgh
  A tailored, matrix free interior point method for fast optimization on gas networks
  online talk, zoom link
   
03.07.2025 Oliver Sander, Technische Universität Dresden
 
Finsler geodesics and finite-strain plasticity 🖉

The theory of energetic rate-independent systems is an elegant way to describe nonlinear systems in mechanics and other fields. One particular advantage is that it yields a natural time discretization that consists of a sequence of minimization problems. Unfortunately, in many interesting cases the objective functional is only given implicitly as the solution of a second minimization problem for a
curve length in the state space. Therefore, its evaluation and obtaining derivatives can be very costly. Instead, we present a transformation based on the Finsler exponential map that turns the second minimization problem into an initial-value-problem for a second-order ODE. Solutions of this can be found much cheaper numerically, or may even be available in closed form. We show examples of this construction, and how to use it to obtain fast and robust Proximal Newton solvers for finite-strain elastoplasticity. 

   
07.07.25 Adrian Schmidt, Humboldt-Universität zu Berlin
  Morse theory for abs-smooth optimization problems 
  Beginn: 12:30 Uhr, Achtung: Abweichender Termin
   
15.07.25 Yves Jäckle, Zuse-Institut Berlin
 
BerLean 🖉

tLean is an interactive theorem prover and programming language. It allows its users to write definitions and proofs from mathematics or computer science in a formal language, so that they may be verified. We'll introduce formalization in Lean with a simple study of linked lists. Then, we'll proceed with a discussion of what it means to verify algorithms in Lean, and how it differs from algorithms that produce proofs. Finally, we'll implement a variant of subgradient descent in Lean and prove a convergence theorem for it.

  Beginn: 15:00 Uhr, Achtung: Abweichender Termin