Search Results for author: Julius Berner

Found 14 papers, 8 papers with code

Pretraining Codomain Attention Neural Operators for Solving Multiphysics PDEs

1 code implementation19 Mar 2024 Md Ashiqur Rahman, Robert Joseph George, Mogab Elleithy, Daniel Leibovici, Zongyi Li, Boris Bonev, Colin White, Julius Berner, Raymond A. Yeh, Jean Kossaifi, Kamyar Azizzadenesheli, Anima Anandkumar

On complex downstream tasks with limited data, such as fluid flow simulations and fluid-structure interactions, we found CoDA-NO to outperform existing methods on the few-shot learning task by over $36\%$.

Few-Shot Learning Self-Supervised Learning

DPOT: Auto-Regressive Denoising Operator Transformer for Large-Scale PDE Pre-Training

1 code implementation6 Mar 2024 Zhongkai Hao, Chang Su, Songming Liu, Julius Berner, Chengyang Ying, Hang Su, Anima Anandkumar, Jian Song, Jun Zhu

Pre-training has been investigated to improve the efficiency and performance of training neural operators in data-scarce settings.

Denoising

Neural Operators with Localized Integral and Differential Kernels

no code implementations26 Feb 2024 Miguel Liu-Schiaffini, Julius Berner, Boris Bonev, Thorsten Kurth, Kamyar Azizzadenesheli, Anima Anandkumar

In this work, we present a principled approach to operator learning that can capture local features under two frameworks by learning differential operators and integral operators with locally supported kernels.

Operator learning

Large Language Models for Mathematicians

no code implementations7 Dec 2023 Simon Frieder, Julius Berner, Philipp Petersen, Thomas Lukasiewicz

Large language models (LLMs) such as ChatGPT have received immense interest for their general-purpose language understanding and, in particular, their ability to generate high-quality text or computer code.

Improved sampling via learned diffusions

1 code implementation3 Jul 2023 Lorenz Richter, Julius Berner, Guan-Horng Liu

Recently, a series of papers proposed deep learning-based approaches to sample from unnormalized target densities using controlled diffusion processes.

Mathematical Capabilities of ChatGPT

2 code implementations NeurIPS 2023 Simon Frieder, Luca Pinchetti, Alexis Chevalier, Ryan-Rhys Griffiths, Tommaso Salvatori, Thomas Lukasiewicz, Philipp Christian Petersen, Julius Berner

We investigate the mathematical capabilities of two iterations of ChatGPT (released 9-January-2023 and 30-January-2023) and of GPT-4 by testing them on publicly available datasets, as well as hand-crafted ones, using a novel methodology.

Elementary Mathematics Math +2

An optimal control perspective on diffusion-based generative modeling

1 code implementation2 Nov 2022 Julius Berner, Lorenz Richter, Karen Ullrich

In particular, we derive a Hamilton-Jacobi-Bellman equation that governs the evolution of the log-densities of the underlying SDE marginals.

Robust SDE-Based Variational Formulations for Solving Linear PDEs via Deep Learning

1 code implementation21 Jun 2022 Lorenz Richter, Julius Berner

The combination of Monte Carlo methods and deep learning has recently led to efficient algorithms for solving partial differential equations (PDEs) in high dimensions.

Learning ReLU networks to high uniform accuracy is intractable

1 code implementation26 May 2022 Julius Berner, Philipp Grohs, Felix Voigtlaender

Statistical learning theory provides bounds on the necessary number of training samples needed to reach a prescribed accuracy in a learning problem formulated over a given target class.

Learning Theory Vocal Bursts Intensity Prediction

Numerically Solving Parametric Families of High-Dimensional Kolmogorov Partial Differential Equations via Deep Learning

1 code implementation NeurIPS 2020 Julius Berner, Markus Dablander, Philipp Grohs

We show that a single deep neural network trained on simulated data is capable of learning the solution functions of an entire family of PDEs on a full space-time region.

How degenerate is the parametrization of neural networks with the ReLU activation function?

no code implementations NeurIPS 2019 Julius Berner, Dennis Elbrächter, Philipp Grohs

Approximation capabilities of neural networks can be used to deal with the latter non-convexity, which allows us to establish that for sufficiently large networks local minima of a regularized optimization problem on the realization space are almost optimal.

Towards a regularity theory for ReLU networks -- chain rule and global error estimates

no code implementations13 May 2019 Julius Berner, Dennis Elbrächter, Philipp Grohs, Arnulf Jentzen

Although for neural networks with locally Lipschitz continuous activation functions the classical derivative exists almost everywhere, the standard chain rule is in general not applicable.

Analysis of the Generalization Error: Empirical Risk Minimization over Deep Artificial Neural Networks Overcomes the Curse of Dimensionality in the Numerical Approximation of Black-Scholes Partial Differential Equations

no code implementations9 Sep 2018 Julius Berner, Philipp Grohs, Arnulf Jentzen

It can be concluded that ERM over deep neural network hypothesis classes overcomes the curse of dimensionality for the numerical solution of linear Kolmogorov equations with affine coefficients.

Cannot find the paper you are looking for? You can Submit a new open access paper.