Search Results for author: Lingjiong Zhu

Found 29 papers, 3 papers with code

Differential Privacy of Noisy (S)GD under Heavy-Tailed Perturbations

no code implementations4 Mar 2024 Umut Şimşekli, Mert Gürbüzbalaban, Sinan Yildirim, Lingjiong Zhu

Injecting heavy-tailed noise to the iterates of stochastic gradient descent (SGD) has received increasing attention over the past few years.

Learning Theory

Short-maturity asymptotics for option prices with interest rates effects

no code implementations21 Feb 2024 Dan Pirjol, Lingjiong Zhu

We derive the short-maturity asymptotics for option prices in the local volatility model in a new short-maturity limit $T\to 0$ at fixed $\rho = (r-q) T$, where $r$ is the interest rate and $q$ is the dividend yield.

Intriguing Differences Between Zero-Shot and Systematic Evaluations of Vision-Language Transformer Models

no code implementations13 Feb 2024 Shaeke Salman, Md Montasir Bin Shams, Xiuwen Liu, Lingjiong Zhu

Transformer-based models have dominated natural language processing and other areas in the last few years due to their superior (zero-shot) performance on benchmark datasets.

Language Modelling Zero-Shot Learning

Convergence Analysis for General Probability Flow ODEs of Diffusion Models in Wasserstein Distances

no code implementations31 Jan 2024 Xuefeng Gao, Lingjiong Zhu

Score-based generative modeling with probability flow ordinary differential equations (ODEs) has achieved remarkable success in a variety of applications.

Wasserstein Convergence Guarantees for a General Class of Score-Based Generative Models

no code implementations18 Nov 2023 Xuefeng Gao, Hoang M. Nguyen, Lingjiong Zhu

We find that the experimental results are in good agreement with our theoretical predictions on the iteration complexity, and the models with our newly proposed forward processes can outperform existing models.

Image Generation Unconditional Image Generation

Asymptotics for Short Maturity Asian Options in Jump-Diffusion models with Local Volatility

no code implementations30 Aug 2023 Dan Pirjol, Lingjiong Zhu

We present a study of the short maturity asymptotics for Asian options in a jump-diffusion model with a local volatility component, where the jumps are modeled as a compound Poisson process.

Asymptotics for the Laplace transform of the time integral of the geometric Brownian motion

no code implementations15 Jun 2023 Dan Pirjol, Lingjiong Zhu

We present an asymptotic result for the Laplace transform of the time integral of the geometric Brownian motion $F(\theta, T) = \mathbb{E}[e^{-\theta X_T}]$ with $X_T = \int_0^T e^{\sigma W_s + ( a - \frac12 \sigma^2)s} ds$, which is exact in the limit $\sigma^2 T \to 0$ at fixed $\sigma^2 \theta T^2$ and $aT$.

Cyclic and Randomized Stepsizes Invoke Heavier Tails in SGD than Constant Stepsize

no code implementations10 Feb 2023 Mert Gürbüzbalaban, Yuanhan Hu, Umut Şimşekli, Lingjiong Zhu

Our results bring a new understanding of the benefits of cyclic and randomized stepsizes compared to constant stepsize in terms of the tail behavior.

Scheduling

Algorithmic Stability of Heavy-Tailed SGD with General Loss Functions

no code implementations27 Jan 2023 Anant Raj, Lingjiong Zhu, Mert Gürbüzbalaban, Umut Şimşekli

Very recently, new generalization bounds have been proven, indicating a non-monotonic relationship between the generalization error and heavy tails, which is more pertinent to the reported empirical observations.

Generalization Bounds

Sensitivities of Asian options in the Black-Scholes model

no code implementations16 Jan 2023 Dan Pirjol, Lingjiong Zhu

We propose analytical approximations for the sensitivities (Greeks) of the Asian options in the Black-Scholes model, following from a small maturity/volatility approximation for the option prices which has the exact short maturity limit, obtained using large deviations theory.

A delayed dual risk model

no code implementations16 Jan 2023 Lingjiong Zhu

In this paper, we study a dual risk model with delays in the spirit of Dassios-Zhao.

Penalized Overdamped and Underdamped Langevin Monte Carlo Algorithms for Constrained Sampling

no code implementations29 Nov 2022 Mert Gürbüzbalaban, Yuanhan Hu, Lingjiong Zhu

When $f$ is smooth and gradients are available, we get $\tilde{\mathcal{O}}(d/\varepsilon^{10})$ iteration complexity for PLD to sample the target up to an $\varepsilon$-error where the error is measured in the TV distance and $\tilde{\mathcal{O}}(\cdot)$ hides logarithmic factors.

Algorithmic Stability of Heavy-Tailed Stochastic Gradient Descent on Least Squares

no code implementations2 Jun 2022 Anant Raj, Melih Barsbey, Mert Gürbüzbalaban, Lingjiong Zhu, Umut Şimşekli

Recent studies have shown that heavy tails can emerge in stochastic optimization and that the heaviness of the tails have links to the generalization error.

Stochastic Optimization

Heavy-Tail Phenomenon in Decentralized SGD

no code implementations13 May 2022 Mert Gurbuzbalaban, Yuanhan Hu, Umut Simsekli, Kun Yuan, Lingjiong Zhu

To have a more explicit control on the tail exponent, we then consider the case where the loss at each node is a quadratic, and show that the tail-index can be estimated as a function of the step-size, batch-size, and the topological properties of the network of the computational nodes.

Stochastic Optimization

Fractal Structure and Generalization Properties of Stochastic Optimization Algorithms

no code implementations NeurIPS 2021 Alexander Camuto, George Deligiannidis, Murat A. Erdogdu, Mert Gürbüzbalaban, Umut Şimşekli, Lingjiong Zhu

As our main contribution, we prove that the generalization error of a stochastic optimization algorithm can be bounded based on the `complexity' of the fractal structure that underlies its invariant measure.

Generalization Bounds Learning Theory +1

Convergence Rates of Stochastic Gradient Descent under Infinite Noise Variance

no code implementations NeurIPS 2021 Hongjian Wang, Mert Gürbüzbalaban, Lingjiong Zhu, Umut Şimşekli, Murat A. Erdogdu

In this paper, we provide convergence guarantees for SGD under a state-dependent and heavy-tailed noise with a potentially infinite variance, for a class of strongly convex objectives.

Asymmetric Heavy Tails and Implicit Bias in Gaussian Noise Injections

1 code implementation13 Feb 2021 Alexander Camuto, Xiaoyu Wang, Lingjiong Zhu, Chris Holmes, Mert Gürbüzbalaban, Umut Şimşekli

In this paper we focus on the so-called `implicit effect' of GNIs, which is the effect of the injected noise on the dynamics of SGD.

Decentralized Stochastic Gradient Langevin Dynamics and Hamiltonian Monte Carlo

no code implementations1 Jul 2020 Mert Gürbüzbalaban, Xuefeng Gao, Yuanhan Hu, Lingjiong Zhu

Stochastic gradient Langevin dynamics (SGLD) and stochastic gradient Hamiltonian Monte Carlo (SGHMC) are two popular Markov Chain Monte Carlo (MCMC) algorithms for Bayesian inference that can scale to large datasets, allowing to sample from the posterior distribution of the parameters of a statistical model given the input data and the prior distribution over the model parameters.

Bayesian Inference regression

The Heavy-Tail Phenomenon in SGD

1 code implementation8 Jun 2020 Mert Gurbuzbalaban, Umut Şimşekli, Lingjiong Zhu

We claim that depending on the structure of the Hessian of the loss at the minimum, and the choices of the algorithm parameters $\eta$ and $b$, the SGD iterates will converge to a \emph{heavy-tailed} stationary distribution.

Non-Convex Optimization via Non-Reversible Stochastic Gradient Langevin Dynamics

no code implementations6 Apr 2020 Yuanhan Hu, Xiaoyu Wang, Xuefeng Gao, Mert Gurbuzbalaban, Lingjiong Zhu

In this paper, we study the non reversible Stochastic Gradient Langevin Dynamics (NSGLD) which is based on discretization of the non-reversible Langevin diffusion.

Stochastic Optimization

Asymptotics of the time-discretized log-normal SABR model: The implied volatility surface

no code implementations27 Jan 2020 Dan Pirjol, Lingjiong Zhu

We derive an almost sure limit and a large deviations result for the log-asset price in the limit of large number of time steps.

Robust Distributed Accelerated Stochastic Gradient Methods for Multi-Agent Networks

no code implementations19 Oct 2019 Alireza Fallah, Mert Gurbuzbalaban, Asuman Ozdaglar, Umut Simsekli, Lingjiong Zhu

When gradients do not contain noise, we also prove that distributed accelerated methods can \emph{achieve acceleration}, requiring $\mathcal{O}(\kappa \log(1/\varepsilon))$ gradient evaluations and $\mathcal{O}(\kappa \log(1/\varepsilon))$ communications to converge to the same fixed point with the non-accelerated variant where $\kappa$ is the condition number and $\varepsilon$ is the target accuracy.

Stochastic Optimization

Accelerated Linear Convergence of Stochastic Momentum Methods in Wasserstein Distances

no code implementations22 Jan 2019 Bugra Can, Mert Gurbuzbalaban, Lingjiong Zhu

In the special case of strongly convex quadratic objectives, we can show accelerated linear rates in the $p$-Wasserstein metric for any $p\geq 1$ with improved sensitivity to noise for both AG and HB through a non-asymptotic analysis under some additional assumptions on the noise structure.

Breaking Reversibility Accelerates Langevin Dynamics for Global Non-Convex Optimization

no code implementations19 Dec 2018 Xuefeng Gao, Mert Gurbuzbalaban, Lingjiong Zhu

We study two variants that are based on non-reversible Langevin diffusions: the underdamped Langevin dynamics (ULD) and the Langevin dynamics with a non-symmetric drift (NLD).

Global Convergence of Stochastic Gradient Hamiltonian Monte Carlo for Non-Convex Stochastic Optimization: Non-Asymptotic Performance Bounds and Momentum-Based Acceleration

no code implementations12 Sep 2018 Xuefeng Gao, Mert Gürbüzbalaban, Lingjiong Zhu

We provide finite-time performance bounds for the global convergence of both SGHMC variants for solving stochastic non-convex optimization problems with explicit constants.

Stochastic Optimization

Asymptotic Analysis for Optimal Dividends in a Dual Risk Model

no code implementations13 Jan 2016 Arash Fahim, Lingjiong Zhu

The dual risk model is a popular model in finance and insurance, which is often used to model the wealth process of a venture capital or high tech company.

Optimal Investment in a Dual Risk Model

no code implementations16 Oct 2015 Arash Fahim, Lingjiong Zhu

In this paper, we propose to study the optimal investment strategy on research and development for the dual risk models to minimize the ruin probability of the underlying company.

A State-Dependent Dual Risk Model

no code implementations13 Oct 2015 Lingjiong Zhu

In a dual risk model, the premiums are considered as the costs and the claims are regarded as the profits.

Cannot find the paper you are looking for? You can Submit a new open access paper.