Stochastic Optimization

282 papers with code • 12 benchmarks • 11 datasets

Stochastic Optimization is the task of optimizing certain objective functional by generating and using stochastic random variables. Usually the Stochastic Optimization is an iterative process of generating random variables that progressively finds out the minima or the maxima of the objective functional. Stochastic Optimization is usually applied in the non-convex functional spaces where the usual deterministic optimization such as linear or quadratic programming or their variants cannot be used.

Source: ASOC: An Adaptive Parameter-free Stochastic Optimization Techinique for Continuous Variables

Libraries

Use these libraries to find Stochastic Optimization models and implementations

Why Do We Need Weight Decay in Modern Deep Learning?

tml-epfl/why-weight-decay 6 Oct 2023

In this work, we highlight that the role of weight decay in modern deep learning is different from its regularization effect studied in classical learning theory.

35
06 Oct 2023

Smoothing Methods for Automatic Differentiation Across Conditional Branches

philipp-andelfinger/discograd 5 Oct 2023

We detail the effects of the approximations made for tractability in SI and propose a novel Monte Carlo estimator that avoids the underlying assumptions by estimating the smoothed programs' gradients through a combination of AD and sampling.

7
05 Oct 2023

Quasi-Monte Carlo for 3D Sliced Wasserstein

khainb/quasi-sw 21 Sep 2023

Monte Carlo (MC) integration has been employed as the standard approximation method for the Sliced Wasserstein (SW) distance, whose analytical expression involves an intractable expectation.

5
21 Sep 2023

Landscape-Sketch-Step: An AI/ML-Based Metaheuristic for Surrogate Optimization Problems

rafael-a-monteiro-math/landscape_sketch_and_step 14 Sep 2023

In this paper, we introduce a new heuristics for global optimization in scenarios where extensive evaluations of the cost function are expensive, inaccessible, or even prohibitive.

0
14 Sep 2023

A Distributed Data-Parallel PyTorch Implementation of the Distributed Shampoo Optimizer for Training Neural Networks At-Scale

facebookresearch/optimizers 12 Sep 2023

It constructs a block-diagonal preconditioner where each block consists of a coarse Kronecker product approximation to full-matrix AdaGrad for each parameter of the neural network.

199
12 Sep 2023

PROMISE: Preconditioned Stochastic Optimization Methods by Incorporating Scalable Curvature Estimates

udellgroup/promise 5 Sep 2023

This paper introduces PROMISE ($\textbf{Pr}$econditioned Stochastic $\textbf{O}$ptimization $\textbf{M}$ethods by $\textbf{I}$ncorporating $\textbf{S}$calable Curvature $\textbf{E}$stimates), a suite of sketching-based preconditioned stochastic gradient algorithms for solving large-scale convex optimization problems arising in machine learning.

1
05 Sep 2023

Likelihood-based inference and forecasting for trawl processes: a stochastic optimization approach

danleonte/ambit_stochastics 30 Aug 2023

In this paper, we develop the first likelihood-based methodology for the inference of real-valued trawl processes and introduce novel deterministic and probabilistic forecasting methods.

3
30 Aug 2023

Integrating LLMs and Decision Transformers for Language Grounded Generative Quality-Diversity

salehiac/languagegroundedqd 25 Aug 2023

Quality-Diversity is a branch of stochastic optimization that is often applied to problems from the Reinforcement Learning and control domains in order to construct repertoires of well-performing policies/skills that exhibit diversity with respect to a behavior space.

6
25 Aug 2023

Serverless Federated AUPRC Optimization for Multi-Party Collaborative Imbalanced Data Mining

xidongwu/d-auprc 6 Aug 2023

To address the above challenge, we study the serverless multi-party collaborative AUPRC maximization problem since serverless multi-party collaborative training can cut down the communications cost by avoiding the server node bottleneck, and reformulate it as a conditional stochastic optimization problem in a serverless multi-party collaborative learning setting and propose a new ServerLess biAsed sTochastic gradiEnt (SLATE) algorithm to directly optimize the AUPRC.

0
06 Aug 2023

A stochastic optimization approach to train non-linear neural networks with a higher-order variation regularization

oknakfm/hovr 4 Aug 2023

While the $(k, q)$-VR terms applied to general parametric models are computationally intractable due to the integration, this study provides a stochastic optimization algorithm, that can efficiently train general models with the $(k, q)$-VR without conducting explicit numerical integration.

1
04 Aug 2023