Skip to content

shoelim/mpgd

Repository files navigation

MPGD: Chaotic Regularization and Heavy-Tailed Limits for Deterministic Gradient Descent

Multiscale perturbed gradient descent (MPGD) is an optimization framework where the gradient descent recursion is augmented with chaotic perturbations that evolve via an independent dynamical system.

This repository contains the implementation of MPGD and the code used for the results in the paper. Please refer to the paper for an introduction to the optimization tasks and other details.

Requirements

  • python 3
  • pyTorch 1.9.*
  • hydra 1.* (via pip install hydra-core --upgrade)
  • sklearn
  • numpy
  • scipy
  • pandas
  • math
  • statistics

Instructions and Usage

Minimizing the widening valley loss

python minimizing_widening_valley_loss.py

See also the Google Colab version here

Airfoil Self-Noise regression

python train.py

Electrocardiogram (ECG) classification

python ecg_classification_mlps.py

See also the Google Colab version here

CIFAR-10 classification

Scripts for training runs can be found in train.sh. Please check and specify the parameters appropriately before running.

About

MPGD: Chaotic Regularization and Heavy-Tailed Limits for Deterministic Gradient Descent

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published