This story on HackerNoon has a decentralized backup on Sia.
Transaction ID: Jtpn2O4mFEz3HHcfuNnZ2otKbc6_-M_Y7KkaYxViYWY
Cover

Stable Nonconvex-Nonconcave Training via Linear Interpolation: Approximating the resolvent

Written by @interpolation | Published on 2024/3/7

TL;DR
This paper presents a theoretical analysis of linear interpolation as a principled method for stabilizing (large-scale) neural network training.

This paper is available on arxiv under CC 4.0 license.

Authors:

(1) Thomas Pethick, EPFL (LIONS) thomas.pethick@epfl.ch;

(2) Wanyun Xie, EPFL (LIONS) wanyun.xie@epfl.ch;

(3) Volkan Cevher, EPFL (LIONS) volkan.cevher@epfl.ch.

5 Approximating the resolvent

This can be approximated with a fixed point iteration of

which is a contraction for small enough γ since F is Lipschitz continuous. It follows from Banach’s fixed-point theorem Banach (1922) that the sequence converges linearly. We formalize this in the following theorem, which additionally applies when only stochastic feedback is available.

The resulting update in Algorithm 1 is identical to GDA but crucially always steps from z. We use this as a subroutine in RAPP to get convergence under a cohypomonotone operator while only suffering a logarithmic factor in the rate.

[story continues]


Written by
@interpolation
#1 Publication focused exclusively on Interpolation, ie determining value from the existing values in a given data set.

Topics and
tags
linear-interpolation|nonexpansive-operators|rapp|cohypomonotone-problems|lookahead-algorithms|rapp-and-lookahead|training-gans|nonmonotone-class
This story on HackerNoon has a decentralized backup on Sia.
Transaction ID: Jtpn2O4mFEz3HHcfuNnZ2otKbc6_-M_Y7KkaYxViYWY