LiyuanLucasLiu / RAdam-Tensorflow

Simple Tensorflow implementation of "On The Variance Of The Adaptive Learning Rate And Beyond"

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

RAdam-Tensorflow

On the Variance of the Adaptive Learning Rate and Beyond

Usage

from RAdam import RAdamOptimizer

train_op = RAdamOptimizer(learning_rate=0.001, beta1=0.9, beta2=0.999, weight_decay=0.0).minimize(loss)

Algorithm

Result

result

Author

Junho Kim

About

Simple Tensorflow implementation of "On The Variance Of The Adaptive Learning Rate And Beyond"

License:MIT License


Languages

Language:Python 100.0%