There are 1 repository under sgd topic.
LSTM and QRNN Language Model Toolkit for PyTorch
NFNets and Adaptive Gradient Clipping for SGD implemented in PyTorch. Find explanation at tourdeml.github.io/blog/
MATLAB/Octave library for stochastic optimization algorithms: Version 1.0.20
Machine learning algorithms in Dart programming language
This repository contains the results for the paper: "Descending through a Crowded Valley - Benchmarking Deep Learning Optimizers"
Keras/TF implementation of AdamW, SGDW, NadamW, Warm Restarts, and Learning Rate multipliers
A Deep Learning and preprocessing framework in Rust with support for CPU and GPU.
A tour of different optimization algorithms in PyTorch.
Rethinking the Smaller-Norm-Less-Informative Assumption in Channel Pruning of Convolution Layers https://arxiv.org/abs/1802.00124
Java based sample code for developing on Android. The demos in this repository are stored on separate branches. To navigate to a demo, please click branches.
PyTorch implementation of Federated Learning algorithms FedSGD, FedAvg, FedAvgM, FedIR, FedVC, FedProx and standard SGD, applied to visual classification. Client distributions are synthesized with arbitrary non-identicalness and imbalance (Dirichlet priors). Client systems can be arbitrarily heterogeneous. Several mobile-friendly models are provided
Riemannian stochastic optimization algorithms: Version 1.0.3
Unofficial implementation of Switching from Adam to SGD optimization in PyTorch.
Distributed Learning by Pair-Wise Averaging
A C++ toolkit for Convex Optimization (Logistic Loss, SVM, SVR, Least Squares etc.), Convex Optimization algorithms (LBFGS, TRON, SGD, AdsGrad, CG, Nesterov etc.) and Classifiers/Regressors (Logistic Regression, SVMs, Least Squares Regression etc.)
R/Rcpp implementation of the 'Follow-the-Regularized-Leader' algorithm
Implementation of key concepts of neuralnetwork via numpy
Code for paper "On Sampling Strategies for Neural Network-based Collaborative Filtering"
Stochastic gradient descent from scratch for linear regression
Amortized version of the differentially private SGD algorithm published in "Deep Learning with Differential Privacy" by Abadi et al. Enforces privacy by clipping and sanitising the gradients with Gaussian noise during training.
vector quantization for stochastic gradient descent.
SGD with large step sizes learns sparse features [ICML 2023]
Notes & Code to go over "Grokking Deep Learning" Book by Andrew Trask
Distributed Momentum for Byzantine-resilient Stochastic Gradient Descent (ICLR 2021)
implementation of factorization machine, support classification.
Adaptive Reinforcement Learning of curious AI basketball agents
A sentiment classifier on mixed language (and mixed script) reviews in Tamil, Malayalam and English