Miao Lu's starred repositories
Diff4RLSurvey
This repository contains a collection of resources and papers on Diffusion Models for RL, accompanying the paper "Diffusion Models for Reinforcement Learning: A Survey"
direct-preference-optimization
Reference implementation for DPO (Direct Preference Optimization)
alignment-handbook
Robust recipes to align language models with human and AI preferences
decision-pretrained-transformer
Implemention of the Decision-Pretrained Transformer (DPT) from the paper Supervised Pretraining Can Learn In-Context Reinforcement Learning.
awesome-RLHF
A curated list of reinforcement learning with human feedback resources (continually updated)
awesome-causality-data
A data index for learning causality.
NeuralCausalModels
Neural Causal Model (NCM) implementation by the authors of The Causal Neural Connection.
fullbatchtraining
Training vision models with full-batch gradient descent and regularization
RL-for-Markov-Exchange-Economy
Codes for the ICML 2022 accepted paper: *Welfare Maximization in Competitive Equilibrium: Reinforcement Learning for Markov Exchange Economy*.
learn2branch-ecole
Reimplementation of "Exact Combinatorial Optimization with Graph Convolutional Neural Networks" (NeurIPS 2019)
cs-self-learning
计算机自学指南
neural-tangents
Fast and Easy Infinite Neural Networks in Python