PKU-Alignment

PKU-Alignment

Geek Repo

Loves Sharing and Open-Source, Making AI Safer.

Location:China

Github PK Tool:Github PK Tool

PKU-Alignment's repositories

safe-rlhf

Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback

Language:PythonLicense:Apache-2.0Stargazers:1180Issues:17Issues:82

omnisafe

OmniSafe is an infrastructural framework for accelerating SafeRL research.

Language:PythonLicense:Apache-2.0Stargazers:861Issues:38Issues:90

safety-gymnasium

NeurIPS 2023: Safety-Gymnasium: A Unified Safe Reinforcement Learning Benchmark

Language:PythonLicense:Apache-2.0Stargazers:326Issues:9Issues:23

Safe-Policy-Optimization

NeurIPS 2023: Safe Policy Optimization: A benchmark repository for safe reinforcement learning algorithms

Language:PythonLicense:Apache-2.0Stargazers:295Issues:6Issues:10

AlignmentSurvey

AI Alignment: A Comprehensive Survey

beavertails

BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).

Language:MakefileLicense:Apache-2.0Stargazers:81Issues:5Issues:5

ProAgent

ProAgent: Building Proactive Cooperative Agents with Large Language Models

Language:JavaScriptLicense:MITStargazers:37Issues:9Issues:1

SafeDreamer

ICLR 2024: SafeDreamer: Safe Reinforcement Learning with World Models

Language:PythonLicense:Apache-2.0Stargazers:28Issues:3Issues:2

ReDMan

ReDMan is an open-source simulation platform that provides a standardized implementation of safe RL algorithms for Reliable Dexterous Manipulation.

Language:PythonLicense:Apache-2.0Stargazers:15Issues:3Issues:0
Stargazers:0Issues:2Issues:0