PKU-Alignment

Results 10 repositories owned by PKU-Alignment
trafficstars

omnisafe

911
Stars
130
Forks
Watchers

JMLR: OmniSafe is an infrastructural framework for accelerating SafeRL research.

Safe-Policy-Optimization

296
Stars
39
Forks
Watchers

NeurIPS 2023: Safe Policy Optimization: A benchmark repository for safe reinforcement learning algorithms

AlignmentSurvey

116
Stars
0
Forks
Watchers

AI Alignment: A Comprehensive Survey

safe-rlhf

1.3k
Stars
119
Forks
Watchers

Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback

safety-gymnasium

383
Stars
53
Forks
Watchers

NeurIPS 2023: Safety-Gymnasium: A Unified Safe Reinforcement Learning Benchmark

beavertails

105
Stars
3
Forks
Watchers

BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).

ProAgent

52
Stars
4
Forks
Watchers

ProAgent: Building Proactive Cooperative Agents with Large Language Models

SafeDreamer

43
Stars
6
Forks
Watchers

ICLR 2024: SafeDreamer: Safe Reinforcement Learning with World Models

safe-sora

33
Stars
5
Forks
33
Watchers

SafeSora is a human preference dataset designed to support safety alignment research in the text-to-video generation field, aiming to enhance the helpfulness and harmlessness of Large Vision Models (L...

align-anything

104
Stars
28
Forks
Watchers

Align Anything: Training All-modality Model with Feedback