-
awesome-RLHF
A curated list of reinforcement learning with human feedback resources (continually updated)
-
hh-rlhf
Human preference data for "Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback"
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
Recently, OpenDILab made a paper collection about Reinforcement Learning with Human Feedback (RLHF) and it has been open-sourced on GitHub. This repository is dedicated to helping researchers to collect the latest papers on RLHF, so that they can get to know this area better and more easily.
Title: Training a Helpful and Harmless Assistant with Reinforcement Learning from Human Feedback
Related posts
-
Awesome RLHF (RL with Human Feedback) This is a collection of research papers for Reinforcement Learning with Human Feedback (RLHF). And the repository will be continuously updated to track the frontier of RLHF. Welcome to follow and star! https://github.com/opendilab/awesome-RLHF
-
OpenDILab Awesome Paper Collection: RL with Human Feedback (1)
-
A collection of research papers for Reinforcement Learning with Human Feedback (RLHF)
-
Deep Reinforcement Learning: Zero to Hero
-
How do I change the maximum number of steps for training