-
awesome-RLHF
A curated list of reinforcement learning with human feedback resources (continually updated)
-
visual-chatgpt
Discontinued Official repo for the paper: Visual ChatGPT: Talking, Drawing and Editing with Visual Foundation Models [Moved to: https://github.com/microsoft/TaskMatrix]
-
InfluxDB
Power Real-Time Data Analytics at Scale. Get real-time insights from all types of time series data with InfluxDB. Ingest, query, and analyze billions of data points in real-time with unbounded cardinality.
Here we’re gonna introduce a new repository open-sourced by OpenDILab. Recently, OpenDILab made a paper collection about Reinforcement Learning with Human Feedback (RLHF) and it has been open-sourced on GitHub. This repository is dedicated to helping researchers to collect the latest papers on RLHF, so that they can get to know this area better and more easily. About RLHF Reinforcement Learning with Human Feedback (RLHF) is an extended branch of Reinforcement Learning (RL) that allows the RLHF family of methods to incorporate human feedback into the training process by using this feedback to construct By using this feedback to build a reward model neural network that provides reward signals to help RL intelligences learn, human needs, preferences, and perceptions can be more naturally communicated to the intelligence in an interactive learning manner, aligning the optimization goals between humans and artificial intelligence to produce systems that behave in a manner consistent with human values. Reinforcement Learning with Human Feedback (RLHF) is an extended branch of Reinforcement Learning. When the optimization goal is abstract and it's very difficult to define the specific reward function, RLHF can help to put human feedback into the training process. This feedback can be constructed into a reward neural network model so that RL agents can learn from the given reward signal and naturally convey human needs, preference and attitude to agents through interactive learning.
Found relevant code at https://github.com/microsoft/visual-chatgpt + all code implementations here
Related posts
-
OpenDILab Awesome Paper Collection: RL with Human Feedback (3)
-
Awesome RLHF (RL with Human Feedback) This is a collection of research papers for Reinforcement Learning with Human Feedback (RLHF). And the repository will be continuously updated to track the frontier of RLHF. Welcome to follow and star! https://github.com/opendilab/awesome-RLHF
-
A collection of research papers for Reinforcement Learning with Human Feedback (RLHF)
-
How do I change the maximum number of steps for training
-
[P] Introducing PPO and Rainbow DQN to our super fast evolutionary HPO reinforcement learning framework