RLHFlow
Code for the Workflow of Reinforcement Learning from Human Feedback (RLHF)
- 20 followers
- United States of America
- https://huggingface.co/RLHFlow
- @RLHFlow
- rlhflow.ai@gmail.com
Popular repositories
-
RLHF-Reward-Modeling
RLHF-Reward-Modeling PublicA recipe to train reward models for RLHF.
-
-
Directional-Preference-Alignment
Directional-Preference-Alignment PublicDirectional Preference Alignment
-
-
Repositories
Showing 5 of 5 repositories