[ICML 2024] Code for the paper "Confronting Reward Overoptimization for Diffusion Models: A Perspective of Inductive and Primacy Biases"
-
Updated
May 20, 2024 - Python
[ICML 2024] Code for the paper "Confronting Reward Overoptimization for Diffusion Models: A Perspective of Inductive and Primacy Biases"
Aligning LLM Agents by Learning Latent Preference from User Edits
Reinforcement Learning from Human Feedback with 🤗 TRL
A curated list of reinforcement learning with human feedback resources[awesome-RLHF-Turkish] (continually updated)
[ NeurIPS 2023 ] Official Codebase for "Aligning Synthetic Medical Images with Clinical Knowledge using Human Feedback"
The Prism Alignment Project
BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).
[CVPR 2024] Code for the paper "Using Human Feedback to Fine-tune Diffusion Models without Any Reward Model"
Implementation of Reinforcement Learning from Human Feedback (RLHF)
The ParroT framework to enhance and regulate the Translation Abilities during Chat based on open-sourced LLMs (e.g., LLaMA-7b, Bloomz-7b1-mt) and human written translation and evaluation data.
Product analytics for AI Assistants
Let's build better datasets, together!
Open-source pre-training implementation of Google's LaMDA in PyTorch. Adding RLHF similar to ChatGPT.
A curated list of reinforcement learning with human feedback resources (continually updated)
Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM
Add a description, image, and links to the human-feedback topic page so that developers can more easily learn about it.
To associate your repository with the human-feedback topic, visit your repo's landing page and select "manage topics."