Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
-
Updated
Apr 20, 2024 - Python
Safe RLHF: Constrained Value Alignment via Safe Reinforcement Learning from Human Feedback
💨 A real time messaging system to build a scalable in-app notifications, multiplayer games, chat apps in web and mobile apps.
Building Ubuntu 18 Bionic vagrant boxes using packer
BeaverTails is a collection of datasets designed to facilitate research on safety alignment in large language models (LLMs).
Site em desenvolvimento
Client-side utility to maintain an up-to-date hyperlocal context graph by consuming the real-time data stream from Pareto Anywhere APIs. We believe in an open Internet of Things.
Add a description, image, and links to the beaver topic page so that developers can more easily learn about it.
To associate your repository with the beaver topic, visit your repo's landing page and select "manage topics."