RewardBench: the first evaluation tool for reward models.
-
Updated
Dec 11, 2024 - Python
RewardBench: the first evaluation tool for reward models.
Free and open source code of the https://tournesol.app platform. Meet the community on Discord https://discord.gg/WvcSG55Bf3
Explore concepts like Self-Correct, Self-Refine, Self-Improve, Self-Contradict, Self-Play, and Self-Knowledge, alongside o1-like reasoning elevation🍓 and hallucination alleviation🍄.
The MAGICAL benchmark suite for robust imitation learning (NeurIPS 2020)
This repository contains the source code for our paper: "NaviSTAR: Socially Aware Robot Navigation with Hybrid Spatio-Temporal Graph Transformer and Preference Learning". For more details, please refer to our project website at https://sites.google.com/view/san-navistar.
Python-based GUI to collect Feedback of Chemist in Molecules
Official implementation of Bootstrapping Language Models via DPO Implicit Rewards
Code for the paper "Aligning LLM Agents by Learning Latent Preference from User Edits".
Official code for ICML 2024 paper, "RIME: Robust Preference-based Reinforcement Learning with Noisy Preferences" (ICML 2024 Spotlight)
PyTorch implementations for Offline Preference-Based RL (PbRL) algorithms
Data and models for the paper "Configurable Safety Tuning of Language Models with Synthetic Preference Data"
Code for "Monocular Depth Estimation via Listwise Ranking using the Plackett-Luce Model" as published at CVPR 2021.
This repository contains the source code for our paper: "Feedback-efficient Active Preference Learning for Socially Aware Robot Navigation", accepted to IROS-2022. For more details, please refer to our project website at https://sites.google.com/view/san-fapl.
Preference Learning with Gaussian Processes and Bayesian Optimization
Java framework for Preference Learning
A paper under AAAI-20 review
[P]reference and [R]ule [L]earning algorithm implementation for Python 3 (https://arxiv.org/abs/1812.07895)
Code for the project: "Analysis of Recommendation-systems based on User Preferences".
Code for the paper "Reward Design for Justifiable Sequential Decision-Making"; ICLR 2024
(AISTATS 2024) "Looping in the Human: Collaborative and Explainable Bayesian Optimization"
Add a description, image, and links to the preference-learning topic page so that developers can more easily learn about it.
To associate your repository with the preference-learning topic, visit your repo's landing page and select "manage topics."