MT Bench
Compare model answers to questions
A curated collection of datasets, models, Spaces, and papers on Reinforcement Learning from Human Feedback (RLHF).
Compare model answers to questions
Note A multi-turn evaluation benchmark for chatbots which uses GPT-4 as a judge to evaluate the quality of responses.
Note A high quality blend of human and synthetic datasets focused on dialogues and reasoning abilities. A good source for SFT and PPO.
Note The first series of open access models trained at scale using RLHF. Based on https://huggingface.co/papers/2307.09288
Note The first series of open access models trained at scale using RLHF. Based on https://huggingface.co/papers/2307.09288
Note An RLHF tuned version of Falcon 40B
Note A dataset of dialogues between human annotators and a 52B parameter language model from Anthropic. Contains "helpfulness" and "harmlessness" subsets that can be used for training reward models. Basis for https://huggingface.co/papers/2112.00861 and https://huggingface.co/papers/2204.05862