POTION Collection These are the flagship POTION models. Load them and use them with model2vec (https://github.com/MinishLab/model2vec) or sentence-transformers β’ 5 items β’ Updated Feb 3 β’ 10
REINFORCE++: A Simple and Efficient Approach for Aligning Large Language Models Paper β’ 2501.03262 β’ Published Jan 4 β’ 92
Reward Bench Collection Datasets, spaces, and models for the reward model benchmark! β’ 5 items β’ Updated about 1 hour ago β’ 9
view article Article Accelerated Inference with Optimum and Transformers Pipelines May 10, 2022 β’ 2