Friday, June 6

Join our daily and weekly newsletters to obtain the latest updates and exclusive content on the coverage of the industry leader. Get more information


Companies should know if the models that drive their applications and agents work on real -life scenarios. This type of evaluation can sometimes be complex because it is difficult to predict specific scenarios. A renewed version of the rewards point of reference seeks to give organizations a better idea or real life performance of a model.

The Ales Institute of AI (AI2) launched rewards 2, an updated version of its reward model reference, rewards, which according to them provides a more holistic vision of model performance and evaluates aligning with the objectives and standards of a company.

AI2 built rewards with classification tasks that measure correlations through the inference of computation and subsequent training. Rewarding the bank is mainly reward models (RM), which can act as judges and evaluate the results of LLM. RMS assigns a score or a “reward” that guides the learning of reinforcement with human comments (RHLF)

Nathan Lambert, a AI2 senior research scientist, told Venturebeat that the first rewards bank worked as planned when it was launched. Even so, the model environment evolved quickly, and their reference points should also.

“As the reward models became more advanced and the most nuanced cases, we quickly recognized with the community that the first version did not completely capture the complexity of the human preferences of the real world,” said Hey.

Lambert added that with reportsbench 2, “we set out to improve both the amplitude and the depth of the evaluation, incorporating more diverse and challenging indications and refining the methodology to better reflect how humans really judge the results of AI in practice.” He said that the second version uses invisible human indications, has a more challenging score configuration and new domains.

Use of evaluations for models that evaluate

While the reward models prove how well the models work, it is also important that the RMS align with the company’s values; Otherwise, the adjustment and reinforcement learning process can reinforce bad behavior, such as hallucinations, reduce generalization and obtain too high harmful responses.

RepensHbench 2 covers six different domains: reality, precise instruction, mathematics, security, focus and ties.

“Companies must use repursbench 2 in two different ways, depending on their application. If RLHF are doing in themselves, they should adopt the best practices and data sets of the leading models that need pipelines of Bercose pipes models that reflect the model they try to train with RL) for the filtration of inference time data, reward 2 has shown that they can select the best model for their domain and see their domain The correlated performance, “Lammert,” he says.

Lambert said that the reference points such as rewards sacrifice users a way to evaluate the models they are choosing based on the “dimensions that care most, instead of trusting a narrow score of unique size.” He said that the idea of ​​performance, that many evaluation methods claim to evaluate, is very subjective because a good response of a model depends largely on the context and the user’s objectives. At the same time, human preferences are killed very.

AI 2 launched the first version of rewards in March 2024. At that time, the company said it was the first point of reference and the classification for rewards models. Since then, several methods for comparative evaluation and RM improvement have emerged. The Meta Fair researchers went out with Rewordbench. Deepseek launched a new technique called Tuning of Critics’ Criticism for Emarter and Scalable RM.Super excited that our second evaluation of the reward model is out. It is substantially harder, much cleaner and well correlated with the sampling PPO/bon downstream.

How the models performed

Since RepensHbench 2 is an updated version of rewards, AI2 tested both existing and newly trained models to see if they continue to classify high. These included a variety of models, such as Gemini, Claude, GPT-4.1 and call-3.1 versions, along with data sets and models such as Qwen, Skywork and its own Tulu.

The company discovered that the largest reward models work better at the point of reference because its base models are stronger. In general, the strongest performance models are the variars or the instructions of call-3.1. In terms of approach and security, Skywork data “are particularly useful,” and Tulu had a good performance in the invoice.

AI2 SAID THAT WHILE they BELIEVE REWARDBENCH 2 This Works This Works Works Works Works Works Works Works Works Works Works Works This Work Works This Work Works Works Works.

 
Exit mobile version