Scale Events
+00:00 GMT
Sign in or Join the community to continue

Aligning LLMs with Representation Learning

Posted Invalid date | Views 183
Share
speakers
avatar
Sean Hendryx
Head of Generative AI ML @ Scale AI

Sean Hendryx leads frontier data research and applied ML for generative AI at Scale. He has been developing models for collaborative human-AI systems for over six years. He is currently focused on researching systems that most effectively leverage human supervision for the advancement of frontier models and has publications on alignment, frontier evaluations, test time compute scaling, meta-learning, and online learning.

+ Read More
avatar
Vaskar Nath
ML Research Engineer @ Scale AI

Vaskar Nath is a Machine Learning Research Engineer at Scale AI, where he works on advancing the capabilities of language models. His research focuses on improving reward modeling, enhancing tool usage, and strengthening reasoning abilities. Vaskar holds a bachelor’s degree from the University of Toronto, where he double-majored in computer science and mathematics with a minor in statistics. Prior to joining Scale, he gained experience through internships at Meta, Nuro, AWS, and Intel.

+ Read More
avatar
Jessica Zhu
Tech Lead, Generative AI Growth @ Scale AI

Jessica Zhu is the tech lead of Scale's Generative AI Growth team. She has spent the past few years working on various products at the forefront of AI, including document processing, computer vision, and LLM infrastructure. In her free time she enjoys traveling, boxing, and playing board games.

+ Read More
SUMMARY

Join Scale AI researchers as they present their NeurIPS 2024 main track paper, Learning Goal-Conditioned Representations for Language Reward Models, introducing a novel approach to improving LLM alignment. This technical session will explore how goal-conditioned representations can enhance reward modeling and significantly reduce computational costs. Through a detailed examination of the methodology and results, we'll demonstrate how this approach achieves substantial improvements in both model performance and decoding efficiency. The presentation will be followed by an in-depth discussion of practical implementation considerations and future research directions.

+ Read More

Watch More

0:42
OpenAI’s InstructGPT: Aligning Language Models With Human Intent
Posted Sep 09, 2022 | Views 44.6K
# Large Language Models (LLMs)
# Natural Language Processing (NLP)
A Machine Learning Infrastructure Playbook with Lambda
Posted Oct 06, 2021 | Views 2.6K
# TransformX 2021
# Breakout Session