Research

Ridge Rider: Finding Diverse Solutions by Following Eigenvectors of the Hessian

November 25, 2020

Abstract

Over the last decade, a single algorithm has changed many facets of our lives - Stochastic Gradient Descent (SGD). In the era of ever decreasing loss functions, SGD and its various offspring have become the go-to optimization tool in machine learning and are a key component of the success of deep neural networks (DNNs). While SGD is guaranteed to converge to a local optimum (under loose assumptions), in some cases it may matter which local optimum is found, and this is often context dependent. Examples frequently arise in machine learning, from shape- versus texture-features to ensemble methods and zero-shot coordination. In these settings, there are desired solutions which SGD on ‘standard’ loss functions will not find, since it instead converges to the ‘easy’ solutions. In this paper, we present a different approach. Rather than following the gradient, which corresponds to a locally greedy direction, we instead follow the eigenvectors of the Hessian. By iteratively following and branching amongst the ridges, we effectively span the loss surface to find qualitatively different solutions. We show both theoretically and experimentally that our method, called Ridge Rider (RR), offers a promising direction for a variety of challenging problems.

Download the Paper

AUTHORS

Written by

Jack Parker-Holder

Luke Metz

Cinjon Resnick

Hengyuan Hu

Adam Lerer

Alistair Letcher

Alex Peysakhovich

Aldo Pacchiano

Jakob Foerster

Research Topics

Reinforcement Learning

Related Publications

November 30, 2020

Theory

Ranking & Recommendations

On ranking via sorting by estimated expected utility

Nicolas Usunier, Clément Calauzènes

November 30, 2020

February 01, 2021

Ranking & Recommendations

Anytime Inference with Distilled Hierarchical Neural Ensembles

Adria Ruiz, Jakob Verbeek

February 01, 2021

November 01, 2018

Ranking & Recommendations

Horizon: Facebook's Open Source Applied Reinforcement Learning Platform | Facebook AI Research

Jason Gauci, Edoardo Conti, Yitao Liang, Kittipat Virochsiri, Yuchen He, Zachary Kaden, Vivek Narayanan, Xiaohui Ye

November 01, 2018

December 03, 2018

Ranking & Recommendations

NLP

Training with Low-precision Embedding Tables | Facebook AI Research

Jian Zhang, Jiyan Yang, Hector Yuen

December 03, 2018

May 03, 2019

Ranking & Recommendations

Multi-Perspective Relevance Matching with Hierarchical ConvNets for Social Media Search | Facebook AI Research

Jinfeng Rao, Wei Yang, Yuhao Zhang, Ferhan Ture, Jimmy Lin

May 03, 2019

Help Us Pioneer The Future of AI

We share our open source frameworks, tools, libraries, and models for everything from research exploration to large-scale production deployment.