End-to-end Algorithm Synthesis with Recurrent Networks: Logical Extrapolation Without Overthinking

02/11/2022
by   Arpit Bansal, et al.
4

Machine learning systems perform well on pattern matching tasks, but their ability to perform algorithmic or logical reasoning is not well understood. One important reasoning capability is logical extrapolation, in which models trained only on small/simple reasoning problems can synthesize complex algorithms that scale up to large/complex problems at test time. Logical extrapolation can be achieved through recurrent systems, which can be iterated many times to solve difficult reasoning problems. We observe that this approach fails to scale to highly complex problems because behavior degenerates when many iterations are applied – an issue we refer to as "overthinking." We propose a recall architecture that keeps an explicit copy of the problem instance in memory so that it cannot be forgotten. We also employ a progressive training routine that prevents the model from learning behaviors that are specific to iteration number and instead pushes it to learn behaviors that can be repeated indefinitely. These innovations prevent the overthinking problem, and enable recurrent systems to solve extremely hard logical extrapolation tasks, some requiring over 100K convolutional layers, without overthinking.

READ FULL TEXT

page 1

page 14

page 15

research
02/22/2021

Thinking Deeply with Recurrence: Generalizing from Easy to Hard Sequential Reasoning Problems

Deep neural networks are powerful machines for visual pattern recognitio...
research
05/29/2019

SATNet: Bridging deep learning and logical reasoning using a differentiable satisfiability solver

Integrating logical reasoning within deep learning architectures has bee...
research
06/08/2021

Can You Learn an Algorithm? Generalizing from Easy to Hard Problems with Recurrent Networks

Deep neural networks are powerful machines for visual pattern recognitio...
research
02/09/2011

From Machine Learning to Machine Reasoning

A plausible definition of "reasoning" could be "algebraically manipulati...
research
05/23/2022

On the Paradox of Learning to Reason from Data

Logical reasoning is needed in a wide range of NLP tasks. Can a BERT mod...
research
11/14/2022

Logical Tasks for Measuring Extrapolation and Rule Comprehension

Logical reasoning is essential in a variety of human activities. A repre...
research
06/29/2020

Switchblade – a Neural Network for Hard 2D Tasks

Convolutional neural networks have become the main tools for processing ...

Please sign up or login with your details

Forgot password? Click here to reset