Great Power, Great Responsibility: Recommendations for Reducing Energy for Training Language Models

05/19/2022
by   Joseph McDonald, et al.
0

The energy requirements of current natural language processing models continue to grow at a rapid, unsustainable pace. Recent works highlighting this problem conclude there is an urgent need for methods that reduce the energy needs of NLP and machine learning more broadly. In this article, we investigate techniques that can be used to reduce the energy consumption of common NLP applications. In particular, we focus on techniques to measure energy usage and different hardware and datacenter-oriented settings that can be tuned to reduce energy consumption for training and inference for language models. We characterize the impact of these settings on metrics such as computational performance and energy consumption through experiments conducted on a high performance computing system as well as popular cloud computing platforms. These techniques can lead to significant reduction in energy consumption when training language models or their use for inference. For example, power-capping, which limits the maximum power a GPU can consume, can enable a 15% decrease in energy usage with marginal increase in overall computation time when training a transformer-based language model.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/30/2021

EnergyVis: Interactively Tracking and Exploring Energy Consumption for ML Models

The advent of larger machine learning (ML) models have improved state-of...
research
06/10/2022

Measuring the Carbon Intensity of AI in Cloud Instances

By providing unprecedented access to computational resources, cloud comp...
research
08/20/2023

A Human-on-the-Loop Optimization Autoformalism Approach for Sustainability

This paper outlines a natural conversational approach to solving persona...
research
11/02/2018

Progress and Tradeoffs in Neural Language Models

In recent years, we have witnessed a dramatic shift towards techniques d...
research
05/01/2023

Performance and Energy Consumption of Parallel Machine Learning Algorithms

Machine learning models have achieved remarkable success in various real...
research
01/28/2022

Benchmarking Resource Usage for Efficient Distributed Deep Learning

Deep learning (DL) workflows demand an ever-increasing budget of compute...
research
06/02/2021

IrEne: Interpretable Energy Prediction for Transformers

Existing software-based energy measurements of NLP models are not accura...

Please sign up or login with your details

Forgot password? Click here to reset