numpywren: serverless linear algebra

10/23/2018
by   Vaishaal Shankar, et al.
0

Linear algebra operations are widely used in scientific computing and machine learning applications. However, it is challenging for scientists and data analysts to run linear algebra at scales beyond a single machine. Traditional approaches either require access to supercomputing clusters, or impose configuration and cluster management challenges. In this paper we show how the disaggregation of storage and compute resources in so-called "serverless" environments, combined with compute-intensive workload characteristics, can be exploited to achieve elastic scalability and ease of management. We present numpywren, a system for linear algebra built on a serverless architecture. We also introduce LAmbdaPACK, a domain-specific language designed to implement highly parallel linear algebra algorithms in a serverless setting. We show that, for certain linear algebra algorithms such as matrix multiply, singular value decomposition, and Cholesky decomposition, numpywren's performance (completion time) is within 33 efficiency (total CPU-hours) is up to 240 providing an easier to use interface and better fault tolerance. At the same time, we show that the inability of serverless runtimes to exploit locality across the cores in a machine fundamentally limits their network efficiency, which limits performance on other algorithms such as QR factorization. This highlights how cloud providers could better support these types of computations through small changes in their infrastructure.

READ FULL TEXT

page 1

page 2

page 5

page 6

page 8

page 9

page 11

page 13

research
06/03/2018

Alchemist: An Apache Spark <=> MPI Interface

The Apache Spark framework for distributed computation is popular in the...
research
01/13/2023

A hybrid probabilistic domain decomposition algorithm suited for very large-scale elliptic PDEs

State of the art domain decomposition algorithms for large-scale boundar...
research
02/20/2022

Benchmarking the Linear Algebra Awareness of TensorFlow and PyTorch

Linear algebra operations, which are ubiquitous in machine learning, for...
research
03/30/2023

Matrix diagonalization and singular value decomposition: Static SageMath and dynamic ChatGPT juxtaposed

We investigated some difficulties that students often face when studying...
research
09/09/2019

Scheduling optimization of parallel linear algebra algorithms using Supervised Learning

Linear algebra algorithms are used widely in a variety of domains, e.g m...
research
01/01/2022

Matrix Decomposition and Applications

In 1954, Alston S. Householder published Principles of Numerical Analysi...
research
05/16/2022

Cloud Matrix Machine for Julia and Implicit Parallelization for Matrix Languages

Matrix computations are widely used in increasing sizes and complexity i...

Please sign up or login with your details

Forgot password? Click here to reset