Adopting Two Supervisors for Efficient Use of Large-Scale Remote Deep Neural Networks

04/05/2023
by   Michael Weiss, et al.
0

Recent decades have seen the rise of large-scale Deep Neural Networks (DNNs) to achieve human-competitive performance in a variety of artificial intelligence tasks. Often consisting of hundreds of millions, if not hundreds of billion parameters, these DNNs are too large to be deployed to, or efficiently run on resource-constrained devices such as mobile phones or IoT microcontrollers. Systems relying on large-scale DNNs thus have to call the corresponding model over the network, leading to substantial costs for hosting and running the large-scale remote model, costs which are often charged on a per-use basis. In this paper, we propose BiSupervised, a novel architecture, where, before relying on a large remote DNN, a system attempts to make a prediction on a small-scale local model. A DNN supervisor monitors said prediction process and identifies easy inputs for which the local prediction can be trusted. For these inputs, the remote model does not have to be invoked, thus saving costs, while only marginally impacting the overall system accuracy. Our architecture furthermore foresees a second supervisor to monitor the remote predictions and identify inputs for which not even these can be trusted, allowing to raise an exception or run a fallback strategy instead. We evaluate the cost savings, and the ability to detect incorrectly predicted inputs on four diverse case studies: IMDB movie review sentiment classification, Github issue triaging, Imagenet image classification, and SQuADv2 free-text question answering

READ FULL TEXT
research
08/24/2022

CheapET-3: Cost-Efficient Use of Remote DNN Models

On complex problems, state of the art prediction accuracy of Deep Neural...
research
05/02/2022

Simple Techniques Work Surprisingly Well for Neural Network Test Prioritization and Active Learning (Replicability Study)

Test Input Prioritizers (TIP) for Deep Neural Networks (DNN) are an impo...
research
08/31/2018

Rx-Caffe: Framework for evaluating and training Deep Neural Networks on Resistive Crossbars

Deep Neural Networks (DNNs) are widely used to perform machine learning ...
research
12/20/2019

Lightweight and Unobtrusive Privacy Preservation for Remote Inference via Edge Data Obfuscation

The growing momentum of instrumenting the Internet of Things (IoT) with ...
research
05/31/2019

Multi-Precision Quantized Neural Networks via Encoding Decomposition of -1 and +1

The training of deep neural networks (DNNs) requires intensive resources...
research
06/18/2021

Quantized Neural Networks via -1, +1 Encoding Decomposition and Acceleration

The training of deep neural networks (DNNs) always requires intensive re...
research
02/01/2021

Fail-Safe Execution of Deep Learning based Systems through Uncertainty Monitoring

Modern software systems rely on Deep Neural Networks (DNN) when processi...

Please sign up or login with your details

Forgot password? Click here to reset