DeepAI AI Chat
Log In Sign Up

Parallelizing Machine Learning as a Service for the End-User

by   Daniela Loreti, et al.

As ML applications are becoming ever more pervasive, fully-trained systems are made increasingly available to a wide public, allowing end-users to submit queries with their own data, and to efficiently retrieve results. With increasingly sophisticated such services, a new challenge is how to scale up to evergrowing user bases. In this paper, we present a distributed architecture that could be exploited to parallelize a typical ML system pipeline. We start from a case study consisting of a text mining service and then discuss how the proposed method can be generalized to many similar applications. We demonstrate the significance of the computational gain boosted by the distributed architecture by way of an extensive experimental evaluation.


page 1

page 2

page 3

page 4


Minerva: A Portable Machine Learning Microservice Framework for Traditional Enterprise SaaS Applications

In traditional SaaS enterprise applications, microservices are an essent...

Training and Serving Machine Learning Models at Scale

In recent years, Web services are becoming more and more intelligent (e....

SystemDS: A Declarative Machine Learning System for the End-to-End Data Science Lifecycle

Machine learning (ML) applications become increasingly common in many do...

Machine Learning for Metaverse-enabled Wireless Systems: Vision, Requirements, and Challenges

Today's wireless systems are posing key challenges in terms of quality o...

SLO-ML: A Language for Service Level Objective Modelling in Multi-cloud Applications

Cloud modelling languages (CMLs) are designed to assist customers in tac...
08/24/2017 Towards Multi-tenant Resource Sharing for Machine Learning Workloads

We present, a declarative machine learning service platform we b...

fairDMS: Rapid Model Training by Data and Model Reuse

Extracting actionable information from data sources such as the Linac Co...