LLMs Can Understand Encrypted Prompt: Towards Privacy-Computing Friendly Transformers

05/28/2023
by   Xuanqi Liu, et al.
0

Prior works have attempted to build private inference frameworks for transformer-based large language models (LLMs) in a server-client setting, where the server holds the model parameters and the client inputs the private data for inference. However, these frameworks impose significant overhead when the private inputs are forward propagated through the original LLMs. In this paper, we show that substituting the computation- and communication-heavy operators in the transformer architecture with privacy-computing friendly approximations can greatly reduce the private inference costs with minor impact on model performance. Compared to the state-of-the-art Iron (NeurIPS 2022), our privacy-computing friendly model inference pipeline achieves a 5× acceleration in computation and an 80% reduction in communication overhead, while retaining nearly identical accuracy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/23/2023

Primer: Fast Private Transformer Inference on Encrypted Data

It is increasingly important to enable privacy-preserving inference for ...
research
06/14/2022

Private Set Matching Protocols

We introduce Private Set Matching (PSM) problems, in which a client aims...
research
11/10/2020

A Systematic Comparison of Encrypted Machine Learning Solutions for Image Classification

This work provides a comprehensive review of existing frameworks based o...
research
10/24/2022

Real-time Speech Interruption Analysis: From Cloud to Client Deployment

Meetings are an essential form of communication for all types of organiz...
research
03/27/2018

Cloud-based MPC with Encrypted Data

This paper explores the privacy of cloud outsourced Model Predictive Con...
research
12/26/2022

LOCKS: User Differentially Private and Federated Optimal Client Sampling

With changes in privacy laws, there is often a hard requirement for clie...
research
08/25/2023

Falcon: Accelerating Homomorphically Encrypted Convolutions for Efficient Private Mobile Network Inference

Efficient networks, e.g., MobileNetV2, EfficientNet, etc, achieves state...

Please sign up or login with your details

Forgot password? Click here to reset