Are You Copying My Model? Protecting the Copyright of Large Language Models for EaaS via Backdoor Watermark

05/17/2023
by   Wenjun Peng, et al.
0

Large language models (LLMs) have demonstrated powerful capabilities in both text understanding and generation. Companies have begun to offer Embedding as a Service (EaaS) based on these LLMs, which can benefit various natural language processing (NLP) tasks for customers. However, previous studies have shown that EaaS is vulnerable to model extraction attacks, which can cause significant losses for the owners of LLMs, as training these models is extremely expensive. To protect the copyright of LLMs for EaaS, we propose an Embedding Watermark method called EmbMarker that implants backdoors on embeddings. Our method selects a group of moderate-frequency words from a general text corpus to form a trigger set, then selects a target embedding as the watermark, and inserts it into the embeddings of texts containing trigger words as the backdoor. The weight of insertion is proportional to the number of trigger words included in the text. This allows the watermark backdoor to be effectively transferred to EaaS-stealer's model for copyright verification while minimizing the adverse impact on the original embeddings' utility. Our extensive experiments on various datasets show that our method can effectively protect the copyright of EaaS models without compromising service quality.

READ FULL TEXT
research
08/17/2022

Neural Embeddings for Text

We propose a new kind of embedding for natural language text that deeply...
research
03/18/2021

Model Extraction and Adversarial Transferability, Your BERT is Vulnerable!

Natural language processing (NLP) tasks, ranging from text classificatio...
research
06/21/2021

Membership Inference on Word Embedding and Beyond

In the text processing context, most ML models are built on word embeddi...
research
08/16/2019

Learning Conceptual-Contexual Embeddings for Medical Text

External knowledge is often useful for natural language understanding ta...
research
02/06/2023

Protecting Language Generation Models via Invisible Watermarking

Language generation models have been an increasingly powerful enabler fo...
research
12/01/2022

Embedding generation for text classification of Brazilian Portuguese user reviews: from bag-of-words to transformers

Text classification is a natural language processing (NLP) task relevant...
research
09/16/2022

Negation, Coordination, and Quantifiers in Contextualized Language Models

With the success of contextualized language models, much research explor...

Please sign up or login with your details

Forgot password? Click here to reset