SafeSpeech:

Robust and Universal Voice Protection Against Malicious Speech Synthesis

[Paper][GitHub]

Zhisheng Zhang, Derui Wang, Qianyi Yang, Pengyang Huang, Junhan Pu, Yuxin Cao, Kai Ye, Jie Hao, Yixian Yang

Abstract: Speech synthesis technology has brought great convenience, while the widespread usage of realistic deepfake audio has triggered hazards. Malicious adversaries may unauthorizedly collect victims' speeches and clone a similar voice for illegal exploitation (e.g., telecom fraud). However, the existing defense methods cannot effectively prevent deepfake exploitation and are vulnerable to robust training techniques. Therefore, a more effective and robust data protection method is urgently needed. In response, we propose a defensive framework, SafeSpeech, which protects the users' audio before uploading by embedding imperceptible perturbations on original speeches to prevent high-quality synthetic speech. In SafeSpeech, we devise a robust and universal proactive protection technique, Speech PErturbative Concealment (SPEC), that leverages a surrogate model to generate universally applicable perturbation for generative synthetic models. Moreover, we optimize the human perception of embedded perturbation in terms of time and frequency domains. To evaluate our method comprehensively, we conduct extensive experiments across advanced models and datasets, both subjectively and objectively. Our experimental results demonstrate that SafeSpeech achieves state-of-the-art (SOTA) voice protection effectiveness and transferability and is highly robust against advanced adaptive adversaries. Moreover, SafeSpeech has real-time capability in real-world tests. The source code is available at https://github.com/wxzyd123/SafeSpeech.

Contents

SafeSpeech Workflow

Figure 1. The SafeSpeech safeguards voice by constructing a surrogate TTS model that minimizes the designed objectives (L_mel and L_noise with perception constraint L_perception detailed in Section 4). Despite attackers fine-tuning advanced TTS models from social platforms, they cannot produce high-quality synthetic speech to circumvent voiceprint locks or deceive victims' families.

Original and Protected Speech

These samples are undefended and protected samples from LibriTTS dataset of speaker 5339, respectively.

Method Sample 1 Sample 2 Sample 3
Original
Protected random noise
AdvPoison
SEP
PTA
AttackVC
AntiFake
SafeSpeech
denoised

Synthesized Speech based on Fine-tuning

BERT-VITS2

Method Sample 1 Sample 2 Sample 3
clean
random-noise
AdvPoison
SEP
PTA
AntiFake
AttackVC
SafeSpeech

Other Models with Fine-tuning Capabilities

Models Method Sample 1 Sample 2
StyleTTS 2 clean
SafeSpeech
MB-iSTFT-VITS clean
SafeSpeech
VITS clean
SafeSpeech
GlowTTS clean
SafeSpeech

Synthesized Speech based on Zero-shot

These are synthesized speech from advanced zero-shot TTS models.

Models Method Sample 1 Sample 2
Original
TorToise-TTS clean
SafeSpeech
XTTS clean
SafeSpeech
OpenVoice clean
SafeSpeech
FishSpeech clean
SafeSpeech
F5-TTS clean
SafeSpeech