Nothing Special   »   [go: up one dir, main page]

Comprehensive Analysis of Privacy Leakage in Vertical Federated Learning During Prediction

Authors: Xue Jiang (Technical University of Munich; Huawei Technologies Düsseldorf GmbH), Xuebing Zhou (Huawei Technologies Düsseldorf GmbH), Jens Grossklags (Technical University of Munich)

Volume: 2022
Issue: 2
Pages: 263–281
DOI: https://doi.org/10.2478/popets-2022-0045

Download PDF

Abstract: Vertical federated learning (VFL), a variant of federated learning, has recently attracted increasing attention. An active party having the true labels jointly trains a model with other parties (referred to as passive parties) in order to use more features to achieve higher model accuracy. During the prediction phase, all the parties collaboratively compute the predicted confidence scores of each target record and the results will be finally returned to the active party. However, a recent study by Luo et al. [28] pointed out that the active party can use these confidence scores to reconstruct passiveparty features and cause severe privacy leakage. In this paper, we conduct a comprehensive analysis of privacy leakage in VFL frameworks during the prediction phase. Our study improves on previous work [28] regarding two aspects. We first design a general gradientbased reconstruction attack framework that can be flexibly applied to simple logistic regression models as well as multi-layer neural networks. Moreover, besides performing the attack under the white-box setting, we give the first attempt to conduct the attack under the blackbox setting. Extensive experiments on a number of realworld datasets show that our proposed attack is effective under different settings and can achieve at best twice or thrice of a reduction of attack error compared to previous work [28]. We further analyze a list of potential mitigation approaches and compare their privacyutility performances. Experimental results demonstrate that privacy leakage from the confidence scores is a substantial privacy risk in VFL frameworks during the prediction phase, which cannot be simply solved by crypto-based confidentiality approaches. On the other hand, processing the confidence scores with information compression and randomization approaches can provide strengthened privacy protection.

Keywords: vertical federated learning, privacy attacks

Copyright in PoPETs articles are held by their authors. This article is published under a Creative Commons Attribution-NonCommercial-NoDerivs 3.0 license.