TY - GEN
T1 - Differentially-Private Neural Network Training with Private Features and Public Labels
AU - Monir, Islam A.
AU - Ghinita, Gabriel
N1 - Publisher Copyright:
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2024.
PY - 2024
Y1 - 2024
N2 - Training neural networks (NN) with differential privacy (DP) protection has been extensively studied in the past decade, with the DP-SGD (stochastic gradient descent) mechanism representing the benchmark approach. Conventional DP-SGD assumes that both the features and the labels of training samples must be protected. A recent variation of DP-SGD considers training when the input sample features are non-private, and only labels must be protected, which improves accuracy by reducing the amount of noise injected by DP. We argue that in some scenarios, the converse holds, namely the labels may be publicly known, while the features themselves are sensitive. We provide a customized technique for this setting, we identify several design trade-offs, and we show how one can factor in such trade-offs to revise the architecture of the NN in order to improve accuracy. Extensive experiments on real data show that our approach significantly outperforms the DP-SGD baseline.
AB - Training neural networks (NN) with differential privacy (DP) protection has been extensively studied in the past decade, with the DP-SGD (stochastic gradient descent) mechanism representing the benchmark approach. Conventional DP-SGD assumes that both the features and the labels of training samples must be protected. A recent variation of DP-SGD considers training when the input sample features are non-private, and only labels must be protected, which improves accuracy by reducing the amount of noise injected by DP. We argue that in some scenarios, the converse holds, namely the labels may be publicly known, while the features themselves are sensitive. We provide a customized technique for this setting, we identify several design trade-offs, and we show how one can factor in such trade-offs to revise the architecture of the NN in order to improve accuracy. Extensive experiments on real data show that our approach significantly outperforms the DP-SGD baseline.
KW - Differential Privacy
KW - Machine Learning
KW - Neural Networks
UR - http://www.scopus.com/inward/record.url?scp=85202192103&partnerID=8YFLogxK
U2 - 10.1007/978-3-031-68323-7_16
DO - 10.1007/978-3-031-68323-7_16
M3 - Conference contribution
AN - SCOPUS:85202192103
SN - 9783031683220
VL - 14912
T3 - Lecture Notes In Computer Science
SP - 208
EP - 222
BT - Big Data Analytics And Knowledge Discovery, Dawak 2024
A2 - Wrembel, R
A2 - Chiusano, S
A2 - Kotsis, G
A2 - Tjoa, AM
A2 - Khalil, I
PB - Springer Science and Business Media Deutschland GmbH
T2 - 26th International Conference on Data Warehousing and Knowledge Discovery, DaWaK 2024
Y2 - 26 August 2024 through 28 August 2024
ER -