JOURNAL ARTICLE

Differentially Private Distributed Online Learning

Chencheng LiPan ZhouLi XiongQian WangTing Wang

Year: 2018 Journal:   IEEE Transactions on Knowledge and Data Engineering Vol: 30 (8)Pages: 1440-1453   Publisher: IEEE Computer Society

Abstract

In the big data era, the generation of data presents some new characteristics, including wide distribution, high velocity, high dimensionality, and privacy concern. To address these challenges for big data analytics, we develop a privacy-preserving distributed online learning framework on the data collected from distributed data sources. Specifically, each node (i.e., data source) has the capacity of learning a model from its local dataset, and exchanges intermediate parameters with a random part of their own neighboring (logically connected) nodes. Hence, the topology of the communications in our distributed computing framework is unfixed in practice. As online learning always performs on the sensitive data, we introduce the notion of differential privacy (DP) into our distributed online learning algorithm (DOLA) to protect the data privacy during the learning, which prevents an adversary from inferring any significant sensitive information. Our model is of general value for big data analytics in the distributed setting, because it can provide rigorous and scalable privacy proof and have much less computational complexity when compared to classic schemes, e.g., secure multiparty computation (SMC). To tackle high-dimensional incoming data entries, we study a sparse version of the DOLA with novel DP techniques to save the computing resources and improve the utility. Furthermore, we present two modified private DOLAs to meet the need of practical applications. One is to convert the DOLA to distributed stochastic optimization in an offline setting, the other is to use the mini-batches approach to reduce the amount of the perturbation noise and improve the utility. We conduct experiments on real datasets in a configured distributed platform. Numerical experiment results validate the feasibility of our private DOLAs.

Keywords:
Computer science Differential privacy Big data Scalability Information privacy Distributed learning Curse of dimensionality Distributed database Distributed computing Distributed algorithm Theoretical computer science Data mining Artificial intelligence Computer security

Metrics

113
Cited By
8.54
FWCI (Field Weighted Citation Impact)
59
Refs
0.97
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Privacy-Preserving Technologies in Data
Physical Sciences →  Computer Science →  Artificial Intelligence
Stochastic Gradient Optimization Techniques
Physical Sciences →  Computer Science →  Artificial Intelligence
Cryptography and Data Security
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

JOURNAL ARTICLE

Differentially Private Distributed Learning

Yaqin ZhouShaojie Tang

Journal:   INFORMS journal on computing Year: 2020 Vol: 32 (3)Pages: 779-789
JOURNAL ARTICLE

Locally Differentially Private Distributed Online Learning With Guaranteed Optimality

Ziqin ChenYongqiang Wang

Journal:   IEEE Transactions on Automatic Control Year: 2024 Vol: 70 (4)Pages: 2521-2536
JOURNAL ARTICLE

Differentially private distributed online mirror descent algorithm

Meng YuanJinlong LeiYiguang Hong

Journal:   Neurocomputing Year: 2023 Vol: 551 Pages: 126531-126531
JOURNAL ARTICLE

Differentially private distributed estimation and learning

Marios PapachristouM. Amin Rahimian

Journal:   IISE Transactions Year: 2024 Vol: 57 (7)Pages: 756-772
JOURNAL ARTICLE

Differentially private distributed online learning over time‐varying digraphs via dual averaging

Dongyu HanKun LiuYeming LinYuanqing Xia

Journal:   International Journal of Robust and Nonlinear Control Year: 2021 Vol: 32 (5)Pages: 2485-2499
© 2026 ScienceGate Book Chapters — All rights reserved.