JOURNAL ARTICLE

A Sharper Generalization Bound for Divide-and-Conquer Ridge Regression

Shusen Wang

Year: 2019 Journal:   Proceedings of the AAAI Conference on Artificial Intelligence Vol: 33 (01)Pages: 5305-5312   Publisher: Association for the Advancement of Artificial Intelligence

Abstract

We study the distributed machine learning problem where the n feature-response pairs are partitioned among m machines uniformly at random. The goal is to approximately solve an empirical risk minimization (ERM) problem with the minimum amount of communication. The divide-and-conquer (DC) method, which was proposed several years ago, lets every worker machine independently solve the same ERM problem using its local feature-response pairs and the driver machine combine the solutions. This approach is in one-shot and thereby extremely communication-efficient. Although the DC method has been studied by many prior works, reasonable generalization bound has not been established before this work.For the ridge regression problem, we show that the prediction error of the DC method on unseen test samples is at most ε times larger than the optimal. There have been constantfactor bounds in the prior works, their sample complexities have a quadratic dependence on d, which does not match the setting of most real-world problems. In contrast, our bounds are much stronger. First, our 1 + ε error bound is much better than their constant-factor bounds. Second, our sample complexity is merely linear with d.

Keywords:
Divide and conquer algorithms Generalization Empirical risk minimization Feature (linguistics) Computer science Upper and lower bounds Quadratic equation Quadratic growth Sample size determination Regression Contrast (vision) Constant (computer programming) Mathematics Ridge Sample (material) Algorithm Minification Generalization error Linear regression Artificial intelligence Mathematical optimization Machine learning Statistics Artificial neural network

Metrics

10
Cited By
1.18
FWCI (Field Weighted Citation Impact)
58
Refs
0.83
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Stochastic Gradient Optimization Techniques
Physical Sciences →  Computer Science →  Artificial Intelligence
Machine Learning and Algorithms
Physical Sciences →  Computer Science →  Artificial Intelligence
Sparse and Compressive Sensing Techniques
Physical Sciences →  Engineering →  Computational Mechanics

Related Documents

JOURNAL ARTICLE

Divide and conquer kernel ridge regression

ZhangYuchenDuchiJohnWainwrightMartin

Journal:   Journal of Machine Learning Research Year: 2015
JOURNAL ARTICLE

Divide and conquer local average regression

Xiangyu ChangShao-Bo LinYao Wang

Journal:   Electronic Journal of Statistics Year: 2017 Vol: 11 (1)
JOURNAL ARTICLE

Divide-and-conquer random sketched kernel ridge regression for large-scale data analysis

Jongkyeong KangMyoungshic Jhun

Journal:   Journal of the Korean Data and Information Science Society Year: 2020 Vol: 31 (1)Pages: 15-23
JOURNAL ARTICLE

Distributed Generalized Cross-Validation for Divide-and-Conquer Kernel Ridge Regression and Its Asymptotic Optimality

Ganggang XuZuofeng ShangGuang Cheng

Journal:   Journal of Computational and Graphical Statistics Year: 2019 Vol: 28 (4)Pages: 891-908
JOURNAL ARTICLE

Divide and conquer for generalized approximately expectile regression

Zhen ZengWeixin Yao

Journal:   Journal of Statistical Planning and Inference Year: 2025 Vol: 240 Pages: 106300-106300
© 2026 ScienceGate Book Chapters — All rights reserved.