BOOK-CHAPTER

High-Dimensional Graphical Model Selection Using l1 -Regularized Logistic Regression

Abstract

We consider the problem of estimating the graph structure associated with a discrete Markov random field. We describe a method based on ℓ1-regularized logistic regression, in which the neighborhood of any given node is estimated by performing logistic regression subject to an ℓ1-constraint. Our framework applies to the high-dimensional setting, in which both the number of nodes p and maximum neighborhood sizes d are allowed to grow as a function of the number of observations n. Our main results provide sufficient conditions on the triple (n, p, d) for the method to succeed in consistently estimating the neighborhood of every node in the graph simultaneously. Under certain assumptions on the population Fisher information matrix, we prove that consistent neighborhood selection can be obtained for sample sizes n = Ω(d3 log p), with the error decaying as O(exp(−Cn/d3)) for some constant C. If these same assumptions are imposed directly on the sample matrices, we show that n= Ω(d2 log p) samples are sufficient.

Keywords:
Logistic regression Logistic model tree Selection (genetic algorithm) Computer science Graphical model Artificial intelligence Statistics Mathematics Pattern recognition (psychology) Machine learning

Metrics

42
Cited By
4.22
FWCI (Field Weighted Citation Impact)
0
Refs
0.95
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Face and Expression Recognition
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Neural Networks and Applications
Physical Sciences →  Computer Science →  Artificial Intelligence

Related Documents

© 2026 ScienceGate Book Chapters — All rights reserved.