Learning Networks from Wide-Sense Stationary Stochastic Processes

dc.contributor.authorRayas, Anirudh
dc.contributor.authorCheng, Jiajun
dc.contributor.authorAnguluri, Rajasekhar
dc.contributor.authorDeka, Deepjyoti
dc.contributor.authorDasarathy, Gautam
dc.date.accessioned2025-01-31T18:24:07Z
dc.date.available2025-01-31T18:24:07Z
dc.date.issued2024-12-04
dc.description.abstractComplex networked systems driven by latent inputs are common in fields like neuroscience, finance, and engineering. A key inference problem here is to learn edge connectivity from node outputs (potentials). We focus on systems governed by steady-state linear conservation laws: Xₜ = L*Yₜ, where Xₜ, Yₜ ∈ Rᵖ denote inputs and potentials, respectively, and the sparsity pattern of the p × p Laplacian L* encodes the edge structure. Assuming Xₜ to be a wide-sense stationary stochastic process with a known spectral density matrix, we learn the support of L* from temporally correlated samples of Yₜ via an ℓ₁-regularized Whittle’s maximum likelihood estimator (MLE). The regularization is particularly useful for learning large-scale networks in the high-dimensional setting where the network size p significantly exceeds the number of samples n. We show that the MLE problem is strictly convex, admitting a unique solution. Under a novel mutual incoherence condition and certain sufficient conditions on (n, p, d), we show that the ML estimate recovers the sparsity pattern of L* with high probability, where d is the maximum degree of the graph underlying L*. We provide recovery guarantees for L* in element-wise maximum, Frobenius, and operator norms. Finally, we complement our theoretical results with several simulation studies on synthetic and benchmark datasets, including engineered systems (power and water networks), and real-world datasets from neural systems (such as the human brain).
dc.description.sponsorshipThis work was supported in part by the National Science Foundation (NSF) award CCF-2048223 and the National Institutes of Health (NIH) under the award 1R01GM140468-01. D. Deka acknowledges the funding provided by LANL’s Directed Research and Development (LDRD) project: “High-Performance Artificial Intelligence” (20230771DI).
dc.description.urihttp://arxiv.org/abs/2412.03768
dc.format.extent34 pages
dc.genrejournal articles
dc.genrepreprints
dc.identifierdoi:10.13016/m22bie-jxyx
dc.identifier.urihttps://doi.org/10.1016/bs.armc.2021.09.001
dc.identifier.urihttp://hdl.handle.net/11603/37551
dc.language.isoen_US
dc.relation.isAvailableAtThe University of Maryland, Baltimore County (UMBC)
dc.relation.ispartofUMBC Computer Science and Electrical Engineering Department
dc.relation.ispartofUMBC Faculty Collection
dc.rightsThis item is likely protected under Title 17 of the U.S. Copyright Law. Unless on a Creative Commons license, for uses protected by Copyright Law, contact the copyright holder or the author.
dc.subjectComputer Science - Machine Learning
dc.subjectStatistics - Machine Learning
dc.subjectElectrical Engineering and Systems Science - Signal Processing
dc.titleLearning Networks from Wide-Sense Stationary Stochastic Processes
dc.typeText
dcterms.creatorhttps://orcid.org/0000-0003-2537-2778

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
2412.03768v1.pdf
Size:
3.61 MB
Format:
Adobe Portable Document Format