TY - CONF
T1 - Online factorization and partition of complex networks from random walks
AU - Yang, Lin F.
AU - Braverman, Vladimir
AU - Zhao, Tuo
AU - Wang, Mengdi
N1 - Funding Information:
This material is based upon work supported by the NSF Grants IIS-1447639, EAGER CCF- 1650041, and CAREER CCF-1652257.
Funding Information:
*Lin Yang and Mengdi Wang are affiliated with Department of Operations Research and Financial Engineering, Princeton University. Vladimir Braverman is affiliated with Department of Computer Science at Johns Hopkins University; Tuo Zhao is affiliated with School of Industrial and Systems Engineering at Georgia Institute of Technology; Email: lin.yang@princeton.edu, mengdiw@princeton.edu; Mengdi Wang is the corresponding author. This material is based upon work supported by the NSF Grants IIS-1447639, EAGER CCF-1650041, and CAREER CCF-1652257.
Publisher Copyright:
© 2019 Association For Uncertainty in Artificial Intelligence (AUAI). All rights reserved.
PY - 2019
Y1 - 2019
N2 - Finding the reduced-dimensional structure is critical to understanding complex networks. Existing approaches such as spectral clustering are applicable only when the full network is explicitly observed. In this paper, we focus on the online factorization and partition of implicit large-scale networks based on observations from an associated random walk. We formulate this into a nonconvex stochastic factorization problem and propose an efficient and scalable stochastic generalized Hebbian algorithm. The algorithm is able to process dependent state-transition data dynamically generated by the underlying network and learn a low-dimensional representation for each vertex. By applying a diffusion approximation analysis, we show that the continuous-time limiting process of the stochastic algorithm converges globally to the “principal components” of the Markov chain and achieves a nearly optimal sample complexity. Once given the learned low-dimensional representations, we further apply clustering techniques to recover the network partition. We show that when the associated Markov process is lumpable, one can recover the partition exactly with high probability. We apply the proposed approach to model the traffic flow of Manhattan as city-wide random walks. By using our algorithm to analyze the taxi trip data, we discover a latent partition of the Manhattan city that closely matches the traffic dynamics.
AB - Finding the reduced-dimensional structure is critical to understanding complex networks. Existing approaches such as spectral clustering are applicable only when the full network is explicitly observed. In this paper, we focus on the online factorization and partition of implicit large-scale networks based on observations from an associated random walk. We formulate this into a nonconvex stochastic factorization problem and propose an efficient and scalable stochastic generalized Hebbian algorithm. The algorithm is able to process dependent state-transition data dynamically generated by the underlying network and learn a low-dimensional representation for each vertex. By applying a diffusion approximation analysis, we show that the continuous-time limiting process of the stochastic algorithm converges globally to the “principal components” of the Markov chain and achieves a nearly optimal sample complexity. Once given the learned low-dimensional representations, we further apply clustering techniques to recover the network partition. We show that when the associated Markov process is lumpable, one can recover the partition exactly with high probability. We apply the proposed approach to model the traffic flow of Manhattan as city-wide random walks. By using our algorithm to analyze the taxi trip data, we discover a latent partition of the Manhattan city that closely matches the traffic dynamics.
UR - http://www.scopus.com/inward/record.url?scp=85084011938&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85084011938&partnerID=8YFLogxK
M3 - Paper
AN - SCOPUS:85084011938
T2 - 35th Conference on Uncertainty in Artificial Intelligence, UAI 2019
Y2 - 22 July 2019 through 25 July 2019
ER -