|国家预印本平台
首页|Initialization and training of matrix product state probabilistic models

Initialization and training of matrix product state probabilistic models

Initialization and training of matrix product state probabilistic models

来源:Arxiv_logoArxiv
英文摘要

Modeling probability distributions via the wave function of a quantum state is central to quantum-inspired generative modeling and quantum state tomography (QST). We investigate a common failure mode in training randomly initialized matrix product states (MPS) using gradient descent. The results show that the trained MPS models do not accurately predict the strong interactions between boundary sites in periodic spin chain models. In the case of the Born machine algorithm, we further identify a causality trap, where the trained MPS models resemble causal models that ignore the non-local correlations in the true distribution. We propose two complementary strategies to overcome the training failure -- one through optimization and one through initialization. First, we develop a natural gradient descent (NGD) method, which approximately simulates the gradient flow on tensor manifolds and significantly enhances training efficiency. Numerical experiments show that NGD avoids local minima in both Born machines and in general MPS tomography. Remarkably, we show that NGD with line search can converge to the global minimum in only a few iterations. Second, for the BM algorithm, we introduce a warm-start initialization based on the TTNS-Sketch algorithm. We show that gradient descent under a warm initialization does not encounter the causality trap and admits rapid convergence to the ground truth.

Xun Tang、Yuehaw Khoo、Lexing Ying

物理学计算技术、计算机技术

Xun Tang,Yuehaw Khoo,Lexing Ying.Initialization and training of matrix product state probabilistic models[EB/OL].(2025-05-09)[2025-06-05].https://arxiv.org/abs/2505.06419.点此复制

评论