The future forecasting ability of machine learning (ML) makes ML a promising tool for predicting long-time quantum dissipative dynamics of open systems. In this Article, we employ nonparametric machine learning algorithm (kernel ridge regression as a representative of the kernel methods) to study the quantum dissipative dynamics of the widely-used spin-boson model. Our ML model takes short-time dynamics as an input and is used for fast propagation of the long-time dynamics, greatly reducing the computational effort in comparison with the traditional approaches. Presented results show that the ML model performs well in both symmetric and asymmetric spin-boson models. Our approach is not limited to spin-boson model and can be extended to complex systems.
1. updated discussion and schemes to clarify method. 2. demonstrated machine learning dynamics for times longer than those present in the training set