A high-level overview of the QBM training loop. We propose using a coreset in place of the full data set to minimize the number of iterations needed to reach a target model quality.

Recent work has proposed and explored using coreset techniques for quantum algorithms that operate on classical data sets to accelerate the applicability of these algorithms on near-term quantum devices. We apply these ideas to Quantum Boltzmann Machines (QBM) where gradient-based steps which require Gibbs state sampling are the main computational bottle-neck during training. By using a coreset in place of the full data set, we try to minimize the number of steps needed and accelerate the overall training time. In a regime where computational time on quantum computers is a precious resource, we propose this might lead to substantial practical savings. We evaluate this approach on 6x6 binary images from an augmented bars and stripes data set using a QBM with 36 visible units and 8 hidden units. Using an Inception score inspired metric, we compare QBM training times with and without using coresets. DOI 10.1109/QCE53715.2022.00049

J. Viszlai, T. Tomesh, P. Gokhale, E. Anschuetz and F. T. Chong