TY - JOUR
T1 - Online learning of quantum states
AU - Aaronson, Scott
AU - Hazan, Elad
AU - Chen, Xinyi
AU - Kale, Satyen
AU - Nayak, Ashwin
N1 - Funding Information:
⇤Supported by a Vannevar Bush Faculty Fellowship from the US Department of Defense. Part of this work was done while the author was supported by an NSF Alan T. Waterman Award. †Part of this work was done when the author was a research assistant at Princeton University. ‡Research supported in part by NSERC Canada.
Publisher Copyright:
© 2018 Curran Associates Inc.All rights reserved.
PY - 2018
Y1 - 2018
N2 - Suppose we have many copies of an unknown n-qubit state . We measure some copies of using a known two-outcome measurement E1, then other copies using a measurement E2, and so on. At each stage t, we generate a current hypothesis !t about the state , using the outcomes of the previous measurements. We show that it is possible to do this in a way that guarantees that |Tr(Ei!t) Tr(Ei)|, the error in our prediction for the next measurement, is at least " at most On/"2 times. Even in the “non-realizable” setting-where there could be arbitrary noise in the measurement outcomes-we show how to output hypothesis states that incur at most O(pTn ) excess loss over the best possible state on the first T measurements. These results generalize a 2007 theorem by Aaronson on the PAC-learnability of quantum states, to the online and regret-minimization settings. We give three different ways to prove our results-using convex optimization, quantum postselection, and sequential fat-shattering dimension-which have different advantages in terms of parameters and portability.
AB - Suppose we have many copies of an unknown n-qubit state . We measure some copies of using a known two-outcome measurement E1, then other copies using a measurement E2, and so on. At each stage t, we generate a current hypothesis !t about the state , using the outcomes of the previous measurements. We show that it is possible to do this in a way that guarantees that |Tr(Ei!t) Tr(Ei)|, the error in our prediction for the next measurement, is at least " at most On/"2 times. Even in the “non-realizable” setting-where there could be arbitrary noise in the measurement outcomes-we show how to output hypothesis states that incur at most O(pTn ) excess loss over the best possible state on the first T measurements. These results generalize a 2007 theorem by Aaronson on the PAC-learnability of quantum states, to the online and regret-minimization settings. We give three different ways to prove our results-using convex optimization, quantum postselection, and sequential fat-shattering dimension-which have different advantages in terms of parameters and portability.
UR - http://www.scopus.com/inward/record.url?scp=85064820108&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85064820108&partnerID=8YFLogxK
M3 - Conference article
AN - SCOPUS:85064820108
SN - 1049-5258
VL - 2018-December
SP - 8962
EP - 8972
JO - Advances in Neural Information Processing Systems
JF - Advances in Neural Information Processing Systems
T2 - 32nd Conference on Neural Information Processing Systems, NeurIPS 2018
Y2 - 2 December 2018 through 8 December 2018
ER -