towards quantum machine learning with tensor networks · 01/08/2018  · machine learning with...

12
Towards Quantum Machine Learning with Tensor Networks William Huggins, 1 Piyush Patil, 1 Bradley Mitchell, 1 K. Birgitta Whaley, 1 and E. Miles Stoudenmire 2 1 University of California Berkeley, Berkeley, CA 94720 USA 2 Center for Computational Quantum Physics, Flatiron Institute, 162 5th Avenue, New York, NY 10010, USA (Dated: August 1, 2018) Machine learning is a promising application of quantum computing, but challenges remain as near-term devices will have a limited number of physical qubits and high error rates. Motivated by the usefulness of tensor networks for machine learning in the classical context, we propose quantum computing approaches to both discriminative and generative learning, with circuits based on tree and matrix product state tensor networks that could have benefits for near-term devices. The result is a unified framework where classical and quantum computing can benefit from the same theoretical and algorithmic developments, and the same model can be trained classically then transferred to the quantum setting for additional optimization. Tensor network circuits can also provide qubit- efficient schemes where, depending on the architecture, the number of physical qubits required scales only logarithmically with, or independently of the input or output data sizes. We demonstrate our proposals with numerical experiments, training a discriminative model to perform handwriting recognition using a optimization procedure that could be carried out on quantum hardware, and testing the noise resilience of the trained model. I. INTRODUCTION For decades, quantum computing has promised to rev- olutionize certain computational tasks. It now appears that we stand on the eve of the first experimental demon- stration of a quantum advantage [1]. With noisy, inter- mediate scale quantum computers around the corner, it is natural to investigate the most promising applications of quantum computers and to determine how best to har- ness the limited, yet powerful resources they offer. Machine learning is a very appealing application for quantum computers because the theories of learning and of quantum mechanics both involve statistics at a fun- damental level, and machine learning techniques are in- herently resilient to noise, which may allow realization by near-term quantum computers operating without er- ror correction. But major obstacles include the limited number of qubits in near-term devices and the chal- lenges of working with real data. Real data sets may contain millions of samples, and individual samples are typically vectors with hundreds or thousands of compo- nents. Therefore one would like to find quantum algo- rithms that can perform meaningful tasks for large sets of high-dimensional samples even with a small number of noisy qubits. The quantum algorithms we propose in this work implement machine learning tasks—both discriminative and generative—using circuits equivalent to tensor net- works [2–4], specifically tree tensor networks [5–8] and matrix product states [2, 9, 10]. Tensor networks have recently been proposed as a promising architecture for machine learning with classical computers [11–13], and provide good results for both discriminative [12–18] and generative learning tasks [19, 20]. The circuits we will study contain many parameters which are not determined at the outset, in contrast to quantum algorithms such as Grover search or Shor factor- ization [21, 22]. Only the circuit geometry is fixed, while D =4 h0| h0| h0| h0| h0| h0| h0| h0| = FIG. 1. The quantum state of N qubits corresponding to a tree tensor network (left) can be realized as a quantum circuit acting on N qubits (right). The circuit is read from top to bottom, with the yellow bars representing unitary gates. The bond dimension D connecting two nodes of the tensor network is determined by number of qubits V connecting two sequential unitaries in the circuit, with D =2 V . the parameters determining the unitary operations must be optimized for the specific machine learning task. Our approach is therefore conceptually related to the quan- tum variational eigensolver [23, 24] and to the quantum approximate optimization algorithms [25], where quan- tum circuit parameters are discovered with the help of an auxiliary classical algorithm. The application of such hybrid quantum-classical al- gorithms to machine learning was recently investigated by several groups for labeling [26, 27] or generating data [28–30]. The proposals of Refs. 26, 27, 29, and 30 are re- lated to approaches we propose below, but consider very general classes of quantum circuits. This motivates the question: is there a subset of quantum circuits which are especially natural or advantageous for machine learn- ing tasks? Tensor network circuits might provide a com- pelling answer, for three main reasons: 1. Tensor network models could be implemented on small, near-term quantum devices for input and output dimensions far exceeding the number of physical qubits. If the hardware permits the mea- surement of one of the qubits separately from the arXiv:1803.11537v2 [quant-ph] 31 Jul 2018

Upload: others

Post on 05-Oct-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Towards Quantum Machine Learning with Tensor Networks · 01/08/2018  · machine learning with classical computers [11{13], and provide good results for both discriminative [12{18]

Towards Quantum Machine Learning with Tensor Networks

William Huggins,1 Piyush Patil,1 Bradley Mitchell,1 K. Birgitta Whaley,1 and E. Miles Stoudenmire2

1University of California Berkeley, Berkeley, CA 94720 USA2Center for Computational Quantum Physics, Flatiron Institute, 162 5th Avenue, New York, NY 10010, USA

(Dated: August 1, 2018)

Machine learning is a promising application of quantum computing, but challenges remain asnear-term devices will have a limited number of physical qubits and high error rates. Motivated bythe usefulness of tensor networks for machine learning in the classical context, we propose quantumcomputing approaches to both discriminative and generative learning, with circuits based on treeand matrix product state tensor networks that could have benefits for near-term devices. The resultis a unified framework where classical and quantum computing can benefit from the same theoreticaland algorithmic developments, and the same model can be trained classically then transferred tothe quantum setting for additional optimization. Tensor network circuits can also provide qubit-efficient schemes where, depending on the architecture, the number of physical qubits required scalesonly logarithmically with, or independently of the input or output data sizes. We demonstrateour proposals with numerical experiments, training a discriminative model to perform handwritingrecognition using a optimization procedure that could be carried out on quantum hardware, andtesting the noise resilience of the trained model.

I. INTRODUCTION

For decades, quantum computing has promised to rev-olutionize certain computational tasks. It now appearsthat we stand on the eve of the first experimental demon-stration of a quantum advantage [1]. With noisy, inter-mediate scale quantum computers around the corner, itis natural to investigate the most promising applicationsof quantum computers and to determine how best to har-ness the limited, yet powerful resources they offer.

Machine learning is a very appealing application forquantum computers because the theories of learning andof quantum mechanics both involve statistics at a fun-damental level, and machine learning techniques are in-herently resilient to noise, which may allow realizationby near-term quantum computers operating without er-ror correction. But major obstacles include the limitednumber of qubits in near-term devices and the chal-lenges of working with real data. Real data sets maycontain millions of samples, and individual samples aretypically vectors with hundreds or thousands of compo-nents. Therefore one would like to find quantum algo-rithms that can perform meaningful tasks for large setsof high-dimensional samples even with a small number ofnoisy qubits.

The quantum algorithms we propose in this workimplement machine learning tasks—both discriminativeand generative—using circuits equivalent to tensor net-works [2–4], specifically tree tensor networks [5–8] andmatrix product states [2, 9, 10]. Tensor networks haverecently been proposed as a promising architecture formachine learning with classical computers [11–13], andprovide good results for both discriminative [12–18] andgenerative learning tasks [19, 20].

The circuits we will study contain many parameterswhich are not determined at the outset, in contrast toquantum algorithms such as Grover search or Shor factor-ization [21, 22]. Only the circuit geometry is fixed, while

D = 4<latexit sha1_base64="JFP3kCp4Tz8drI6GelidaLuTbAk=">AAACuHicdVFNSwMxEE3X7/pV9egluAieym4RVEQQ9OBR0arQXWQ2na6hSXZJskpZ+xO86tW/5b8xXXtQWwcCjzfv5U0ySS64sUHwWfNmZufmFxaX6ssrq2vrjY3NW5MVmmGbZSLT9wkYFFxh23Ir8D7XCDIReJf0z0b9uyfUhmfqxg5yjCWkivc4A+uo6/OT/YeGHzSDqugkCMfAJ+O6fNiofUTdjBUSlWUCjOmEQW7jErTlTOCwHhUGc2B9SLHjoAKJJi6rWYd01zFd2su0O8rSiv3pKEEaM5CJU0qwj+Zvb0RO63UK2zuMS67ywqJi30G9QlCb0dHDaZdrZFYMHACmuZuVskfQwKz7nnqk8JllUoLqllEf7bATxmWEyhQaR1nlix9GGlTqHjj8rU40TKgjUUn98GWKurq+NdVAncNv0X+S4Cn9L8kZf7jcTsO/G5wE7VbzqBlc7funZ+PlLpJtskP2SEgOyCm5IJekTRhJySt5I+/esQde6vFvqVcbe7bIr/L0F6Yz3Ac=</latexit><latexit sha1_base64="JFP3kCp4Tz8drI6GelidaLuTbAk=">AAACuHicdVFNSwMxEE3X7/pV9egluAieym4RVEQQ9OBR0arQXWQ2na6hSXZJskpZ+xO86tW/5b8xXXtQWwcCjzfv5U0ySS64sUHwWfNmZufmFxaX6ssrq2vrjY3NW5MVmmGbZSLT9wkYFFxh23Ir8D7XCDIReJf0z0b9uyfUhmfqxg5yjCWkivc4A+uo6/OT/YeGHzSDqugkCMfAJ+O6fNiofUTdjBUSlWUCjOmEQW7jErTlTOCwHhUGc2B9SLHjoAKJJi6rWYd01zFd2su0O8rSiv3pKEEaM5CJU0qwj+Zvb0RO63UK2zuMS67ywqJi30G9QlCb0dHDaZdrZFYMHACmuZuVskfQwKz7nnqk8JllUoLqllEf7bATxmWEyhQaR1nlix9GGlTqHjj8rU40TKgjUUn98GWKurq+NdVAncNv0X+S4Cn9L8kZf7jcTsO/G5wE7VbzqBlc7funZ+PlLpJtskP2SEgOyCm5IJekTRhJySt5I+/esQde6vFvqVcbe7bIr/L0F6Yz3Ac=</latexit><latexit sha1_base64="JFP3kCp4Tz8drI6GelidaLuTbAk=">AAACuHicdVFNSwMxEE3X7/pV9egluAieym4RVEQQ9OBR0arQXWQ2na6hSXZJskpZ+xO86tW/5b8xXXtQWwcCjzfv5U0ySS64sUHwWfNmZufmFxaX6ssrq2vrjY3NW5MVmmGbZSLT9wkYFFxh23Ir8D7XCDIReJf0z0b9uyfUhmfqxg5yjCWkivc4A+uo6/OT/YeGHzSDqugkCMfAJ+O6fNiofUTdjBUSlWUCjOmEQW7jErTlTOCwHhUGc2B9SLHjoAKJJi6rWYd01zFd2su0O8rSiv3pKEEaM5CJU0qwj+Zvb0RO63UK2zuMS67ywqJi30G9QlCb0dHDaZdrZFYMHACmuZuVskfQwKz7nnqk8JllUoLqllEf7bATxmWEyhQaR1nlix9GGlTqHjj8rU40TKgjUUn98GWKurq+NdVAncNv0X+S4Cn9L8kZf7jcTsO/G5wE7VbzqBlc7funZ+PlLpJtskP2SEgOyCm5IJekTRhJySt5I+/esQde6vFvqVcbe7bIr/L0F6Yz3Ac=</latexit> h0|

<latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit>

h0|<latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit>

h0|<latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit>

h0|<latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit>

h0|<latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit>

h0|<latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit>

h0|<latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit>

h0|<latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit><latexit sha1_base64="Jt93F4dYNDzyEEPAyamQ54qlMrI=">AAACvHicdVHLSsNAFJ3Gd33r0s1gEFyVpAjqSsGNSwVrhSbIzfS2DpmZhJlJpcR8hFvd+Fv+jdPYhdp6YeBw7jlzX0kuuLFB8NnwFhaXlldW15rrG5tb2zu7e/cmKzTDDstEph8SMCi4wo7lVuBDrhFkIrCbpFeTfHeE2vBM3dlxjrGEoeIDzsA6qhslGsqgetzxg1ZQB50F4RT4ZBo3j7uNj6ifsUKiskyAMb0wyG1cgracCayaUWEwB5bCEHsOKpBo4rLut6JHjunTQabdU5bW7E9HCdKYsUycUoJ9Mn9zE3JerlfYwVlccpUXFhX7LjQoBLUZnQxP+1wjs2LsADDNXa+UPYEGZt2KmpHCZ5ZJCapfRinaqhfGZYTKFBontcoXP4w0qKEbsPqtdmucUUeilvrhyxx1/X17roE6h9+m/1SC0fC/Ss74w+VuGv694CzotFvnreD2xL+8mh53lRyQQ3JMQnJKLsk1uSEdwkhKXskbefcuPPRST35LvcbUs09+hTf6AvII3jM=</latexit>

=<latexit sha1_base64="uFbKRTBq720KdpN34ecdDTho5nE=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHoRCLx4VrArdpcym0zU0yS5JtlLW/gKvevdv+W9M1x7U1oHA4817eZNMkgtubBB81ryl5ZXVtfWN+ubW9s7u3v7BvckKzbDDMpHpxwQMCq6wY7kV+JhrBJkIfEiG7Wn/YYTa8Ezd2XGOsYRU8QFnYB11e9Xb84NGUBWdB+EM+GRWN7392kfUz1ghUVkmwJhuGOQ2LkFbzgRO6lFhMAc2hBS7DiqQaOKymnRCTxzTp4NMu6MsrdifjhKkMWOZOKUE+2T+9qbkol63sIOLuOQqLywq9h00KAS1GZ0+m/a5RmbF2AFgmrtZKXsCDcy6z6lHCp9ZJiWofhkN0U66YVxGqEyhcZpVvvhhpEGl7oGT3+pEw5w6EpXUD18WqKvrmwsN1Dn8Jv0nCUbpf0nO+MPldhr+3eA86DQbl43g9sxvtWfLXSdH5JickpCckxa5JjekQxhB8kreyLt36fU89NJvqVebeQ7Jr/LyL0OJ23s=</latexit><latexit sha1_base64="uFbKRTBq720KdpN34ecdDTho5nE=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHoRCLx4VrArdpcym0zU0yS5JtlLW/gKvevdv+W9M1x7U1oHA4817eZNMkgtubBB81ryl5ZXVtfWN+ubW9s7u3v7BvckKzbDDMpHpxwQMCq6wY7kV+JhrBJkIfEiG7Wn/YYTa8Ezd2XGOsYRU8QFnYB11e9Xb84NGUBWdB+EM+GRWN7392kfUz1ghUVkmwJhuGOQ2LkFbzgRO6lFhMAc2hBS7DiqQaOKymnRCTxzTp4NMu6MsrdifjhKkMWOZOKUE+2T+9qbkol63sIOLuOQqLywq9h00KAS1GZ0+m/a5RmbF2AFgmrtZKXsCDcy6z6lHCp9ZJiWofhkN0U66YVxGqEyhcZpVvvhhpEGl7oGT3+pEw5w6EpXUD18WqKvrmwsN1Dn8Jv0nCUbpf0nO+MPldhr+3eA86DQbl43g9sxvtWfLXSdH5JickpCckxa5JjekQxhB8kreyLt36fU89NJvqVebeQ7Jr/LyL0OJ23s=</latexit><latexit sha1_base64="uFbKRTBq720KdpN34ecdDTho5nE=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHoRCLx4VrArdpcym0zU0yS5JtlLW/gKvevdv+W9M1x7U1oHA4817eZNMkgtubBB81ryl5ZXVtfWN+ubW9s7u3v7BvckKzbDDMpHpxwQMCq6wY7kV+JhrBJkIfEiG7Wn/YYTa8Ezd2XGOsYRU8QFnYB11e9Xb84NGUBWdB+EM+GRWN7392kfUz1ghUVkmwJhuGOQ2LkFbzgRO6lFhMAc2hBS7DiqQaOKymnRCTxzTp4NMu6MsrdifjhKkMWOZOKUE+2T+9qbkol63sIOLuOQqLywq9h00KAS1GZ0+m/a5RmbF2AFgmrtZKXsCDcy6z6lHCp9ZJiWofhkN0U66YVxGqEyhcZpVvvhhpEGl7oGT3+pEw5w6EpXUD18WqKvrmwsN1Dn8Jv0nCUbpf0nO+MPldhr+3eA86DQbl43g9sxvtWfLXSdH5JickpCckxa5JjekQxhB8kreyLt36fU89NJvqVebeQ7Jr/LyL0OJ23s=</latexit>

FIG. 1. The quantum state of N qubits corresponding toa tree tensor network (left) can be realized as a quantumcircuit acting on N qubits (right). The circuit is read from topto bottom, with the yellow bars representing unitary gates.The bond dimension D connecting two nodes of the tensornetwork is determined by number of qubits V connecting twosequential unitaries in the circuit, with D = 2V .

the parameters determining the unitary operations mustbe optimized for the specific machine learning task. Ourapproach is therefore conceptually related to the quan-tum variational eigensolver [23, 24] and to the quantumapproximate optimization algorithms [25], where quan-tum circuit parameters are discovered with the help ofan auxiliary classical algorithm.

The application of such hybrid quantum-classical al-gorithms to machine learning was recently investigatedby several groups for labeling [26, 27] or generating data[28–30]. The proposals of Refs. 26, 27, 29, and 30 are re-lated to approaches we propose below, but consider verygeneral classes of quantum circuits. This motivates thequestion: is there a subset of quantum circuits whichare especially natural or advantageous for machine learn-ing tasks? Tensor network circuits might provide a com-pelling answer, for three main reasons:

1. Tensor network models could be implemented onsmall, near-term quantum devices for inputand output dimensions far exceeding the number ofphysical qubits. If the hardware permits the mea-surement of one of the qubits separately from the

arX

iv:1

803.

1153

7v2

[qu

ant-

ph]

31

Jul 2

018

Page 2: Towards Quantum Machine Learning with Tensor Networks · 01/08/2018  · machine learning with classical computers [11{13], and provide good results for both discriminative [12{18]

2

others, then the number of physical qubits neededcan be made to scale either logarithmically withthe size of the processed data, or independently ofthe data size depending on the particular tensornetwork architecture. Models based on tensor net-works may also have an inherent resilience to noise.We explore both of these aspects in Section IV.

2. There is a gradual crossover from classically sim-ulable tensor network circuits to circuits that re-quire a quantum computer to evaluate. With clas-sical resources, tensor network models already givevery good results for supervised [12, 13, 15, 17]and unsupervised [17, 19] learning tasks. Thesame models—with the same dataset size and datadimension—can be used to initialize more expres-sive models requiring quantum hardware, mak-ing the optimization of the quantum-based modelfaster and more likely to succeed. Algorithmic im-provements in the classical setting can be readilytransferred to the quantum setting as well.

3. There is a rich theoretical understanding ofthe properties of tensor networks [2–4, 10, 31, 32],and their relative mathematical simplicity (involv-ing only linear operations) will likely facilitate fur-ther conceptual developments in the machine learn-ing context, such as interpretability and generaliza-tion. Properties of tensor networks, such as localityof correlations, may provide a favorable inductivebias for processing natural data [14]. One can proverigorous bounds on the noise-resilience of quantumcircuits based on tensor networks [33].

All of the experimental operations necessary to imple-ment tensor network circuits are available for near-termquantum hardware. The capabilities required are prepa-ration of product states; one- and two-qubit unitary op-erations; and measurement in the computational basis.

In what follows, we first describe our proposed frame-works for discriminative and generative learning tasks inSection II. Then we present results of a numerical exper-iment which demonstrates the feasibility of the approachusing operations that could be carried out with an actualquantum device in Section III. We conclude by discussinghow the learning approaches could be implemented witha small number of physical qubits and by addressing theirresilience to noise in Section IV.

II. LEARNING WITH TENSOR NETWORKQUANTUM CIRCUITS

The family of tensor networks we will consider—treetensor networks and matrix product states—can alwaysbe realized precisely by a quantum circuit; see Fig. 1.Typically, the quantum circuits corresponding to tensornetworks are carefully devised to make them efficient toprepare and manipulate with classical computers [34].

With increasing bond dimension, tree and matrix prod-uct state tensor gradually capture a wider range of states,which translates into more expressive and powerful mod-els within the context of machine learning.

For very large bond dimensions, tree and matrix prod-uct tensor networks can eventually encompass the en-tire state space. But when the bond dimensions becometoo high, the cost of the classical approach becomes pro-hibitive. By implementing tensor network circuits onquantum hardware instead, one could go far beyond thespace of classically tractable models.

In this section, we first describe our tensor-networkbased proposal for performing discriminative tasks withquantum hardware. The goal of a discriminative model isto produce a specific output given a certain class of input;for example, assigning labels to images. Then we describeour proposal for generative tasks, where the goal is togenerate samples from a probability distribution inferredfrom a data set. For more background on various typesof machine learning tasks, see the recent review Ref. 35.

For clarity of presentation, we shall make use of multi-qubit unitary operations in this work. However we recog-nize that in practice such unitaries must be implementedusing a more limited set of few-qubit operations, suchas the universal gate sets of one- and two-qubit opera-tors. Whether it is more productive to classically opti-mize over more general unitaries then “compile” theseinto few-qubit operations as a separate step, or to pa-rameterize the models in terms of fewer operations fromthe outset remains an interesting and important practicalquestion for further work.

A. Discriminative Algorithm

To explain the discriminative tensor network frame-work that we propose here, assume that the input tothe algorithm takes the form of a vector of N real num-bers x = (x1, x2, . . . , xN ), with each component normal-ized such that xi ∈ [0, 1]. For example, such an inputcould correspond to a grayscale image withN pixels, withindividual entries encoding normalized grayscale values.We map this vector x ∈ RN to a product state on Nqubits according to the feature map proposed in Ref. 13:

x →

|Φ(x)〉=[cos(π2x1

)sin(π2x1

)]⊗[cos(π2x2

)sin(π2x2

)]⊗ · · · ⊗[cos(π2xN

)sin(π2xN

)] .(1)

Such a state can be prepared by starting from the com-

putational basis state |0〉⊗N , then applying a single qubitunitary to each qubit n = 1, 2, . . . , N .

The model we then propose can be seen as an iterativecoarse-graining procedure that parameterizes a CPTP(completely positive trace preserving) map from an N-qubit input space to a small number of output qubitsencoding the different possible class labels. The circuit

Page 3: Towards Quantum Machine Learning with Tensor Networks · 01/08/2018  · machine learning with classical computers [11{13], and provide good results for both discriminative [12{18]

3

`<latexit sha1_base64="4GGj+A11S286gICQQPOBvxkjn1g=">AAACuXicdVFNSwMxEE3Xr1o/q0cvwUXwVHaLoIIHwYtHBatCd5HZdNrGJtklySpl7V/wqkf/lv/G7NqD2joQeLx5L2+SSTLBjQ2Cz5q3sLi0vFJfbaytb2xubTd3bk2aa4YdlopU3ydgUHCFHcutwPtMI8hE4F0yuij7d0+oDU/VjR1nGEsYKN7nDGxJRSjEw7YftIKq6CwIp8An07p6aNY+ol7KconKMgHGdMMgs3EB2nImcNKIcoMZsBEMsOugAokmLqphJ/TAMT3aT7U7ytKK/ekoQBozlolTSrBD87dXkvN63dz2T+KCqyy3qNh3UD8X1Ka0fDntcY3MirEDwDR3s1I2BA3Muv9pRAqfWSolqF4RjdBOumFcRKhMrrHMKl78MNKgBu6Bk9/qRMOMOhKV1A9f5qir69tzDdQ5/Db9JwmeBv8lOeMPl9tp+HeDs6DTbp22gusj//xiutw62SP75JCE5Jick0tyRTqEkSF5JW/k3TvzEm/oPX5LvdrUs0t+lWe+AN6U3PU=</latexit><latexit sha1_base64="4GGj+A11S286gICQQPOBvxkjn1g=">AAACuXicdVFNSwMxEE3Xr1o/q0cvwUXwVHaLoIIHwYtHBatCd5HZdNrGJtklySpl7V/wqkf/lv/G7NqD2joQeLx5L2+SSTLBjQ2Cz5q3sLi0vFJfbaytb2xubTd3bk2aa4YdlopU3ydgUHCFHcutwPtMI8hE4F0yuij7d0+oDU/VjR1nGEsYKN7nDGxJRSjEw7YftIKq6CwIp8An07p6aNY+ol7KconKMgHGdMMgs3EB2nImcNKIcoMZsBEMsOugAokmLqphJ/TAMT3aT7U7ytKK/ekoQBozlolTSrBD87dXkvN63dz2T+KCqyy3qNh3UD8X1Ka0fDntcY3MirEDwDR3s1I2BA3Muv9pRAqfWSolqF4RjdBOumFcRKhMrrHMKl78MNKgBu6Bk9/qRMOMOhKV1A9f5qir69tzDdQ5/Db9JwmeBv8lOeMPl9tp+HeDs6DTbp22gusj//xiutw62SP75JCE5Jick0tyRTqEkSF5JW/k3TvzEm/oPX5LvdrUs0t+lWe+AN6U3PU=</latexit><latexit sha1_base64="4GGj+A11S286gICQQPOBvxkjn1g=">AAACuXicdVFNSwMxEE3Xr1o/q0cvwUXwVHaLoIIHwYtHBatCd5HZdNrGJtklySpl7V/wqkf/lv/G7NqD2joQeLx5L2+SSTLBjQ2Cz5q3sLi0vFJfbaytb2xubTd3bk2aa4YdlopU3ydgUHCFHcutwPtMI8hE4F0yuij7d0+oDU/VjR1nGEsYKN7nDGxJRSjEw7YftIKq6CwIp8An07p6aNY+ol7KconKMgHGdMMgs3EB2nImcNKIcoMZsBEMsOugAokmLqphJ/TAMT3aT7U7ytKK/ekoQBozlolTSrBD87dXkvN63dz2T+KCqyy3qNh3UD8X1Ka0fDntcY3MirEDwDR3s1I2BA3Muv9pRAqfWSolqF4RjdBOumFcRKhMrrHMKl78MNKgBu6Bk9/qRMOMOhKV1A9f5qir69tzDdQ5/Db9JwmeBv8lOeMPl9tp+HeDs6DTbp22gusj//xiutw62SP75JCE5Jick0tyRTqEkSF5JW/k3TvzEm/oPX5LvdrUs0t+lWe+AN6U3PU=</latexit>

=<latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit><latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit><latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit> prepared input qubit

=<latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit><latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit><latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit> unitary transformation =<latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit><latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit><latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit> traced/unobserved qubit

=<latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit><latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit><latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit> measured/observed qubit

(a) (b) p` =<latexit sha1_base64="r/Np+kwPs3sdsXNzIemkxgxIqiQ=">AAACvnicdVFNTxsxEHUWCjQUSuixF4sVEqdoN0KCHpBSceFIpaaAsqto1pksVvyxsr1AtORX9Eql/q3+mzqbHAIJI1l6evOe39iTFYJbF0X/GsHG5oet7Z2Pzd1Pe/ufD1qHv6wuDcMe00Kb2wwsCq6w57gTeFsYBJkJvMnGl7P+zQMay7X66SYFphJyxUecgfPUXTFIUAh6QQcHYdSO6qKrIF6AkCzqetBq/E2GmpUSlWMCrO3HUeHSCozjTOC0mZQWC2BjyLHvoQKJNq3qiaf02DNDOtLGH+VozS47KpDWTmTmlRLcvX3bm5Hrev3Sjc7TiquidKjYPGhUCuo0nT2fDrlB5sTEA2CG+1kpuwcDzPlPaiYKH5mWEtSwSsbopv04rRJUtjQ4y6qewzgxoHL/wOlrdWZgRZ2IWhrGz2vU9fWdtQbqHWGHvpMED/l7Sd645PI7jd9ucBX0Ou1v7ejHadi9XCx3h3wlR+SExOSMdMkVuSY9wogkv8kL+RN8D/JABnouDRoLzxfyqoKn/6k13nM=</latexit><latexit sha1_base64="r/Np+kwPs3sdsXNzIemkxgxIqiQ=">AAACvnicdVFNTxsxEHUWCjQUSuixF4sVEqdoN0KCHpBSceFIpaaAsqto1pksVvyxsr1AtORX9Eql/q3+mzqbHAIJI1l6evOe39iTFYJbF0X/GsHG5oet7Z2Pzd1Pe/ufD1qHv6wuDcMe00Kb2wwsCq6w57gTeFsYBJkJvMnGl7P+zQMay7X66SYFphJyxUecgfPUXTFIUAh6QQcHYdSO6qKrIF6AkCzqetBq/E2GmpUSlWMCrO3HUeHSCozjTOC0mZQWC2BjyLHvoQKJNq3qiaf02DNDOtLGH+VozS47KpDWTmTmlRLcvX3bm5Hrev3Sjc7TiquidKjYPGhUCuo0nT2fDrlB5sTEA2CG+1kpuwcDzPlPaiYKH5mWEtSwSsbopv04rRJUtjQ4y6qewzgxoHL/wOlrdWZgRZ2IWhrGz2vU9fWdtQbqHWGHvpMED/l7Sd645PI7jd9ucBX0Ou1v7ejHadi9XCx3h3wlR+SExOSMdMkVuSY9wogkv8kL+RN8D/JABnouDRoLzxfyqoKn/6k13nM=</latexit><latexit sha1_base64="r/Np+kwPs3sdsXNzIemkxgxIqiQ=">AAACvnicdVFNTxsxEHUWCjQUSuixF4sVEqdoN0KCHpBSceFIpaaAsqto1pksVvyxsr1AtORX9Eql/q3+mzqbHAIJI1l6evOe39iTFYJbF0X/GsHG5oet7Z2Pzd1Pe/ufD1qHv6wuDcMe00Kb2wwsCq6w57gTeFsYBJkJvMnGl7P+zQMay7X66SYFphJyxUecgfPUXTFIUAh6QQcHYdSO6qKrIF6AkCzqetBq/E2GmpUSlWMCrO3HUeHSCozjTOC0mZQWC2BjyLHvoQKJNq3qiaf02DNDOtLGH+VozS47KpDWTmTmlRLcvX3bm5Hrev3Sjc7TiquidKjYPGhUCuo0nT2fDrlB5sTEA2CG+1kpuwcDzPlPaiYKH5mWEtSwSsbopv04rRJUtjQ4y6qewzgxoHL/wOlrdWZgRZ2IWhrGz2vU9fWdtQbqHWGHvpMED/l7Sd645PI7jd9ucBX0Ou1v7ejHadi9XCx3h3wlR+SExOSMdMkVuSY9wogkv8kL+RN8D/JABnouDRoLzxfyqoKn/6k13nM=</latexit>

`<latexit sha1_base64="4GGj+A11S286gICQQPOBvxkjn1g=">AAACuXicdVFNSwMxEE3Xr1o/q0cvwUXwVHaLoIIHwYtHBatCd5HZdNrGJtklySpl7V/wqkf/lv/G7NqD2joQeLx5L2+SSTLBjQ2Cz5q3sLi0vFJfbaytb2xubTd3bk2aa4YdlopU3ydgUHCFHcutwPtMI8hE4F0yuij7d0+oDU/VjR1nGEsYKN7nDGxJRSjEw7YftIKq6CwIp8An07p6aNY+ol7KconKMgHGdMMgs3EB2nImcNKIcoMZsBEMsOugAokmLqphJ/TAMT3aT7U7ytKK/ekoQBozlolTSrBD87dXkvN63dz2T+KCqyy3qNh3UD8X1Ka0fDntcY3MirEDwDR3s1I2BA3Muv9pRAqfWSolqF4RjdBOumFcRKhMrrHMKl78MNKgBu6Bk9/qRMOMOhKV1A9f5qir69tzDdQ5/Db9JwmeBv8lOeMPl9tp+HeDs6DTbp22gusj//xiutw62SP75JCE5Jick0tyRTqEkSF5JW/k3TvzEm/oPX5LvdrUs0t+lWe+AN6U3PU=</latexit><latexit sha1_base64="4GGj+A11S286gICQQPOBvxkjn1g=">AAACuXicdVFNSwMxEE3Xr1o/q0cvwUXwVHaLoIIHwYtHBatCd5HZdNrGJtklySpl7V/wqkf/lv/G7NqD2joQeLx5L2+SSTLBjQ2Cz5q3sLi0vFJfbaytb2xubTd3bk2aa4YdlopU3ydgUHCFHcutwPtMI8hE4F0yuij7d0+oDU/VjR1nGEsYKN7nDGxJRSjEw7YftIKq6CwIp8An07p6aNY+ol7KconKMgHGdMMgs3EB2nImcNKIcoMZsBEMsOugAokmLqphJ/TAMT3aT7U7ytKK/ekoQBozlolTSrBD87dXkvN63dz2T+KCqyy3qNh3UD8X1Ka0fDntcY3MirEDwDR3s1I2BA3Muv9pRAqfWSolqF4RjdBOumFcRKhMrrHMKl78MNKgBu6Bk9/qRMOMOhKV1A9f5qir69tzDdQ5/Db9JwmeBv8lOeMPl9tp+HeDs6DTbp22gusj//xiutw62SP75JCE5Jick0tyRTqEkSF5JW/k3TvzEm/oPX5LvdrUs0t+lWe+AN6U3PU=</latexit><latexit sha1_base64="4GGj+A11S286gICQQPOBvxkjn1g=">AAACuXicdVFNSwMxEE3Xr1o/q0cvwUXwVHaLoIIHwYtHBatCd5HZdNrGJtklySpl7V/wqkf/lv/G7NqD2joQeLx5L2+SSTLBjQ2Cz5q3sLi0vFJfbaytb2xubTd3bk2aa4YdlopU3ydgUHCFHcutwPtMI8hE4F0yuij7d0+oDU/VjR1nGEsYKN7nDGxJRSjEw7YftIKq6CwIp8An07p6aNY+ol7KconKMgHGdMMgs3EB2nImcNKIcoMZsBEMsOugAokmLqphJ/TAMT3aT7U7ytKK/ekoQBozlolTSrBD87dXkvN63dz2T+KCqyy3qNh3UD8X1Ka0fDntcY3MirEDwDR3s1I2BA3Muv9pRAqfWSolqF4RjdBOumFcRKhMrrHMKl78MNKgBu6Bk9/qRMOMOhKV1A9f5qir69tzDdQ5/Db9JwmeBv8lOeMPl9tp+HeDs6DTbp22gusj//xiutw62SP75JCE5Jick0tyRTqEkSF5JW/k3TvzEm/oPX5LvdrUs0t+lWe+AN6U3PU=</latexit>

FIG. 2. Discriminative tree tensor network model architec-ture, showing an example in which V = 2 qubits connectdifferent subtrees. Figure (a) shows the model implementa-tion as a quantum circuit. Circles indicate inputs preparedin a product state as in Eq. 1; hash marks indicate qubitsthat remain unobserved past a certain point in the circuit. Aparticular pre-determined qubit is sampled (square symbol)and its distribution serves as the output of the model. Figure(b) shows the tensor network diagram for the reduced densitymatrix of the output qubit.

takes the form of a tree, with V qubit lines connectingeach subtree to the rest of the circuit. We call such qubitlines “virtual qubits” to connect with the terminology oftensor networks, where tensor indices internal to the net-work are called virtual indices. A larger V can capturea larger set of functions, just as a tensor network witha sufficiently large bond dimension can parameterize anyN-index tensor.

At each step, we take V of the qubits resulting fromone of the unitary operations of the previous step, or sub-tree, and V from another subtree and act on them withanother parameterized unitary transformation (possiblytogether with some ancilla qubits—not shown). Then Vof the qubits are discarded, while the other V proceed tothe next node of the tree, that is, the next step of the

FIG. 3. The connectivity of nodes of our tree network model,as it would be applied to a 4x4 image. Each step coarse-grainsin either the horizontal or the vertical directions, and thesesteps alternate.

circuit. In our classical simulations we trace over all dis-carded qubits, while on a quantum computer, we wouldbe free to ignore or reset such qubits.

Once all unitary operations defining the circuit havebeen carried out, one or more qubits serve as the outputqubits. (Which qubits are outputs is designated aheadof time.) The most probable state of the output qubitsdetermines the prediction of the model, that is, the la-bel the model assigns to the input. To determine themost probable state of the output qubits, one performsrepeated evaluations of the circuit for the same inputin order to estimate their probability distribution in thecomputational basis.

We show the quantum circuit of our proposed proce-dure in Fig. 2. In the case of image classification, itis natural to always group input qubits based on pixelscoming from nearby regions of the image, with a treestructure illustrated schematically in Fig. 3.

A closely related family of models can be devised basedon matrix product states. An example is illustrated inFig. 4 showing the case of V = 2. Matrix product states(MPS) can be viewed as maximally unbalanced trees, anddiffer from the binary tree models described above in thatafter each unitary operation on 2V inputs only one setof V qubits are passed to the next node of the network.Such models are likely a better fit for data that has a one-dimensional pattern of correlations, such as time-series,language, or audio data.

B. Generative Algorithm

The generative algorithm we propose is nearly the re-verse of the discriminative algorithm, in terms of its cir-cuit architecture. The algorithm produces random sam-ples by first preparing a quantum state then measuring itin the computational basis, putting it within the family of

FIG. 4. Discriminative tensor network model for the caseof a matrix product state (MPS) architecture with V = 2qubits connecting each subtree. The symbols have the samemeaning as in Fig. 2. An MPS can be viewed as a maximallyunbalanced tree.

Page 4: Towards Quantum Machine Learning with Tensor Networks · 01/08/2018  · machine learning with classical computers [11{13], and provide good results for both discriminative [12{18]

4

algorithms recently dubbed “Born machines” [19, 28, 29].But rather than preparing a completely general state,we shall consider specific patterns of state preparationcorresponding to tree and matrix product state tensornetworks. This provides the advantages discussed in theintroduction, such as connections to classical tensor net-work models and the ability to reduce the number ofphysical qubits required, which will be discussed furtherin Section IV.

The generative algorithm based on a tree tensor net-work (shown in Fig. 5) begins by preparing 2V qubits

in a reference computational basis state 〈0|⊗2V , then en-tangling these qubits by unitary operations. Another set

of 2V qubits are prepared in the state 〈0|⊗2V . Half ofthese are grouped with the first V entangled qubits, andhalf with the second V entangled qubits. Two more uni-tary operations are applied to each new grouping of 2Vqubits; the outputs are now split into four groups; andthe process repeats for each group. The process endswhen the total number of qubits processed reaches thesize of the output one wants to generate.

Once all unitaries acting on a certain qubit have beenapplied, this qubit can be measured. The measured out-put of all of the qubits in the computational basis repre-

=<latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit><latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit><latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit> unitary transformation

=<latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit><latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit><latexit sha1_base64="NRrUKEqzmqotoFv+0mTMVdMJHg8=">AAACtnicdVFNSwMxEE3X7/qtRy/BRfBUdougHpSCF48KVoXuUmbT6RqaZJckq5Rtf4FXvfu3/Demaw9q60Dg8ea9vEkmyQU3Ngg+a97C4tLyyupafX1jc2t7Z3fv3mSFZthmmcj0YwIGBVfYttwKfMw1gkwEPiSDq0n/4Rm14Zm6s8McYwmp4n3OwDrq9qK74weNoCo6C8Ip8Mm0brq7tY+ol7FCorJMgDGdMMhtXIK2nAkc16PCYA5sACl2HFQg0cRlNemYHjmmR/uZdkdZWrE/HSVIY4YycUoJ9sn87U3Ieb1OYftncclVXlhU7DuoXwhqMzp5Nu1xjcyKoQPANHezUvYEGph1n1OPFL6wTEpQvTIaoB13wriMUJlC4ySrHPlhpEGl7oHj3+pEw4w6EpXUD0dz1NX1zbkG6hx+k/6TBM/pf0nO+MPldhr+3eAsaDcb543g9sRvXU6Xu0oOyCE5JiE5JS1yTW5ImzCC5JW8kXfv3Ot66KXfUq829eyTX+XlX0II23Y=</latexit> measured/observed qubit

= h0| = [1, 0]<latexit sha1_base64="mA1wXIdvp6YroCJlfjHpPKbzkVs=">AAACyXicdVHBThsxEHUWaNNQSoBjLxYrJA5VtBsh0R6oInHh0AOVGkDaXUWzziRYsb0r20tJlz3xH4hb+SX+BmfJAUgYydLTm/f87Jk0F9zYIHhseCurax8+Nj+11j9vfNlsb22fmazQDPssE5m+SMGg4Ar7lluBF7lGkKnA83RyPOufX6E2PFN/7DTHRMJY8RFnYB01aG8f0TjVUAYVPaJR+I0GyaDtB52gLroIwjnwybxOB1uN+3iYsUKiskyAMVEY5DYpQVvOBFatuDCYA5vAGCMHFUg0SVk/vqJ7jhnSUabdUZbW7EtHCdKYqUydUoK9NG97M3JZLyrs6HtScpUXFhV7DhoVgtqMziZBh1wjs2LqADDN3VspuwQNzLp5tWKFf1kmJahhGU/QVlGYlDEqU2icZZU3fhhrUGP3weq12s1zQR2LWuqHN0vU9fXdpQbqHH6XvpMEV+P3kpzxhcvtNHy7wUXQ73Z+dILfB37v53y5TfKV7JJ9EpJD0iMn5JT0CSPX5I78Jw/eL097196/Z6nXmHt2yKvybp8AfUHhDA==</latexit><latexit sha1_base64="mA1wXIdvp6YroCJlfjHpPKbzkVs=">AAACyXicdVHBThsxEHUWaNNQSoBjLxYrJA5VtBsh0R6oInHh0AOVGkDaXUWzziRYsb0r20tJlz3xH4hb+SX+BmfJAUgYydLTm/f87Jk0F9zYIHhseCurax8+Nj+11j9vfNlsb22fmazQDPssE5m+SMGg4Ar7lluBF7lGkKnA83RyPOufX6E2PFN/7DTHRMJY8RFnYB01aG8f0TjVUAYVPaJR+I0GyaDtB52gLroIwjnwybxOB1uN+3iYsUKiskyAMVEY5DYpQVvOBFatuDCYA5vAGCMHFUg0SVk/vqJ7jhnSUabdUZbW7EtHCdKYqUydUoK9NG97M3JZLyrs6HtScpUXFhV7DhoVgtqMziZBh1wjs2LqADDN3VspuwQNzLp5tWKFf1kmJahhGU/QVlGYlDEqU2icZZU3fhhrUGP3weq12s1zQR2LWuqHN0vU9fXdpQbqHH6XvpMEV+P3kpzxhcvtNHy7wUXQ73Z+dILfB37v53y5TfKV7JJ9EpJD0iMn5JT0CSPX5I78Jw/eL097196/Z6nXmHt2yKvybp8AfUHhDA==</latexit><latexit sha1_base64="mA1wXIdvp6YroCJlfjHpPKbzkVs=">AAACyXicdVHBThsxEHUWaNNQSoBjLxYrJA5VtBsh0R6oInHh0AOVGkDaXUWzziRYsb0r20tJlz3xH4hb+SX+BmfJAUgYydLTm/f87Jk0F9zYIHhseCurax8+Nj+11j9vfNlsb22fmazQDPssE5m+SMGg4Ar7lluBF7lGkKnA83RyPOufX6E2PFN/7DTHRMJY8RFnYB01aG8f0TjVUAYVPaJR+I0GyaDtB52gLroIwjnwybxOB1uN+3iYsUKiskyAMVEY5DYpQVvOBFatuDCYA5vAGCMHFUg0SVk/vqJ7jhnSUabdUZbW7EtHCdKYqUydUoK9NG97M3JZLyrs6HtScpUXFhV7DhoVgtqMziZBh1wjs2LqADDN3VspuwQNzLp5tWKFf1kmJahhGU/QVlGYlDEqU2icZZU3fhhrUGP3weq12s1zQR2LWuqHN0vU9fXdpQbqHH6XvpMEV+P3kpzxhcvtNHy7wUXQ73Z+dILfB37v53y5TfKV7JJ9EpJD0iMn5JT0CSPX5I78Jw/eL097196/Z6nXmHt2yKvybp8AfUHhDA==</latexit>

FIG. 5. Generative tree tensor network model architecture,showing a case with V = 2 qubits connecting each subtree.To sample from the model, qubits are prepared in a referencecomputational basis state 〈0| (left-hand side of circuit). Then2V qubits are entangled via unitary operations at each layerof the tree as shown. The qubits are measured at the pointsin the circuit labeled by square symbols (right-hand side ofcircuit), and the results of these measurements provides theoutput of the model. While all qubits could be entangledbefore being measured, we discuss in Section IV the possi-bility performing opportunistic measurements to reduce thephysical qubit overhead.

sents one sample from the generative model.We illustrate our proposed generative approach for the

case of V = 2 and binary outputs in Fig. 5. As in thediscriminative case, one can also devise an MPS basedgenerative algorithm more suitable for one-dimensionaldata. The circuit for such an algorithm is shown in Fig. 6.

III. NUMERICAL EXPERIMENTS

To show the feasibility of implementing our proposal ona near-term quantum device, we trained a discriminativemodel based on a tree tensor network for a supervisedlearning task, namely labeling image data. The specificnetwork architecture we used is shown as a quantum cir-cuit in Fig. 7. When viewed as a tensor network, thismodel has a bond dimension of D = 2. This stems fromthe fact that after each unitary operation entangles twoqubits, only one of the qubits is acted on at the next scale(next step of the circuit).

A. Loss Function

Our eventual goal is to select the parameters of ourcircuit such that we can confidently assign the correctlabel to a new piece of data by running our circuit asmall number of times. To this end, we choose the lossfunction which we want to minimize starting with thefollowing definitions. Let Λ be the model parameters; dbe an element of the training data set; and let p`(Λ,x) bethe probability of the model to output a label ` for a giveninput x. Because we consider the setting of supervisedlearning, the correct labels are known for the training setinputs, and define `x to be the correct label for the inputx. Now define

plargest false(Λ,x) = max` 6=`x

[p`(Λ,x)

](2)

FIG. 6. Generative tensor network model for the case of amatrix product state (MPS) architecture with V = 2 qubitsconnecting each unitary. The symbols have the same meaningas in Fig. 5.

Page 5: Towards Quantum Machine Learning with Tensor Networks · 01/08/2018  · machine learning with classical computers [11{13], and provide good results for both discriminative [12{18]

5

FIG. 7. Model architecture used in the experiments of Sec-tion III, which is a special case of the model of Fig. 2 withone virtual qubit connecting each subtree. For illustrationpurposes we show a model with 16 inputs and 4 layers above,whereas the actual model used in the experiments had 64 in-puts and 6 layers.

as the probability of the incorrect output state which hasthe highest probability of being observed. Then, definethe loss function for a single input x to be

L(Λ,x) = max(plargest false(Λ,x)− p`x(Λ,x) + λ, 0)η,(3)

and the total loss function to be

L(Λ) =1

|data|∑

x∈data

L(Λ,x). (4)

The “hyper-parameters” λ and η are to be chosen togive good empirical performance on a validation data set.Essentially, we assign a penalty for each element of thetraining set where the gap between probability of assign-ing the true label and the probability of assigning themost likely incorrect label is less than λ. This loss func-tion allows us to concentrate our efforts during trainingon making sure that we are likely to assign the correctlabel after taking the majority vote of several executionsof the model, rather than trying to force the model toalways output the correct label in each separate run.

B. Optimization

Of course, we are interested in training our circuit togeneralize well to unobserved inputs, so instead of opti-mizing over the entire distribution of data as in Eq. 4, weoptimize the loss function over a subset of the trainingdata and compare to a held-out set of test data. Further-more, because the size of the training set for a typicalmachine learning problem is so large (60,000 examples in

the case of the MNIST data set), it would be impracti-cal to calculate the loss over all of the training data ateach optimization step. Instead, we follow a standard ap-proach in machine learning and randomly select a mini-batch of training examples at each iteration. Then, weuse the following stochastic estimate of our true train-ing loss (recalling that Λ represents the current modelparameters):

L(Λ) =1

|mini-batch|∑

x∈mini-batch

L(Λ,x) (5)

In order to faithfully test how our approach would per-form on a near-term quantum computer, we have chosento minimize our loss function using a variant of the simul-taneous perturbation stochastic approximation (SPSA)algorithm which was recently used to find quantum cir-cuits approximating ground states in Ref. 33 and wasoriginally developed in Ref. 36.

Essentially, each step of SPSA estimates the gradient ofthe loss function by performing a finite difference calcu-lation along a random direction and updates the param-eters accordingly. In our experimentation, we have alsofound it helpful to include a momentum term v, whichmixes a fraction of previous update steps into the currentupdate. We outline the algorithm we used in more detailbelow.

1. Initialize the model parameters Λ randomly, andset v to zero.

2. Choose appropriate values for the constants,a, b, A, s, t, γ, n,M that define the optimization pro-cedure.

3. For each k ∈ {0, 1, 2, ...,M}, set αk = a(k+1+A)s and

βk = b(k+1)t , and randomly partition the training

data into mini-batches of n images. Perform thefollowing steps using each mini-batch:

(a) Generate random perturbation ∆ in parame-ter space.

(b) Evaluate g = L(Λold+αk∆)−L(Λold−αk∆)2αk

, with

L(x) defined as in Eq. 5.

(c) Set vnew = γvold − gβk∆(d) Set Λnew = Λold + vnew

C. Results

We trained circuits with a single output qubit at eachnode to recognize grayscale images of size 8 × 8 belong-ing to one of two classes using the SPSA optimizationprocedure described above. The images were obtainedfrom the MNIST data set of handwritten digits [37], andwe show results below for classifiers trained to distin-guish between each of the 45 pairs of handwritten digits0 through 9.

Page 6: Towards Quantum Machine Learning with Tensor Networks · 01/08/2018  · machine learning with classical computers [11{13], and provide good results for both discriminative [12{18]

6

FIG. 8. Test accuracy as a function of the number of SPSAepochs (M = 30, in the language of the previous section)for binary classification of handwritten 0’s and 7’s from theMNIST data set.

The unitary operations U applied at each node in thetree were parameterized by writing them as U = exp(iH)where H is a Hermitian matrix (the matrices H were al-lowed to be different for each node). The free parame-ters were chosen to be the elements forming the diagonaland upper triangle of each Hermitian matrix, resulting inexactly 1008 free parameters for the 8 × 8 image recog-nition task. The mini-batch size and the other hyper-parameters for the training procedure and the loss func-tion were hand-tuned by running a small number of ex-periments, using the SigOpt [38] software package, withthe goal of obtaining the most rapid and consistent per-formance (averaged over the different digit pairs) on avalidation data set.

Ultimately, we found that networks trained with thechoices (λ = .234, η = 5.59, a = 28.0, b = 33.0, A =74.1, s = 4.13, t = .658, γ = 0.882, n = 222) wereable to achieve an average test accuracy above 95%. Theaccuracies of the individual pairwise classifiers are tabu-lated in Fig. 9, and data from a representative exampleof the training process for one of the easier pairs to clas-sify is shown in 8. We observed significant differencesin performance across the different pairs, partly owing,perhaps, to the difficulty of distinguishing similar digitsusing 64 pixel images. We also note that different choicesof hyper-parameters could significantly affect which pairswere classified most accurately.

IV. IMPLEMENTATION ON NEAR-TERMDEVICES

A key advantage of carrying out machine learning taskswith models equivalent to tree or matrix product tensornetworks is that they could be implemented using a very

small number of physical qubits. The key requirementis that the hardware must allow the measurement of in-dividual physical qubits without further disturbing thestate of the other qubits, a capability also required forcertain approaches to quantum error correction [39]. Be-low we will first discuss how the number of qubits neededto implement either a discriminative or generative treetensor network model can be made to scale only loga-rithmically in both the data dimension and in the bonddimension of the network. Then we will discuss the spe-cial case of matrix product state tensor networks, whichcan be implemented with a number of physical qubitsthat is independent of the input or output data dimen-sion.

Another key advantage of using tensor network modelson near-term devices could be their robustness to noise,which will certainly be present in any near-term hard-ware. To explore the noise resilience of our models, wepresent a numerical experiment where we evaluate themodel trained in Section III with random errors, and ob-serve whether it can still produce useful results.

A. Qubit-Efficient Tree Network Models

To discuss the minimum qubit resources needed to im-plement general tree tensor network models, recall thenotion of the virtual qubit number V from Section II.

FIG. 9. The test accuracy for each of the pairwise classifierstrained with the hyper-parameters mentioned in the text. Theaccuracy for each classifier can be found by choosing the po-sition along the x-axis corresponding to one class and theposition on the y-axis corresponding to the other.

Page 7: Towards Quantum Machine Learning with Tensor Networks · 01/08/2018  · machine learning with classical computers [11{13], and provide good results for both discriminative [12{18]

7

(a)

(b)

FIG. 10. Qubit-efficient scheme for evaluating (a) discrimina-tive and (b) generative tree models with V = 2 virtual qubitsand N = 16 inputs or outputs. Note that the two patterns arethe reverse of each other. In (a) qubits indicated with hashmarks are measured and the measurement results discarded.These qubits are then reset and prepared with additional in-put states. In (b) measured qubits are recorded and reset toa reference state 〈0|.

This is the number of qubit lines connecting each subtreeto higher nodes in the tree. Viewed as a tensor network,the bond dimension D, or dimension of the internal ten-sor indices, is given by D = 2V .

For example, the tree shown in Fig. 7 has V = 1 and abond dimension of D = 2. The tree shown in Fig. 10 hasV = 2 and D = 4. When discussing these models in gen-eral terms, it suffices to consider only unitary operationsacting on 2V qubits, since at each node of the tree, twosubtrees (two sets of V qubits) are entangled together.

Given only the ability to perform state preparationand unitary operations, it would take N physical qubitsto evaluate a discriminative tree network model on Ninputs. However, if we also allow the step of measure-ment and resetting of certain qubits, then the number ofphysical qubits Q required to process N inputs given Vvirtual states passing between each node can be signifi-cantly reduced to just Q(N,V ) = V lg(2N/V ).

To see why, consider the circuit showing the mostqubit-efficient scheme for implementing the discrimina-tive case Fig. 10(a). For a given V , the number of in-puts that can be processed by a single unitary is 2V .Then V of the qubits can be measured and reused, butthe other V qubits must remain entangled. So onlyV new qubits must be introduced to process 2V moreinputs. From this line of reasoning and the observa-tion that Q(2V, V ) = 2V , one can deduce the resultQ(N,V ) = V lg(2N/V ).

For generative tree network models, generating N out-

(a)

(b)

FIG. 11. Qubit-efficient scheme for evaluating (a) discrim-inative and (b) generative matrix product state models foran arbitrary number of inputs or outputs. The figure showsthe case of V = 3 qubits connecting each node of the net-work. When evaluating the discriminative model, one of thequbits is measured after each unitary is applied and the re-sult discarded; the qubit is then prepared with the next inputcomponent. To implement the generative model, one of thequbits is measured after each unitary operation and the resultrecorded. The qubit is then reset to the state 〈0|.

puts with V virtual qubits requires the same number ofphysical qubits as for the discriminative case; this can beseen by observing that the pattern of unitaries is just thereverse of the discriminative case for the same N and V .Fig. 10 shows the most qubit-efficient way to sample agenerative tree models for the case of V = 2 virtual andN = 16 output qubits, requiring only Q = 8 physicalqubits.

Though a linear growth of the number of physicalqubits as a function of virtual qubit number V mayseem more prohibitive compared to the logarithmic scal-ing with N , even a small increase in V would lead toa significantly more expressive model. From the pointof view of tensor networks the expressivity of the modelis usually measured by the bond dimension D = 2V .In terms of the bond dimension, the number of qubitsneeded thus scales only as Q(N,D) ∼ lg(D) lg(N). Thelargest bond dimensions used in state-of-the-art classicaltensor network calculations are around D = 215 or about30, 000. So for V = 16 or more virtual qubits one wouldquickly exceed the power of any classical tensor networkcalculation we are aware of.

B. Qubit-Efficient Matrix Product Models

A matrix product state (MPS) tensor network is a spe-cial case of a tree tensor network that is maximally un-balanced. This gives an MPS certain advantages withoutsacrificing expressivity for one-dimensional distributions,as measured by the maximum entanglement entropy itcan carry across bipartitions of the input or output space,meaning a division of (x1, . . . , xj) from (xj+1, . . . , xN ).

Page 8: Towards Quantum Machine Learning with Tensor Networks · 01/08/2018  · machine learning with classical computers [11{13], and provide good results for both discriminative [12{18]

8

(a)

(b)

(c)

(d)

FIG. 12. Mapping of the generative matrix product state(MPS) quantum circuit with V = 3 to a bond dimensionD = 23 MPS tensor network diagram. First (a) interpret thecircuit diagram as a tensor diagram by interpreting referencestates 〈0| as vectors [1, 0]; qubit lines as dimension 2 tensorindices; and measurements as setting indices to fixed values.Then (b) contract the reference states into the unitary tensorsand (c) redraw the tensors in a linear chain. Finally, (d) mergethree D = 2 indices into a single D = 8 dimensional index oneach bond.

Given the ability to measure and reset a subset of phys-ical qubits, a key advantage of implementing a discrim-inative or generative tensor network model based on anMPS is that for a model with V virtual qubits, an arbi-trary number of inputs or outputs can be processed byusing only V +1 physical qubits. The circuits illustratinghow this can be done are shown in Fig. 11.

The implementation of the discriminative algorithmshown in Fig. 11(a) begins by preparing and entanglingV input qubit states. One of the qubits is measured andreset to the next input state. Then all V + 1 qubits areentangled and a single qubit measured and re-prepared.Continuing in this way, one can process all of the inputs.Once all inputs are processed, the model output is ob-tained by sampling one or more of the physical qubits.

To implement the generative MPS algorithm shown inFig. 11(b), one prepares all qubits to a reference state

|0〉⊗V+1and after entangling the qubits, one measures

and records a single qubit to generate the first outputvalue. This qubit is reset to the state |0〉 and all thequbits are then acted on by another (V + 1) qubit uni-tary. A single qubit is again measured to generate thesecond output value, and the algorithm continues untilN outputs have been generated.

To understand the equivalence of the generative circuitof Fig. 11(b) to conventional tensor diagram notation foran MPS, interpret the circuit diagram Fig. 12(a) as a ten-sor network diagram, treating elements such as referencestates 〈0| as tensors or vectors [1, 0]. One can contractor sum over the reference state indices and merge any V

qubit indices into a single index of dimension D = 2V .The result is a standard MPS tensor network diagramFig. 12(d) for the amplitude of observing a particular setof values of the measured qubits.

C. Noise Resilience

Any implementation of our proposed approach onnear-term quantum hardware will have to contend witha significant level of noise due to qubit and gate imper-fections. But one intuition about noise effects in ourtree models is that an error which corrupts a qubit onlyscrambles the information coming from the patch of in-puts belonging to the past “causal cone” of that qubit.And because the vast majority of the operations occurnear the leaves of the tree, the most likely errors there-fore correspond to scrambling only small patches of theinput data. We note that a good classifier should nat-urally be robust to small deformations and corruptionsof the input, and, in fact, adding various kinds of noiseduring training is a commonly used strategy in classicalmachine learning. Based on these intuitions, we expectour circuits could demonstrate a high level of toleranceto noise.

In order to quantitatively understand the robustness ofour proposed approach to noise on quantum hardware,

FIG. 13. The test accuracy for each of the pairwise classifiersunder noise corresponding to a T1 of 5µs, a T2 of 7µs, anda gate time of 200 ns. In most cases, the accuracy is compa-rable to the results from training without noise. Note that itwas necessary to choose a different set of hyper-parameters toenable successful training under noise.

Page 9: Towards Quantum Machine Learning with Tensor Networks · 01/08/2018  · machine learning with classical computers [11{13], and provide good results for both discriminative [12{18]

9

FIG. 14. Success probability of two different pairwise clas-sification circuits prediction on their test sets (sorted by de-creasing probability of success along the x-axis) over a widerange of T1 values (y-axis). For each T1 shown, the proba-bility of successfully classifying each member of the test setis indicated. Note that success probabilities which are largerthan .5 even by a relatively small margin imply that the cor-responding test example could be correctly classified with amajority voting scheme. Gate time Tg = 200ns was heldfixed while T2 was set to be 7

5T1. Noise levels corresponding

to current hardware are approximately two thirds of the wayup the chart. Grey areas indicate regions where the modelwould misclassify the test example.

we study how performance is affected by independentamplitude-damping and dephasing channels applied toeach qubit. In particular, we investigate how this errormodel would affect the pairwise tree network discrimi-native models of the type described in Section II A andshown in Fig. 7.

The specific error model we implemented is the fol-lowing: during the contraction step of node i in themodel evaluation, we compose amplitude damping anddephasing noise channels acting on its left and rightchildren ρiL and ρiR, mapping ρiL → Ea (Ed (ρiL)) andρiR → Ea (Ed (ρiR)). Any completely positive trace-preserving noise channel E(ρ) can be expressed in the

operator-sum representation as E(ρ) =∑aMaρM

†a ,

where∑aMaM

†a = I. Here the Kraus operators Ma for

the amplitude damping channel Ea are (in the z-basis)

M0 =

(1 00√

1− pa

), M1 =

(0√pa

0 0

),

while for the dephasing channel Ed the Kraus operatorsare

M0 =√

1− pd I , M1 =

(√pd 00 0

), M2 =

(0 00√pd

).

To evaluate model performance under realistic val-ues of pa and pd on current hardware, we determine paand pd based on the continuous-time Kraus operatorsof these channels, which depend on the duration of thetwo-qubit gate Tg, the coherence time T1 of the qubits,and the dephasing time T2 of the qubits. Specifically,pa = 1 − e−Tg/T1 and pd = 1 − e−Tg/T2 . Realistic val-ues for the time scales are Tg = 200 ns and T1 = 50µs,T2 = 70µs, corresponding to pa = 0.004 and pd = 0.003.But numerical experiments with these values showed al-most no observable noise effects, so we consider an evenmore conservative parameter set with T1 and T2 reducedby an order of magnitude, such that pa = 0.039 andpd = 0.028.

We plot the resulting test accuracies in Fig. 13, notingthat the Kraus operator formalism allows us to directlycalculate the reduced density matrix of the labeling qubitunder the effects of our noise model, therefore no explicitsampling of noise realizations is needed. Given that thecoherence times used for the plot are easily achievableeven on today’s very early hardware platforms, the re-sults shown in Fig. 13 are encouraging: many of themodels give a test accuracy only slightly reduced fromthe noiseless case Fig. 9. The largest reduction was forthe digit ‘4’ versus digit ‘9’ model, which dropped from atest accuracy of 0.88 to 0.806. Interestingly this was alsothe model with the worst performance in the noiselesscase. The typical change in test accuracy across all ofthe models due to the noise was about 0.004.

To mitigate the effect of noise when classifying a par-ticular image, one can evaluate the quantum circuit somesmall number of times and choose the label which is mostfrequently observed. For example, one could take a ma-jority vote from 500 executions and classify and correctlyclassify an image whose individual probability of successis .55 with almost 99% accuracy. In order to shed a moredetailed light on our approach’s robustness to noise, weplot in Fig. 14 the individual success probabilities forclassifying each test example (x-axis), sorted by theirprobabilities for ease of visualization, over a range of de-coherence times (y-axis). The two panels show two differ-ent models, one trained to distinguish images of digits ‘0’versus ‘1’ ; the other digits ‘9’ versus ‘4’. These modelswere trained and evaluated at various levels of noise us-ing the same training hyper-parameters that were found

Page 10: Towards Quantum Machine Learning with Tensor Networks · 01/08/2018  · machine learning with classical computers [11{13], and provide good results for both discriminative [12{18]

10

to give a good performance at {pa, pd} = {0.039, 0.028}.The ratio between the dephasing time and the coherencetime was held at a fixed ratio T2/T1 = 7/5.

We see that, for the examples which our models cor-rectly classify in the low noise limit, the success probabil-ity remains appreciably greater than .5 for a wide rangeof noise levels. In both diagrams the y-axis is scaledso that coherence times achievable by today’s hardwareoccur two thirds of the way up from the bottom. Inter-estingly, we note that the success probabilities saturateat coherence times much shorter than this, and only dropoff dramatically at T1 values near T1 ∼ 1µ s. The highperformance of our model over a broad swath of testedcoherence and dephasing times suggests that the effects ofnoise on our approach can be dramatically mitigated bythe combination of the hybrid quantum/classical train-ing procedure and a small number of repetitions with amajority voting scheme. We find these results encourag-ing as empirical evidence that the limited-width “causalcone” structure possessed by models of this type mayhave inherent noise robustness properties.

V. DISCUSSION

Many of the features that make tensor networks ap-pealing for classical algorithms also make them a promis-ing framework for quantum computing. Tensor net-works provide a natural hierarchy of increasingly com-plex quantum states, allowing one to choose the appro-priate amount of resources for a given task. They alsoenable specialized algorithms which can make efficientuse of valuable resources, such as reducing the number ofqubits needed to process high dimensional data. An op-timized, classically tractable tensor network can be usedto initialize the parameters of a more powerful modelimplemented on quantum hardware. Doing so would al-leviate issues associated with random initial parameters,which can place circuits in regions of parameter spacewith vanishing gradients [40].

While the approach to optimization we considered inour numerical experiments worked well, algorithms whichare more specialized to the tensor network architecturecould be devised. For example, by defining an objectivefor each subtree of a tree network it could be possible totrain subtrees separately [17]. Likewise, the MPS archi-tecture has certain orthogonality or light-cone propertieswhich mean that only the tensors to the left of a certain

physical index determine its distribution; this propertycould also be exploited for better optimization.

Another very interesting future direction would be togain a better understanding of the noise resilience of ten-sor network machine learning algorithms. We performedsome simple numerical experiments to show that thesealgorithms can tolerate a high level of noise, but addi-tional empirical demonstrations as well as a theoreticalexplanation of how generic this property is would be veryuseful. In an interesting recent work, Kim and Swingleinvestigated tensor networks within a quantum comput-ing framework for finding ground states of local Hamil-tonians [33]. One of their results was a rigorous boundon the sensitivity of the algorithm output to noise, whichrelied on specific properties of tensor networks. It wouldbe very interesting adapt their bound to the machinelearning context.

Other tensor network architectures besides trees andMPS also deserve further investigation in the context ofquantum algorithms. The PEPS family of tensor net-works are specially designed to capture two-dimensionalpatterns of correlations [41, 42]. The MERA family oftensor networks, retain certain benefits of tree tensornetworks but have more expressive power, and admit anatural description as a quantum circuit [33, 34].

Tensor networks strike a careful balance betweenexpressive power and computational efficiency, and canbe viewed as a particularly useful and natural class ofquantum circuits. Based on the rich theoretical under-standing of their properties and powerful algorithms foroptimizing them, we are optimistic they will providemany interesting avenues for quantum machine learningresearch.

ACKNOWLEDGEMENTS

We thank Dave Bacon, Norm Tubman, AlejandroPerdomo-Ortiz, Mark Nowakowski, Vinay Ramasesh,Dar Dahlen, and Lei Wang for helpful discussions. Thework of W. Huggins and K. B. Whaley was supportedby the U.S. Department of Energy, Office of Science, Of-fice of Advanced Scientific Computing Research, Quan-tum Algorithm Teams Program, under contract numberDE-AC02-05CH11231. Computational resources wereprovided by NIH grant S10OD023532, administered bythe Molecular Graphics and Computation Facility at UCBerkeley.

[1] Sergio Boixo, Sergei V. Isakov, Vadim N. Smelyan-skiy, Ryan Babbush, Nan Ding, Zhang Jiang, John M.Martinis, and Hartmut Neven, “Characterizing quan-tum supremacy in near-term devices,” arxiv:1608.00263(2016).

[2] Stellan Ostlund and Stefan Rommer, “Thermodynamiclimit of density matrix renormalization,” Phys. Rev. Lett.75, 3537–3540 (1995).

[3] Roman Orus, “A practical introduction to tensor net-works: Matrix product states and projected entangledpair states,” Annals of Physics 349, 117 – 158 (2014).

Page 11: Towards Quantum Machine Learning with Tensor Networks · 01/08/2018  · machine learning with classical computers [11{13], and provide good results for both discriminative [12{18]

11

[4] F. Verstraete, V. Murg, and J. I. Cirac, “Matrix prod-uct states, projected entangled pair states, and varia-tional renormalization group methods for quantum spinsystems,” Advances in Physics 57, 143–224 (2008).

[5] M. Fannes, B. Nachtergaele, and R. F. Werner, “Finitelycorrelated states on quantum spin chains,” Communica-tions in Mathematical Physics 144, 443–490 (1992).

[6] M.-B. Lepetit, M. Cousy, and G.M. Pastor, “Density-matrix renormalization study of the hubbard model [4]ona bethe lattice,” The European Physical Journal B - Con-densed Matter and Complex Systems 13, 421–427 (2000).

[7] L. Tagliacozzo, G. Evenbly, and G. Vidal, “Simulationof two-dimensional quantum systems using a tree tensornetwork that exploits the entropic area law,” Phys. Rev.B 80, 235127 (2009).

[8] W. Hackbusch and S. Kuhn, “A new scheme for the ten-sor representation,” Journal of Fourier Analysis and Ap-plications 15, 706–722 (2009).

[9] Guifre Vidal, “Efficient classical simulation of slightly en-tangled quantum computations,” Phys. Rev. Lett. 91,147902 (2003).

[10] U. Schollwock, “The density-matrix renormalizationgroup in the age of matrix product states,” Annals ofPhysics 326, 96–192 (2011).

[11] Nadav Cohen, Or Sharir, and Amnon Shashua, “On theexpressive power of deep learning: A tensor analysis,”29th Annual Conference of Learning Theory , 698–728(2016).

[12] Alexander Novikov, Mikhail Trofimov, and IvanOseledets, “Exponential machines,” arxiv:1605.03795(2016).

[13] E.M. Stoudenmire and David J Schwab, “Supervisedlearning with tensor networks,” in Advances In Neu-ral Information Processing Systems 29 (2016) pp. 4799–4807, arxiv:1605.05775.

[14] Yoav Levine, David Yakira, Nadav Cohen, and AmnonShashua, “Deep learning and quantum entanglement:Fundamental connections with implications to networkdesign,” (2017), arxiv:1704.01552.

[15] Ding Liu, Shi-Ju Ran, Peter Wittek, Cheng Peng,Raul Blazquez Garcıa, Gang Su, and Maciej Lewenstein,“Machine learning by two-dimensional hierarchical tensornetworks: A quantum information theoretic perspectiveon deep architectures,” arxiv:1710.04833 (2017).

[16] Valentin Khrulkov, Alexander Novikov, and Ivan Os-eledets, “Expressive power of recurrent neural networks,”arxiv:1711.00811 (2017).

[17] E.M. Stoudenmire, “Learning relevant features of datawith multi-scale tensor networks,” arxiv:1801.00315(2018).

[18] Ivan Glasser, Nicola Pancotti, and J. Ignacio Cirac,“Supervised learning with generalized tensor networks,”arxiv:1806.05964 (2018).

[19] Zhao-Yu Han, Jun Wang, Heng Fan, Lei Wang, and PanZhang, “Unsupervised generative modeling using matrixproduct states,” arxiv:1709.01662 (2017).

[20] Chu Guo, Zhanming Jie, Wei Lu, and Dario Poletti,“Matrix product operators for sequence to sequencelearning,” arxiv:1803.10908 (2018).

[21] Lov K. Grover, “A fast quantum mechanical algorithmfor database search,” in Proceedings of the Twenty-eighth Annual ACM Symposium on Theory of Comput-ing, STOC ’96 (ACM, New York, NY, USA, 1996) pp.212–219.

[22] Peter W. Shor, “Polynomial-time algorithms for primefactorization and discrete logarithms on a quantum com-puter,” SIAM Journal on Computing 26, 1484–1509(1997).

[23] Alberto Peruzzo, Jarrod McClean, Peter Shadbolt, Man-Hong Yung, Xiao-Qi Zhou, Peter J. Love, Alan Aspuru-Guzik, and Jeremy L. O’Brien, “A variational eigenvaluesolver on a photonic quantum processor,” Nature Com-munications 5, 4213 (2014).

[24] Jarrod R McClean, Jonathan Romero, Ryan Babbush,and Alan Aspuru-Guzik, “The theory of variationalhybrid quantum-classical algorithms,” New Journal ofPhysics 18, 023023 (2016).

[25] Edward Farhi, Jeffrey Goldstone, and Sam Gut-mann, “A quantum approximate optimization algo-rithm,” arxiv:1411.4028 (2014).

[26] Edward Farhi and Hartmut Neven, “Classification withquantum neural networks on near term processors,”arxiv:1802.06002 (2018).

[27] Maria Schuld and Nathan Killoran, “Quantum machinelearning in feature hilbert spaces,” arxiv:1803.07128(2018).

[28] Xun Gao, Zhengyu Zhang, and Luming Duan, “An ef-ficient quantum algorithm for generative machine learn-ing,” arxiv:1711.02038 (2017).

[29] Marcello Benedetti, Delfina Garcia-Pintos, YunseongNam, and Alejandro Perdomo-Ortiz, “A generative mod-eling approach for benchmarking and training shallowquantum circuits,” arxiv:1801.07686 (2018).

[30] Kosuke Mitarai, Makoto Negoro, Masahiro Kita-gawa, and Keisuke Fuji, “Quantum circuit learning,”arxiv:1803.00745 (2018).

[31] G. Evenbly and G. Vidal, “Tensor network states andgeometry,” Journal of Statistical Physics 145, 891–918(2011).

[32] M.B. Hastings, “An area law for one-dimensional quan-tum systems,” J. Stat. Mech. 2007, P08024 (2007).

[33] Isaac H. Kim and Brian Swingle, “Robust entangle-ment renormalization on a noisy quantum computer,”arxiv:1711.07500 (2017).

[34] G. Vidal, “Class of quantum many-body states that canbe efficiently simulated,” Phys. Rev. Lett. 101, 110501(2008).

[35] Pankaj Mehta, Marin Bukov, Ching-Hao Wang, Alexan-dre G.R. Day, Clint Richardson, Charles K. Fisher, andDavid J. Schwab, “A high-bias, low-variance introduc-tion to machine learning for physicists,” arxiv:1803.08823(2018).

[36] James C Spall, “An overview of the simultaneous pertur-bation method for efficient optimization,” Johns Hopkinsapl technical digest 19, 482–492 (1998).

[37] Christopher J.C. Burges Yann LeCun,Corinna Cortes, “MNIST handwritten digit database,”http://yann.lecun.com/exdb/mnist/ .

[38] .[39] A. D. Corcoles, Easwar Magesan, Srikanth J. Srinivasan,

Andrew W. Cross, M. Steffen, Jay M. Gambetta, andJerry M. Chow, “Demonstration of a quantum error de-tection code using a square lattice of four supercon-ducting qubits,” Nature Communications 6, 6979 EP –(2015).

[40] Jarrod R. McClean, Sergio Boixo, Vadim N. Smelyan-skiy, Ryan Babbush, and Hartmut Neven, “Barrenplateaus in quantum neural network training land-

Page 12: Towards Quantum Machine Learning with Tensor Networks · 01/08/2018  · machine learning with classical computers [11{13], and provide good results for both discriminative [12{18]

12

scapes,” arxiv:1803.11173 (2018).[41] Martin Schwarz, Kristan Temme, and Frank Verstraete,

“Preparing projected entangled pair states on a quantumcomputer,” Phys. Rev. Lett. 108, 110502 (2012).

[42] Martin Schwarz, Kristan Temme, Frank Verstraete,David Perez-Garcia, and Toby S. Cubitt, “Preparingtopological projected entangled pair states on a quantumcomputer,” Phys. Rev. A 88, 032321 (2013).