Quantum machine learning (QML) has the potential to outperform classical machine learning models, but its trainability is a challenge. Theoretical analysis has shown a tradeoff between the expressive power and trainability of quantum models. One way to address this is by restricting the search space over quantum models. Anschuetz and Gao have significantly contributed by constructing a hierarchy of efficiently trainable quantum neural networks (QNNs) that exhibit large polynomial expressivity separations over classical artificial neural networks (ANNs). This could lead to the development of more efficient and powerful machine learning models.
What is the Tradeoff Between Quantum Neural Networks’ Expressive Power and Trainability?
Quantum machine learning (QML) has been a topic of interest in recent years, with the potential of quantum devices to aid or even replace classical devices in machine learning tasks. Quantum systems can naturally represent complex probability distributions that are believed to be difficult to represent classically. However, recent theoretical analysis of QML algorithms has determined that there are classical learning tasks that are more efficiently performed by quantum machine learning models than classical machine learning models.
The optimization of the extremely complicated nonconvex loss functions that underlie all large-scale learning algorithms for neural networks is efficient in classical machine learning. This phenomenon has only recently been understood theoretically. Unfortunately, these results do not port over to the quantum setting. Generally, the training of QML models is difficult. This line of research in the quantum setting has culminated in the demonstration of a trainability-expressivity tradeoff in quantum machine learning.
Effectively, this is a statement that quantum models that are exponentially more expressive than classical models are not efficiently trainable. This suggests that the superpolynomial separations that rely on reductions to classically hard decision problems may be optimistic as they implicitly require one to have a priori knowledge of what the problem is to avoid optimizing a quantum model over an exponentially large Hilbert space.
How Can We Circumvent the Tradeoff Between Quantum Neural Networks’ Expressive Power and Trainability?
One way to sidestep these issues is by considering restricted QML models. Namely, one can restrict the search space over quantum models to guarantee efficient trainability. This has been recently studied in the context of symmetry-equivariant models, which use symmetries to explicitly restrict the allowed operations of the quantum model. However, recent results on the efficient classical simulability of certain symmetry-equivariant models have made it unclear what symmetries exist that are restrictive enough to allow for efficient trainability while being unrestrictive enough that a quantum advantage is maintained.
A different approach was taken in work by Anschuetz et al, where Gaussian circuits applied to fixed non-Gaussian states were used for the basis of a QML model. There, both efficient trainability and an expressivity separation over classical neural networks were shown. This separation was proved to hold on a certain sequence modeling task even over state-of-the-art classical models such as Transformers. However, this expressivity separation was only quadratic, making an experimental implementation showcasing an advantage impractical due to the large constant overhead of quantum error correction.
What are the Limitations of Current Quantum Neural Networks?
Another drawback of the approach taken by Anschuetz et al was that results were only known for continuous-variable (CV) quantum systems. This makes any experimental demonstration unsuitable for qubit-based quantum architectures due to large constant-factor overheads. The limitations of current quantum neural networks (QNNs) are a significant hurdle in the advancement of quantum machine learning. The tradeoff between the expressive power of QNNs and their trainability is a significant challenge that researchers are trying to overcome.
The limitations of current QNNs are not just theoretical. Practical exponential separations in expressive power over classical machine learning models are believed to be infeasible as such QNNs take a time to train that is exponential in the model size. This makes the implementation of QNNs in practical applications a significant challenge.
What are the Contributions of the Study by Anschuetz and Gao?
In their study, Anschuetz and Gao balance considerations of trainability and expressivity to construct a hierarchy of trainable quantum neural networks (QNNs) that exhibit arbitrarily large polynomial expressivity separations over classical artificial neural networks (ANNs). They show that there exists a hierarchy of sequence modeling tasks indexed by n, k that can be performed to zero error by an efficiently trainable QNN with O(n) quantum neurons. They then show that no ANN with fewer than n^(k-1) neurons can perform this task to any finite cross.
This is a significant contribution to the field of quantum machine learning. By constructing a hierarchy of efficiently trainable QNNs that exhibit unconditionally provable polynomial memory separations of arbitrary constant degree over classical neural networks in performing a classical sequence modeling task, Anschuetz and Gao have made a significant step forward in the development of practical QML models.
What is the Potential Impact of the Study by Anschuetz and Gao?
The potential impact of the study by Anschuetz and Gao is significant. By demonstrating that QNNs can be efficiently trainable and exhibit large polynomial expressivity separations over classical ANNs, they have opened up new possibilities for the application of quantum machine learning. This could potentially lead to the development of more efficient and powerful machine learning models.
Furthermore, their work suggests that quantum contextuality is the source of the expressivity separation. This suggests that other classical sequence learning problems with long-time correlations may be a regime where practical advantages in quantum machine learning may exist. This opens up new avenues of research and potential applications for quantum machine learning.
Publication details: “Arbitrary Polynomial Separations in Trainable Quantum Machine Learning”
Publication Date: 2024-02-13
Authors: Eric R. Anschuetz and Xun Gao
Source: arXiv (Cornell University)
DOI: https://doi.org/10.48550/arxiv.2402.08606
