End-to-End Quantum Vision Transformer: Towards Practical Quantum Speedup in Large-Scale Models

The field of quantum deep learning presents significant opportunities for advancing computational capabilities, yet it faces a major obstacle in the form of the "information loss problem" due to the inherent limitations of the necessary quantum tomography in scaling quantum deep neural networks. This paper introduces an end-to-end Quantum Vision Transformer (QViT), which incorporates an innovative quantum residual connection technique, to overcome these challenges and therefore optimize quantum computing processes in deep learning. Our thorough complexity analysis of the QViT reveals a theoretically exponential and empirically polynomial speedup, showcasing the model's efficiency and potential in quantum computing applications. We conducted extensive numerical tests on modern, large-scale transformers and datasets, establishing the QViT as a pioneering advancement in applying quantum deep neural networks in practical scenarios. Our work provides a comprehensive quantum deep learning paradigm, which not only demonstrates the versatility of current quantum linear algebra algorithms but also promises to enhance future research and development in quantum deep learning..

Medienart:

Preprint

Erscheinungsjahr:

2024

Erschienen:

2024

Enthalten in:

arXiv.org - (2024) vom: 29. Feb. Zur Gesamtaufnahme - year:2024

Sprache:

Englisch

Beteiligte Personen:

Xue, Cheng [VerfasserIn]
Chen, Zhao-Yun [VerfasserIn]
Zhuang, Xi-Ning [VerfasserIn]
Wang, Yun-Jie [VerfasserIn]
Sun, Tai-Ping [VerfasserIn]
Wang, Jun-Chao [VerfasserIn]
Liu, Huan-Yu [VerfasserIn]
Wu, Yu-Chun [VerfasserIn]
Wang, Zi-Lei [VerfasserIn]
Guo, Guo-Ping [VerfasserIn]

Links:

Volltext [kostenfrei]

Themen:

530
Quantum Physics

Förderinstitution / Projekttitel:

PPN (Katalog-ID):

XCH042735955