Skip to main content

Write a PREreview

QGT: A Fully Specified Quantum-Enhanced Transformer for NISQ-era Generative AI

Posted
Server
Preprints.org
DOI
10.20944/preprints202509.2354.v1

Quantum computing holds promise for accelerating Transformer-based generative models, yet existing proposals often remain at the sketch level and lack full specification for near-term devices. We introduce QGT, a fully defined hybrid quantum–classical Transformer tailored to the NISQ-to-simulation regime. Under a k-sparse attention assumption and efficient block-encoding oracles, QGT lowers the per-layer attention cost from \( O(n^2d) \) to \( O(\sqrt{n}\,d) \)​. We provide a unified algorithmic and complexity framework with rigorous theorems and proofs, detailed quantum circuit implementations with parameter-shift gradient derivations and measurement-variance bounds, and comprehensive resource accounting of qubits, gates, and shots. A reproducible classical simulation and ablation study for n = 8 and d = 16 demonstrates that QGT matches classical Transformer performance using only 12 qubits and 40 shots per expectation. QGT thus establishes a concrete foundation for practical quantum-enhanced generative AI on NISQ hardware.

You can write a PREreview of QGT: A Fully Specified Quantum-Enhanced Transformer for NISQ-era Generative AI. A PREreview is a review of a preprint and can vary from a few sentences to a lengthy report, similar to a journal-organized peer-review report.

Before you start

We will ask you to log in with your ORCID iD. If you don’t have an iD, you can create one.

What is an ORCID iD?

An ORCID iD is a unique identifier that distinguishes you from everyone with the same or similar name.

Start now