Attention to quantum complexity

  • Hyejin Kim
  • , Yiqing Zhou
  • , Yichen Xu
  • , Kaarthik Varma
  • , Amir H. Karamlou
  • , Ilan T. Rosen
  • , Jesse C. Hoke
  • , Chao Wan
  • , Jin Peng Zhou
  • , William D. Oliver
  • , Yuri D. Lensky
  • , Kilian Q. Weinberger
  • , Eun Ah Kim

Research output: Contribution to journalArticlepeer-review

Abstract

The imminent era of error-corrected quantum computing demands robust methods to characterize quantum state complexity from limited, noisy measurements. We introduce the Quantum Attention Network (QuAN), a classical artificial intelligence (AI) framework leveraging attention mechanisms tailored for learning quantum complexity. Inspired by large language models, QuAN treats measurement snapshots as tokens while respecting permutation invariance. Combined with our parameter-efficient miniset self-attention block, this enables QuAN to access high-order moments of bit-string distributions and preferentially attend to less noisy snapshots. We test QuAN across three quantum simulation settings: driven hard-core Bose-Hubbard model, random quantum circuits, and toric code under coherent and incoherent noise. QuAN directly learns entanglement and state complexity growth from experimental computational basis measurements, including complexity growth in random circuits from noisy data. In regimes inaccessible to existing theory, QuAN unveils the complete phase diagram for noisy toric code data as a function of both noise types, highlighting AI's transformative potential for assisting quantum hardware.

Original languageEnglish
Pages (from-to)eadu0059
JournalScience Advances
Volume11
Issue number41
DOIs
StatePublished - 10 Oct 2025

Fingerprint

Dive into the research topics of 'Attention to quantum complexity'. Together they form a unique fingerprint.

Cite this