Skip to main content
CenXiv.org
This website is in trial operation, support us!
We gratefully acknowledge support from all contributors.
Contribute
Donate
cenxiv logo > quant-ph > arXiv:2407.00769

Help | Advanced Search

Quantum Physics

arXiv:2407.00769 (quant-ph)
[Submitted on 30 Jun 2024 ]

Title: Achieving Energetic Superiority Through System-Level Quantum Circuit Simulation

Title: 通过系统级量子电路仿真实现能量优势

Authors:Rong Fu, Zhongling Su, Han-Sen Zhong, Xiti Zhao, Jianyang Zhang, Feng Pan, Pan Zhang, Xianhe Zhao, Ming-Cheng Chen, Chao-Yang Lu, Jian-Wei Pan, Zhiling Pei, Xingcheng Zhang, Wanli Ouyang
Abstract: Quantum Computational Superiority boasts rapid computation and high energy efficiency. Despite recent advances in classical algorithms aimed at refuting the milestone claim of Google's sycamore, challenges remain in generating uncorrelated samples of random quantum circuits. In this paper, we present a groundbreaking large-scale system technology that leverages optimization on global, node, and device levels to achieve unprecedented scalability for tensor networks. This enables the handling of large-scale tensor networks with memory capacities reaching tens of terabytes, surpassing memory space constraints on a single node. Our techniques enable accommodating large-scale tensor networks with up to tens of terabytes of memory, reaching up to 2304 GPUs with a peak computing power of 561 PFLOPS half-precision. Notably, we have achieved a time-to-solution of 14.22 seconds with energy consumption of 2.39 kWh which achieved fidelity of 0.002 and our most remarkable result is a time-to-solution of 17.18 seconds, with energy consumption of only 0.29 kWh which achieved a XEB of 0.002 after post-processing, outperforming Google's quantum processor Sycamore in both speed and energy efficiency, which recorded 600 seconds and 4.3 kWh, respectively.
Abstract: 量子计算优势具有快速计算和高能效的特点。 尽管在经典算法方面取得了进展,旨在反驳谷歌Sycamore的里程碑声明,但生成随机量子电路的不相关样本仍然存在挑战。 在本文中,我们提出了一种突破性的大规模系统技术,该技术利用全局、节点和设备级别的优化,实现了张量网络前所未有的可扩展性。 这使得能够处理内存容量达到数十TB的大规模张量网络,超越单个节点的内存空间限制。 我们的技术能够容纳多达数十TB内存的大规模张量网络,最多可使用2304块GPU,峰值计算能力为561 PFLOPS半精度。 值得注意的是,我们在能耗为2.39 kWh的情况下实现了14.22秒的求解时间,达到了0.002的保真度,我们最显著的结果是在后处理后仅消耗0.29 kWh能耗的情况下实现了17.18秒的求解时间,达到了XEB值为0.002,其速度和能效均超过了谷歌的量子处理器Sycamore,后者分别记录了600秒和4.3 kWh。
Subjects: Quantum Physics (quant-ph) ; Distributed, Parallel, and Cluster Computing (cs.DC)
Cite as: arXiv:2407.00769 [quant-ph]
  (or arXiv:2407.00769v1 [quant-ph] for this version)
  https://doi.org/10.48550/arXiv.2407.00769
arXiv-issued DOI via DataCite

Submission history

From: Han-Sen Zhong [view email]
[v1] Sun, 30 Jun 2024 17:14:59 UTC (13,678 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled
  • View Chinese PDF
  • View PDF
  • HTML (experimental)
  • TeX Source
license icon view license
Current browse context:
quant-ph
< prev   |   next >
new | recent | 2024-07
Change to browse by:
cs
cs.DC

References & Citations

  • NASA ADS
  • Google Scholar
  • Semantic Scholar
a export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo

Bibliographic and Citation Tools

Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)

Code, Data and Media Associated with this Article

alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)

Demos

Replicate (What is Replicate?)
Hugging Face Spaces (What is Spaces?)
TXYZ.AI (What is TXYZ.AI?)

Recommenders and Search Tools

Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
IArxiv Recommender (What is IArxiv?)
  • Author
  • Venue
  • Institution
  • Topic

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)
  • About
  • Help
  • contact arXivClick here to contact arXiv Contact
  • subscribe to arXiv mailingsClick here to subscribe Subscribe
  • Copyright
  • Privacy Policy
  • Web Accessibility Assistance
  • arXiv Operational Status
    Get status notifications via email or slack

京ICP备2025123034号