Skip to main content
CenXiv.org
This website is in trial operation, support us!
We gratefully acknowledge support from all contributors.
Contribute
Donate
cenxiv logo > cs > arXiv:2506.10092

Help | Advanced Search

Computer Science > Databases

arXiv:2506.10092 (cs)
[Submitted on 11 Jun 2025 (v1) , last revised 3 Sep 2025 (this version, v2)]

Title: GPU Acceleration of SQL Analytics on Compressed Data

Title: GPU加速的压缩数据SQL分析

Authors:Zezhou Huang, Krystian Sakowski, Hans Lehnert, Wei Cui, Carlo Curino, Matteo Interlandi, Marius Dumitru, Rathijit Sen
Abstract: GPUs are uniquely suited to accelerate (SQL) analytics workloads thanks to their massive compute parallelism and High Bandwidth Memory (HBM) -- when datasets fit in the GPU HBM, performance is unparalleled. Unfortunately, GPU HBMs remain typically small when compared with lower-bandwidth CPU main memory. Besides brute-force scaling across many GPUs, current solutions to accelerate queries on large datasets include leveraging data partitioning and loading smaller data batches in GPU HBM, and hybrid execution with a connected device (e.g., CPUs). Unfortunately, these approaches are exposed to the limitations of lower main memory and host-to-device interconnect bandwidths, introduce additional I/O overheads, or incur higher costs. This is a substantial problem when trying to scale adoption of GPUs on larger datasets. Data compression can alleviate this bottleneck, but to avoid paying for costly decompression/decoding, an ideal solution must include computation primitives to operate directly on data in compressed form. This is the focus of our paper: a set of new methods for running queries directly on light-weight compressed data using schemes such as Run-Length Encoding (RLE), index encoding, bit-width reductions, and dictionary encoding. Our novelty includes operating on multiple RLE columns without decompression, handling heterogeneous column encodings, and leveraging PyTorch tensor operations for portability across devices. Experimental evaluations show speedups of an order of magnitude compared to state-of-the-art commercial CPU-only analytics systems, for real-world queries on a production dataset that would not fit into GPU memory uncompressed. This work paves the road for GPU adoption in a much broader set of use cases, and it is complementary to most other scale-out or fallback mechanisms.
Abstract: GPU由于其巨大的计算并行性和高带宽内存(HBM)而特别适合加速(SQL)分析工作负载——当数据集适合GPU HBM时,性能是无与伦比的。 不幸的是,与低带宽的CPU主存相比,GPU HBM通常仍然较小。 除了在多个GPU上进行暴力扩展外,当前加速大型数据集查询的解决方案包括利用数据分区和将较小的数据批次加载到GPU HBM中,以及与连接设备(例如CPU)的混合执行。 不幸的是,这些方法受到较低主存和主机到设备互连带宽的限制,会引入额外的I/O开销或增加更高的成本。 当试图在更大的数据集上扩展GPU的采用时,这是一个重大问题。 数据压缩可以缓解这一瓶颈,但为了避免支付昂贵的解压/解码费用,理想解决方案必须包括可以直接在压缩数据上运行的计算原语。 这就是我们论文的重点:一种新的方法集合,使用如行程长度编码(RLE)、索引编码、位宽减少和字典编码等方案,在轻量级压缩数据上直接运行查询。 我们的创新包括在不进行解压缩的情况下操作多个RLE列,处理异构列编码,并利用PyTorch张量操作实现跨设备的可移植性。 实验评估显示,与最先进的商业CPU-only分析系统相比,对于无法放入GPU内存的生产数据集上的实际查询,速度提升了数量级。 这项工作为在更广泛的应用场景中采用GPU铺平了道路,并且与其他大多数扩展或回退机制是互补的。
Subjects: Databases (cs.DB)
Cite as: arXiv:2506.10092 [cs.DB]
  (or arXiv:2506.10092v2 [cs.DB] for this version)
  https://doi.org/10.48550/arXiv.2506.10092
arXiv-issued DOI via DataCite

Submission history

From: Rathijit Sen [view email]
[v1] Wed, 11 Jun 2025 18:24:11 UTC (739 KB)
[v2] Wed, 3 Sep 2025 21:07:56 UTC (1,179 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled
  • View Chinese PDF
  • View PDF
  • HTML (experimental)
  • TeX Source
  • Other Formats
view license
Current browse context:
cs.DB
< prev   |   next >
new | recent | 2025-06
Change to browse by:
cs

References & Citations

  • NASA ADS
  • Google Scholar
  • Semantic Scholar
a export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo

Bibliographic and Citation Tools

Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)

Code, Data and Media Associated with this Article

alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)

Demos

Replicate (What is Replicate?)
Hugging Face Spaces (What is Spaces?)
TXYZ.AI (What is TXYZ.AI?)

Recommenders and Search Tools

Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
IArxiv Recommender (What is IArxiv?)
  • Author
  • Venue
  • Institution
  • Topic

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)
  • About
  • Help
  • contact arXivClick here to contact arXiv Contact
  • subscribe to arXiv mailingsClick here to subscribe Subscribe
  • Copyright
  • Privacy Policy
  • Web Accessibility Assistance
  • arXiv Operational Status
    Get status notifications via email or slack

京ICP备2025123034号