Skip to main content
CenXiv.org
This website is in trial operation, support us!
We gratefully acknowledge support from all contributors.
Contribute
Donate
cenxiv logo > cs > arXiv:2506.00528v1

Help | Advanced Search

Computer Science > Machine Learning

arXiv:2506.00528v1 (cs)
[Submitted on 31 May 2025 ]

Title: Ultra-Quantisation: Efficient Embedding Search via 1.58-bit Encodings

Title: 超量化:通过1.58位编码的高效嵌入搜索

Authors:Richard Connor, Alan Dearle, Ben Claydon
Abstract: Many modern search domains comprise high-dimensional vectors of floating point numbers derived from neural networks, in the form of embeddings. Typical embeddings range in size from hundreds to thousands of dimensions, making the size of the embeddings, and the speed of comparison, a significant issue. Quantisation is a class of mechanism which replaces the floating point values with a smaller representation, for example a short integer. This gives an approximation of the embedding space in return for a smaller data representation and a faster comparison function. Here we take this idea almost to its extreme: we show how vectors of arbitrary-precision floating point values can be replaced by vectors whose elements are drawn from the set {-1,0,1}. This yields very significant savings in space and metric evaluation cost, while maintaining a strong correlation for similarity measurements. This is achieved by way of a class of convex polytopes which exist in the high-dimensional space. In this article we give an outline description of these objects, and show how they can be used for the basis of such radical quantisation while maintaining a surprising degree of accuracy.
Abstract: 许多现代搜索领域包含来自神经网络的高维浮点数向量,以嵌入的形式表示。 典型的嵌入大小范围从几百到几千个维度,这使得嵌入的大小以及比较的速度成为一个重要的问题。 量化是一种机制类别,它用更小的表示(例如短整型)替换浮点值。 这样可以以较小的数据表示和更快的比较函数为代价,近似表示嵌入空间。 在这里我们将这个想法几乎推向了极端:我们展示了如何用元素从集合{-1,0,1}中抽取的向量来替代任意精度浮点值的向量。 这种方法在空间和度量评估成本上带来了非常显著的节省,同时保持了相似性测量的强相关性。 这是通过存在于高维空间中的凸多面体类实现的。 在本文中,我们给出了这些对象的大致描述,并展示了它们如何作为这种激进量化的基础,同时保持令人惊讶的准确性。
Comments: Submitted to SISAP25 International Conference on Similarity Search and Applications
Subjects: Machine Learning (cs.LG) ; Databases (cs.DB)
Cite as: arXiv:2506.00528 [cs.LG]
  (or arXiv:2506.00528v1 [cs.LG] for this version)
  https://doi.org/10.48550/arXiv.2506.00528
arXiv-issued DOI via DataCite

Submission history

From: Richard Connor [view email]
[v1] Sat, 31 May 2025 12:22:24 UTC (828 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled
  • View Chinese PDF
  • View PDF
  • HTML (experimental)
  • TeX Source
  • Other Formats
license icon view license
Current browse context:
cs.LG
< prev   |   next >
new | recent | 2025-06
Change to browse by:
cs
cs.DB

References & Citations

  • NASA ADS
  • Google Scholar
  • Semantic Scholar
a export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo

Bibliographic and Citation Tools

Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)

Code, Data and Media Associated with this Article

alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)

Demos

Replicate (What is Replicate?)
Hugging Face Spaces (What is Spaces?)
TXYZ.AI (What is TXYZ.AI?)

Recommenders and Search Tools

Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
IArxiv Recommender (What is IArxiv?)
  • Author
  • Venue
  • Institution
  • Topic

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)
  • About
  • Help
  • contact arXivClick here to contact arXiv Contact
  • subscribe to arXiv mailingsClick here to subscribe Subscribe
  • Copyright
  • Privacy Policy
  • Web Accessibility Assistance
  • arXiv Operational Status
    Get status notifications via email or slack

京ICP备2025123034号