Skip to main content
CenXiv.org
This website is in trial operation, support us!
We gratefully acknowledge support from all contributors.
Contribute
Donate
cenxiv logo > cs > arXiv:2410.12101v1

Help | Advanced Search

Computer Science > Machine Learning

arXiv:2410.12101v1 (cs)
[Submitted on 15 Oct 2024 (this version) , latest version 22 Oct 2024 (v2) ]

Title: The Persian Rug: solving toy models of superposition using large-scale symmetries

Title: 波斯地毯:使用大规模对称性解决叠加的玩具模型

Authors:Aditya Cowsik, Kfir Dolev, Alex Infanger
Abstract: We present a complete mechanistic description of the algorithm learned by a minimal non-linear sparse data autoencoder in the limit of large input dimension. The model, originally presented in arXiv:2209.10652, compresses sparse data vectors through a linear layer and decompresses using another linear layer followed by a ReLU activation. We notice that when the data is permutation symmetric (no input feature is privileged) large models reliably learn an algorithm that is sensitive to individual weights only through their large-scale statistics. For these models, the loss function becomes analytically tractable. Using this understanding, we give the explicit scalings of the loss at high sparsity, and show that the model is near-optimal among recently proposed architectures. In particular, changing or adding to the activation function any elementwise or filtering operation can at best improve the model's performance by a constant factor. Finally, we forward-engineer a model with the requisite symmetries and show that its loss precisely matches that of the trained models. Unlike the trained model weights, the low randomness in the artificial weights results in miraculous fractal structures resembling a Persian rug, to which the algorithm is oblivious. Our work contributes to neural network interpretability by introducing techniques for understanding the structure of autoencoders. Code to reproduce our results can be found at https://github.com/KfirD/PersianRug .
Abstract: 我们提供了一个最小非线性稀疏数据自编码器在输入维度大的极限下的算法的完整机制描述。 该模型最初在arXiv:2209.10652中提出,通过一个线性层压缩稀疏数据向量,并使用另一个线性层后接ReLU激活函数进行解压缩。 我们注意到,当数据是排列对称的(没有输入特征是优先的)时,大模型可靠地学习一种仅通过其大规模统计量对单个权重敏感的算法。 对于这些模型,损失函数变得解析可处理。 利用这种理解,我们给出了高稀疏性下损失的显式缩放关系,并表明该模型在最近提出的架构中接近最优。 特别是,对激活函数进行任何逐元素或过滤操作的更改或添加,最多只能以常数因子改善模型性能。 最后,我们正向工程了一个具有必要对称性的模型,并表明其损失精确匹配训练模型的损失。 与训练模型的权重不同,人工权重中的低随机性导致了类似于波斯地毯的奇迹般的分形结构,而算法对此是无意识的。 我们的工作通过引入理解自编码器结构的技术,为神经网络的可解释性做出了贡献。 可以找到重现我们结果的代码https://github.com/KfirD/PersianRug 。
Subjects: Machine Learning (cs.LG) ; Disordered Systems and Neural Networks (cond-mat.dis-nn); Artificial Intelligence (cs.AI)
Cite as: arXiv:2410.12101 [cs.LG]
  (or arXiv:2410.12101v1 [cs.LG] for this version)
  https://doi.org/10.48550/arXiv.2410.12101
arXiv-issued DOI via DataCite

Submission history

From: Aditya Cowsik [view email]
[v1] Tue, 15 Oct 2024 22:52:45 UTC (3,878 KB)
[v2] Tue, 22 Oct 2024 17:48:56 UTC (8,034 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled
  • View Chinese PDF
  • View PDF
  • HTML (experimental)
  • TeX Source
license icon view license
Current browse context:
cs.LG
< prev   |   next >
new | recent | 2024-10
Change to browse by:
cond-mat
cond-mat.dis-nn
cs
cs.AI

References & Citations

  • NASA ADS
  • Google Scholar
  • Semantic Scholar
a export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo

Bibliographic and Citation Tools

Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)

Code, Data and Media Associated with this Article

alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)

Demos

Replicate (What is Replicate?)
Hugging Face Spaces (What is Spaces?)
TXYZ.AI (What is TXYZ.AI?)

Recommenders and Search Tools

Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
IArxiv Recommender (What is IArxiv?)
  • Author
  • Venue
  • Institution
  • Topic

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)
  • About
  • Help
  • contact arXivClick here to contact arXiv Contact
  • subscribe to arXiv mailingsClick here to subscribe Subscribe
  • Copyright
  • Privacy Policy
  • Web Accessibility Assistance
  • arXiv Operational Status
    Get status notifications via email or slack

京ICP备2025123034号