Skip to main content
CenXiv.org
This website is in trial operation, support us!
We gratefully acknowledge support from all contributors.
Contribute
Donate
cenxiv logo > cs > arXiv:2506.23286

Help | Advanced Search

Computer Science > Machine Learning

arXiv:2506.23286 (cs)
[Submitted on 29 Jun 2025 ]

Title: Not All Explanations for Deep Learning Phenomena Are Equally Valuable

Title: 并非所有对深度学习现象的解释都具有同等价值

Authors:Alan Jeffares, Mihaela van der Schaar
Abstract: Developing a better understanding of surprising or counterintuitive phenomena has constituted a significant portion of deep learning research in recent years. These include double descent, grokking, and the lottery ticket hypothesis -- among many others. Works in this area often develop ad hoc hypotheses attempting to explain these observed phenomena on an isolated, case-by-case basis. This position paper asserts that, in many prominent cases, there is little evidence to suggest that these phenomena appear in real-world applications and these efforts may be inefficient in driving progress in the broader field. Consequently, we argue against viewing them as isolated puzzles that require bespoke resolutions or explanations. However, despite this, we suggest that deep learning phenomena do still offer research value by providing unique settings in which we can refine our broad explanatory theories of more general deep learning principles. This position is reinforced by analyzing the research outcomes of several prominent examples of these phenomena from the recent literature. We revisit the current norms in the research community in approaching these problems and propose practical recommendations for future research, aiming to ensure that progress on deep learning phenomena is well aligned with the ultimate pragmatic goal of progress in the broader field of deep learning.
Abstract: 近年来,深入理解令人惊讶或违反直觉的现象已成为深度学习研究的重要组成部分。 这些包括双重下降、领悟和彩票票根假设——还有许多其他现象。 该领域的研究通常会提出临时假设,试图以孤立的、逐个案例的方式解释这些观察到的现象。 本文认为,在许多显著的情况下,几乎没有证据表明这些现象出现在现实应用中,这些努力可能在推动更广泛领域进展方面效率低下。 因此,我们反对将它们视为需要定制解决方案或解释的孤立谜题。 然而,尽管如此,我们认为深度学习现象仍然具有研究价值,因为它们提供了独特的环境,使我们能够完善对更一般的深度学习原理的广泛解释理论。 通过对近期文献中这些现象的一些显著例子的研究成果进行分析,进一步支持了这一观点。 我们重新审视了研究社区在处理这些问题时的当前规范,并提出了未来研究的实用建议,旨在确保深度学习现象的进展与更广泛深度学习领域进步的最终实际目标保持一致。
Comments: Accepted at ICML 2025 for oral presentation
Subjects: Machine Learning (cs.LG) ; Artificial Intelligence (cs.AI); Machine Learning (stat.ML)
Cite as: arXiv:2506.23286 [cs.LG]
  (or arXiv:2506.23286v1 [cs.LG] for this version)
  https://doi.org/10.48550/arXiv.2506.23286
arXiv-issued DOI via DataCite

Submission history

From: Alan Jeffares [view email]
[v1] Sun, 29 Jun 2025 15:18:56 UTC (104 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled
  • View Chinese PDF
  • View PDF
  • HTML (experimental)
  • TeX Source
license icon view license
Current browse context:
cs.LG
< prev   |   next >
new | recent | 2025-06
Change to browse by:
cs
cs.AI
stat
stat.ML

References & Citations

  • NASA ADS
  • Google Scholar
  • Semantic Scholar
a export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo

Bibliographic and Citation Tools

Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)

Code, Data and Media Associated with this Article

alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)

Demos

Replicate (What is Replicate?)
Hugging Face Spaces (What is Spaces?)
TXYZ.AI (What is TXYZ.AI?)

Recommenders and Search Tools

Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
IArxiv Recommender (What is IArxiv?)
  • Author
  • Venue
  • Institution
  • Topic

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)
  • About
  • Help
  • contact arXivClick here to contact arXiv Contact
  • subscribe to arXiv mailingsClick here to subscribe Subscribe
  • Copyright
  • Privacy Policy
  • Web Accessibility Assistance
  • arXiv Operational Status
    Get status notifications via email or slack

京ICP备2025123034号