Skip to main content
CenXiv.org
This website is in trial operation, support us!
We gratefully acknowledge support from all contributors.
Contribute
Donate
cenxiv logo > cs > arXiv:2504.15251

Help | Advanced Search

Computer Science > Machine Learning

arXiv:2504.15251 (cs)
[Submitted on 21 Apr 2025 ]

Title: On Learning Parallel Pancakes with Mostly Uniform Weights

Title: 关于主要具有均匀权重的平行煎饼学习

Authors:Ilias Diakonikolas, Daniel M. Kane, Sushrut Karmalkar, Jasper C.H. Lee, Thanasis Pittas
Abstract: We study the complexity of learning $k$-mixtures of Gaussians ($k$-GMMs) on $\mathbb{R}^d$. This task is known to have complexity $d^{\Omega(k)}$ in full generality. To circumvent this exponential lower bound on the number of components, research has focused on learning families of GMMs satisfying additional structural properties. A natural assumption posits that the component weights are not exponentially small and that the components have the same unknown covariance. Recent work gave a $d^{O(\log(1/w_{\min}))}$-time algorithm for this class of GMMs, where $w_{\min}$ is the minimum weight. Our first main result is a Statistical Query (SQ) lower bound showing that this quasi-polynomial upper bound is essentially best possible, even for the special case of uniform weights. Specifically, we show that it is SQ-hard to distinguish between such a mixture and the standard Gaussian. We further explore how the distribution of weights affects the complexity of this task. Our second main result is a quasi-polynomial upper bound for the aforementioned testing task when most of the weights are uniform while a small fraction of the weights are potentially arbitrary.
Abstract: 我们研究了在$\mathbb{R}^d$上学习$k$-混合高斯分布($k$-GMMs)的复杂性。此任务在完全一般情况下已知具有$d^{\Omega(k)}$的复杂度。为了绕过对组件数量的指数下界,研究集中于学习满足额外结构特性的GMM族。一个自然的假设认为组件权重不是指数小的,并且组件具有相同的未知协方差。近期的工作为此类GMM提供了一个运行时间为$d^{O(\log(1/w_{\min}))}$的算法,其中$w_{\min}$是最小权重。我们的第一个主要结果是统计查询(SQ)下界,表明这个准多项式上界本质上是最好的,即使对于均匀权重的特殊情况也是如此。具体来说,我们证明了区分这种混合与标准高斯分布是SQ困难的。我们进一步探讨了权重分布如何影响此任务的复杂性。我们的第二个主要结果是当大多数权重是均匀的而一小部分权重可能是任意时,针对上述测试任务的准多项式上界。
Subjects: Machine Learning (cs.LG) ; Data Structures and Algorithms (cs.DS); Statistics Theory (math.ST); Machine Learning (stat.ML)
Cite as: arXiv:2504.15251 [cs.LG]
  (or arXiv:2504.15251v1 [cs.LG] for this version)
  https://doi.org/10.48550/arXiv.2504.15251
arXiv-issued DOI via DataCite

Submission history

From: Thanasis Pittas [view email]
[v1] Mon, 21 Apr 2025 17:31:55 UTC (39 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled
  • View Chinese PDF
  • View PDF
  • HTML (experimental)
  • TeX Source
  • Other Formats
view license
Current browse context:
cs.LG
< prev   |   next >
new | recent | 2025-04
Change to browse by:
cs
cs.DS
math
math.ST
stat
stat.ML
stat.TH

References & Citations

  • NASA ADS
  • Google Scholar
  • Semantic Scholar
a export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo

Bibliographic and Citation Tools

Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)

Code, Data and Media Associated with this Article

alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)

Demos

Replicate (What is Replicate?)
Hugging Face Spaces (What is Spaces?)
TXYZ.AI (What is TXYZ.AI?)

Recommenders and Search Tools

Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
IArxiv Recommender (What is IArxiv?)
  • Author
  • Venue
  • Institution
  • Topic

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)
  • About
  • Help
  • contact arXivClick here to contact arXiv Contact
  • subscribe to arXiv mailingsClick here to subscribe Subscribe
  • Copyright
  • Privacy Policy
  • Web Accessibility Assistance
  • arXiv Operational Status
    Get status notifications via email or slack

京ICP备2025123034号