Skip to main content
CenXiv.org
This website is in trial operation, support us!
We gratefully acknowledge support from all contributors.
Contribute
Donate
cenxiv logo > cs > arXiv:2106.00115

Help | Advanced Search

Computer Science > Machine Learning

arXiv:2106.00115 (cs)
[Submitted on 31 May 2021 ]

Title: Fine-grained Generalization Analysis of Structured Output Prediction

Title: 结构化输出预测的细粒度泛化分析

Authors:Waleed Mustafa, Yunwen Lei, Antoine Ledent, Marius Kloft
Abstract: In machine learning we often encounter structured output prediction problems (SOPPs), i.e. problems where the output space admits a rich internal structure. Application domains where SOPPs naturally occur include natural language processing, speech recognition, and computer vision. Typical SOPPs have an extremely large label set, which grows exponentially as a function of the size of the output. Existing generalization analysis implies generalization bounds with at least a square-root dependency on the cardinality $d$ of the label set, which can be vacuous in practice. In this paper, we significantly improve the state of the art by developing novel high-probability bounds with a logarithmic dependency on $d$. Moreover, we leverage the lens of algorithmic stability to develop generalization bounds in expectation without any dependency on $d$. Our results therefore build a solid theoretical foundation for learning in large-scale SOPPs. Furthermore, we extend our results to learning with weakly dependent data.
Abstract: 在机器学习中,我们经常遇到结构化输出预测问题 (SOPPs),即输出空间具有丰富内部结构的问题。SOPPs 自然出现的应用领域包括自然语言处理、语音识别和计算机视觉。典型的 SOPPs 具有极其庞大的标签集,该标签集的大小随着输出规模呈指数增长。现有的泛化分析表明,泛化界至少依赖于标签集基数 $d$ 的平方根,这在实际应用中可能没有意义。本文中,我们通过开发具有对数依赖性于 $d$ 的高概率界,显著改进了当前的技术水平。此外,我们利用算法稳定性视角,在没有任何依赖于 $d$ 的情况下,发展出期望下的泛化界。因此,我们的成果为大规模 SOPPs 学习奠定了坚实的理论基础。此外,我们将研究扩展到弱相关数据的学习中。
Comments: To appearn in IJCAI 2021
Subjects: Machine Learning (cs.LG) ; Machine Learning (stat.ML)
Cite as: arXiv:2106.00115 [cs.LG]
  (or arXiv:2106.00115v1 [cs.LG] for this version)
  https://doi.org/10.48550/arXiv.2106.00115
arXiv-issued DOI via DataCite

Submission history

From: Waleed Mustafa [view email]
[v1] Mon, 31 May 2021 21:44:14 UTC (59 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled
  • View Chinese PDF
  • View PDF
  • TeX Source
view license
Current browse context:
cs
< prev   |   next >
new | recent | 2021-06
Change to browse by:
cs.LG
stat
stat.ML

References & Citations

  • NASA ADS
  • Google Scholar
  • Semantic Scholar
a export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo

Bibliographic and Citation Tools

Bibliographic Explorer (What is the Explorer?)
Connected Papers (What is Connected Papers?)
Litmaps (What is Litmaps?)
scite Smart Citations (What are Smart Citations?)

Code, Data and Media Associated with this Article

alphaXiv (What is alphaXiv?)
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub (What is DagsHub?)
Gotit.pub (What is GotitPub?)
Hugging Face (What is Huggingface?)
Papers with Code (What is Papers with Code?)
ScienceCast (What is ScienceCast?)

Demos

Replicate (What is Replicate?)
Hugging Face Spaces (What is Spaces?)
TXYZ.AI (What is TXYZ.AI?)

Recommenders and Search Tools

Influence Flower (What are Influence Flowers?)
CORE Recommender (What is CORE?)
IArxiv Recommender (What is IArxiv?)
  • Author
  • Venue
  • Institution
  • Topic

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.

Which authors of this paper are endorsers? | Disable MathJax (What is MathJax?)
  • About
  • Help
  • contact arXivClick here to contact arXiv Contact
  • subscribe to arXiv mailingsClick here to subscribe Subscribe
  • Copyright
  • Privacy Policy
  • Web Accessibility Assistance
  • arXiv Operational Status
    Get status notifications via email or slack

京ICP备2025123034号