From Average Sensitivity to Small-Loss Regret Bounds under Random-Order Model

From Average Sensitivity to Small-Loss Regret Bounds under Random-Order Model
Notice: This research summary and analysis were automatically generated using AI technology. For absolute accuracy, please refer to the [Original Paper Viewer] below or the Original ArXiv Source.

We study online learning in the random-order model, where the multiset of loss functions is chosen adversarially but revealed in a uniformly random order. Building on the batch-to-online conversion by Dong and Yoshida (2023), we show that if an offline algorithm admits a $(1+\varepsilon)$-approximation guarantee and the effect of $\varepsilon$ on its average sensitivity is characterized by a function $φ(\varepsilon)$, then an adaptive choice of $\varepsilon$ yields a small-loss regret bound of $\tilde O(φ^{\star}(\mathrm{OPT}_T))$, where $φ^{\star}$ is the concave conjugate of $φ$, $\mathrm{OPT}_T$ is the offline optimum over $T$ rounds, and $\tilde O$ hides polylogarithmic factors in $T$. Our method requires no regularity assumptions on loss functions, such as smoothness, and can be viewed as a generalization of the AdaGrad-style tuning applied to the approximation parameter $\varepsilon$. Our result recovers and strengthens the $(1+\varepsilon)$-approximate regret bounds of Dong and Yoshida (2023) and yields small-loss regret bounds for online $k$-means clustering, low-rank approximation, and regression. We further apply our framework to online submodular function minimization using $(1\pm\varepsilon)$-cut sparsifiers of submodular hypergraphs, obtaining a small-loss regret bound of $\tilde O(n^{3/4}(1 + \mathrm{OPT}_T^{3/4}))$, where $n$ is the ground-set size. Our approach sheds light on the power of sparsification and related techniques in establishing small-loss regret bounds in the random-order model.


💡 Research Summary

The paper studies online learning in the random‑order model, where an adversary chooses a multiset of loss functions in advance but the order of arrival is uniformly random. Building on the batch‑to‑online conversion introduced by Dong and Yoshida (2023), the authors show that if an offline algorithm provides a (1 + ε)‑approximation and its average sensitivity (the expected total‑variation distance between outputs on datasets that differ by a uniformly random deletion) scales as β(t, ε) = φ(ε)/(2t) for some non‑increasing function φ, then an adaptive choice of ε at each round yields a regret bound of
\


Comments & Academic Discussion

Loading comments...

Leave a Comment