Rademacher learning rates for iterated random functions
Most supervised learning methods assume training data is drawn from an i.i.d. sample. However, real-world problems often exhibit temporal dependence and strong correlations between marginals of the data-generating process, rendering the i.i.d. assumption unrealistic. Such cases naturally involve tim...
Saved in:
| Published in | Journal of Complexity Vol. 91; p. 101971 |
|---|---|
| Main Author | |
| Format | Journal Article |
| Language | English |
| Published |
Elsevier Inc
01.12.2025
|
| Subjects | |
| Online Access | Get full text |
| ISSN | 0885-064X |
| DOI | 10.1016/j.jco.2025.101971 |
Cover
| Summary: | Most supervised learning methods assume training data is drawn from an i.i.d. sample. However, real-world problems often exhibit temporal dependence and strong correlations between marginals of the data-generating process, rendering the i.i.d. assumption unrealistic. Such cases naturally involve time-series processes and Markov chains. The learning rates typically obtained in these settings remain independent of the data distribution, potentially leading to restrictive hypothesis classes and suboptimal sample complexities. We consider training data generated by an iterated random function that need not be irreducible or aperiodic. Assuming the governing function is contractive in its first argument and subject to certain regularity conditions on the hypothesis class, we first establish uniform convergence for the sample error. We then prove learnability of approximate empirical risk minimization and derive its learning rate bound. Both bounds depend explicitly on the data distribution through the Rademacher complexities of the hypothesis class, thereby better capturing properties of the data-generating distribution. |
|---|---|
| ISSN: | 0885-064X |
| DOI: | 10.1016/j.jco.2025.101971 |