July 18, 2021
Many inference problems, such as sequential decision problems like A/B testing, adaptive sampling schemes like bandit selection, are often on- line in nature. The fundamental problem for on- line inference is to provide a sequence of confidence intervals that are valid uniformly over the growing-into-infinity sample sizes. To address this question, we provide a near-optimal confidence sequence for bounded random variables by utilizing Bentkus’ concentration results. We show that it improves on the existing approaches that use the Cramer-Chernoff technique such as the Hoeffding, Bernstein, and Bennett inequalities. The resulting confidence sequence is confirmed to be favorable in synthetic coverage problems, adaptive stopping algorithms, and multi-armed bandit problems.
Written by
Arun Kumar Kuchibhotla
Qinqing Zheng
Publisher
ICML 2021
Research Topics
Core Machine Learning
Theory
November 08, 2022
Ari Morcos, Shashank Shekhar, Surya Ganguli, Ben Sorscher, Robert Geirhos
November 08, 2022
November 30, 2020
Nicolas Usunier, Clément Calauzènes
November 30, 2020
November 30, 2020
Rama Vedantam, David Schwab, Douwe Kiela, Yann Dubois
November 30, 2020
May 03, 2019
Sho Yaida
May 03, 2019
May 08, 2019
May 08, 2019
March 12, 2018
Leon Bottou, Martin Arjovsky, David Lopez-Paz, Maxime Oquab
March 12, 2018
April 30, 2018
Hongyi Zhang, Moustapha Cisse, Yann Dauphin, David Lopez-Paz
April 30, 2018
June 09, 2019
Vikas Verma, Alex Lamb, Christopher Beckham, Amir Najafi, Ioannis Mitliagkas, David Lopez-Paz, Yoshua Bengio
June 09, 2019
Foundational models
Latest news
Foundational models