共 50 条
- [31] Optimal and Adaptive Off-policy Evaluation in Contextual Bandits INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
- [32] Optimal Baseline Corrections for Off-Policy Contextual Bandits PROCEEDINGS OF THE EIGHTEENTH ACM CONFERENCE ON RECOMMENDER SYSTEMS, RECSYS 2024, 2024, : 722 - 732
- [33] Beyond UCB: Optimal and Efficient Contextual Bandits with Regression Oracles INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
- [34] Beyond UCB: Optimal and Efficient Contextual Bandits with Regression Oracles 25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,
- [35] Near Instance Optimal Model Selection for Pure Exploration Linear Bandits INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
- [37] Optimal Algorithms for Multiplayer Multi-Armed Bandits INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108
- [38] Optimal Streaming Algorithms for Multi-Armed Bandits INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
- [39] Unimodal Bandits: Regret Lower Bounds and Optimal Algorithms INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 1), 2014, 32
- [40] Breaking the √T Barrier: Instance-Independent Logarithmic Regret in Stochastic Contextual Linear Bandits INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,