共 50 条
- [1] Minimax-Optimal Off-Policy Evaluation with Linear Function Approximation [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
- [2] Variance-Aware Off-Policy Evaluation with Linear Function Approximation [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
- [3] Average-Reward Off-Policy Policy Evaluation with Function Approximation [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
- [4] The Optimal Approximation Factors in Misspecified Off-Policy Value Function Estimation [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202 : 768 - 790
- [5] Weighted importance sampling for off-policy learning with linear function approximation [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 27 (NIPS 2014), 2014, 27
- [6] Minimax Value Interval for Off-Policy Evaluation and Policy Optimization [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
- [7] On Well-posedness and Minimax Optimal Rates of Nonparametric Q-function Estimation in Off-policy Evaluation [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
- [8] On the role of overparameterization in off-policy Temporal Difference learning with linear function approximation [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
- [10] Generalized gradient emphasis learning for off-policy evaluation and control with function approximation [J]. Neural Computing and Applications, 2023, 35 : 23599 - 23616