Hyperparameter Optimization Can Even Be Harmful in Off-Policy Learning and How to Deal with It
Hyperparameter Optimization Can Even Be Harmful in Off-Policy Learning and How to Deal with It
Yuta Saito, Masahiro Nomura
Proceedings of the Thirty-Third International Joint Conference on Artificial Intelligence
Main Track. Pages 4860-4867.
https://doi.org/10.24963/ijcai.2024/537
There has been a growing interest in off-policy evaluation in the literature such as recommender systems and personalized medicine. We have so far seen significant progress in developing estimators aimed at accurately estimating the effectiveness of counterfactual policies based on biased logged data. However, there are many cases where those estimators are used not only to evaluate the value of decision making policies but also to search for the best hyperparameters from a large candidate space. This work explores the latter hyperparameter optimization (HPO) task for off-policy learning. We empirically show that naively applying an unbiased estimator of the generalization performance as a surrogate objective in HPO can cause an unexpected failure, merely pursuing hyperparameters whose generalization performance is greatly overestimated. We then propose simple and computationally efficient corrections to the typical HPO procedure to deal with the aforementioned issues simultaneously. Empirical investigations demonstrate the effectiveness of our proposed HPO algorithm in situations where the typical procedure fails severely.
Keywords:
Machine Learning: ML: Causality
Machine Learning: ML: Hyperparameter optimization
Machine Learning: ML: Multi-armed bandits
Uncertainty in AI: UAI: Causality, structural causal models and causal inference