Search Results for author: Matej Cief

Found 3 papers, 2 papers with code

Cross-Validated Off-Policy Evaluation

1 code implementation24 May 2024 Matej Cief, Branislav Kveton, Michal Kompan

In this paper, we study the problem of estimator selection and hyper-parameter tuning in off-policy evaluation.

Model Selection Off-policy evaluation

Learning Action Embeddings for Off-Policy Evaluation

1 code implementation6 May 2023 Matej Cief, Jacek Golebiowski, Philipp Schmidt, Ziawasch Abedjan, Artur Bekasov

Off-policy evaluation (OPE) methods allow us to compute the expected reward of a policy by using the logged data collected by a different policy.

Off-policy evaluation

Pessimistic Off-Policy Optimization for Learning to Rank

no code implementations6 Jun 2022 Matej Cief, Branislav Kveton, Michal Kompan

Off-policy learning is a framework for optimizing policies without deploying them, using data collected by another policy.

Learning-To-Rank Recommendation Systems

Cannot find the paper you are looking for? You can Submit a new open access paper.