How to gamble with non-stationary X-armed bandits and have no regrets

Loading...
Thumbnail Image
Date
2020
Volume
2686
Issue
Journal
Series Titel
WIAS Preprints
Book Title
Publisher
Berlin : Weierstraß-Institut für Angewandte Analysis und Stochastik
Abstract

In X-armed bandit problem an agent sequentially interacts with environment which yields a reward based on the vector input the agent provides. The agent's goal is to maximise the sum of these rewards across some number of time steps. The problem and its variations have been a subject of numerous studies, suggesting sub-linear and sometimes optimal strategies. The given paper introduces a new variation of the problem. We consider an environment, which can abruptly change its behaviour an unknown number of times. To that end we propose a novel strategy and prove it attains sub-linear cumulative regret. Moreover, the obtained regret bound matches the best known bound for GP-UCB for a stationary case, and approaches the minimax lower bound in case of highly smooth relation between an action and the corresponding reward. The theoretical result is supported by experimental study.

Description
Keywords
Citation
Avanesov, V. (2020). How to gamble with non-stationary X-armed bandits and have no regrets (Berlin : Weierstraß-Institut für Angewandte Analysis und Stochastik). Berlin : Weierstraß-Institut für Angewandte Analysis und Stochastik. https://doi.org//10.20347/WIAS.PREPRINT.2686
License
This document may be downloaded, read, stored and printed for your own use within the limits of § 53 UrhG but it may not be distributed via the internet or passed on to external parties.
Dieses Dokument darf im Rahmen von § 53 UrhG zum eigenen Gebrauch kostenfrei heruntergeladen, gelesen, gespeichert und ausgedruckt, aber nicht im Internet bereitgestellt oder an Außenstehende weitergegeben werden.