搜索结果: 1-3 共查到“数学 Multi-Armed Bandit”相关记录3条 . 查询时间(0.093 秒)
On the Combinatorial Multi-Armed Bandit Problem with Markovian Rewards
Combinatorial Multi-Armed Bandit Problem Markovian Rewards
2011/1/21
We consider a combinatorial generalization of the classical multi-armed bandit problem that is defined as follows.There is a given bipartite graph of M users and N M resources.
Learning in A Changing World: Non-Bayesian Restless Multi-Armed Bandit
Non-Bayesian Multi-Armed Bandit
2010/11/24
We consider the restless multi-armed bandit (RMAB) problem with unknown dynamics. In this problem, at each time, a player chooses K out of N (N > K) arms to play. The state of each arm determines the ...
The Non-Bayesian Restless Multi-Armed Bandit: a Case of Near-Logarithmic Regret
The Non-Bayesian Restless Multi-Armed Bandit:Near-Logarithmic Regret
2010/11/24
In the classic Bayesian restless multi-armed bandit (RMAB) problem, there are $N$ arms, with rewards on all arms evolving at each time as Markov chains with known parameters. A player seeks to activa...