• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 1,247
Next 10 →

The Irrevocable Multi-Armed Bandit Problem

by unknown authors , 2008
"... This paper considers the multi-armed bandit problem with multiple simultaneous arm pulls and the additional restriction that we do not allow recourse to arms that were pulled at some point in the past but then discarded. This additional restriction is highly desirable from an operational perspective ..."
Abstract - Add to MetaCart
This paper considers the multi-armed bandit problem with multiple simultaneous arm pulls and the additional restriction that we do not allow recourse to arms that were pulled at some point in the past but then discarded. This additional restriction is highly desirable from an operational

MULTI-ARMED BANDIT PROBLEMS

by Aditya Mahajan, Demosthenis Teneketzis
"... Multi-armed bandit (MAB) problems are a class of sequential resource allocation problems concerned with allocating one or more resources among several alternative (competing) projects. Such problems are paradigms of a fundamental conflict between making decisions (allocating resources) that yield ..."
Abstract - Cited by 16 (0 self) - Add to MetaCart
Multi-armed bandit (MAB) problems are a class of sequential resource allocation problems concerned with allocating one or more resources among several alternative (competing) projects. Such problems are paradigms of a fundamental conflict between making decisions (allocating resources) that yield

Multi-Armed Bandit Problem

by Gianpiero Francesca, Giovanni Pini, Arne Brutschy, Marco Dorigo, Mauro Birattari, Université Libre De Bruxelles, Av F. D. Roosevelt, Gianpiero Francesca, Giovanni Pini, Arne Brutschy, Carlo Pinciroli, Marco Dorigo, Mauro Birattari , 2011
"... The information provided is the sole responsibility of the authors and does not necessarily reflect the opinion of the members of IRIDIA. The authors take full responsibility for any copyright breaches that may result from publication of this paper in the IRIDIA – Technical Report Series. IRIDIA is ..."
Abstract - Add to MetaCart
The information provided is the sole responsibility of the authors and does not necessarily reflect the opinion of the members of IRIDIA. The authors take full responsibility for any copyright breaches that may result from publication of this paper in the IRIDIA – Technical Report Series. IRIDIA is not responsible for any use that might be made of

The budgeted multi-armed bandit problem

by Omid Madani, Daniel J. Lizotte, Russell Greiner, N. Pasadena Ave - 2972 KNOWLEDGE GRADIENT FOR SEQUENTIAL SAMPLING Omid , 2004
"... The following coins problem is a version of a multi-armed bandit problem where one has to select from among a set of objects, say classifiers, after an experimentation phase that is constrained by a time or cost budget. The question is how to spend the budget. The problem involves pure exploration o ..."
Abstract - Cited by 9 (1 self) - Add to MetaCart
The following coins problem is a version of a multi-armed bandit problem where one has to select from among a set of objects, say classifiers, after an experimentation phase that is constrained by a time or cost budget. The question is how to spend the budget. The problem involves pure exploration

THE MULTI-ARMED BANDIT PROBLEM WITH COVARIATES

by Vianney Perchet, Philippe Rigollet , 2013
"... We consider a multi-armed bandit problem in a setting where each arm produces a noisy reward realization which depends on an observable random covariate. As opposed to the traditional static multi-armed bandit prob-lem, this setting allows for dynamically changing rewards that better describe applic ..."
Abstract - Cited by 7 (1 self) - Add to MetaCart
We consider a multi-armed bandit problem in a setting where each arm produces a noisy reward realization which depends on an observable random covariate. As opposed to the traditional static multi-armed bandit prob-lem, this setting allows for dynamically changing rewards that better describe

Algorithms for the multi-armed bandit problem

by Volodymyr Kuleshov, Doina Precup - JOURNAL OF MACHINE LEARNING RESEARCH , 2000
"... The stochastic multi-armed bandit problem is an important model for studying the exploration-exploitation tradeoff in reinforcement learning. Although many algorithms for the problem are well-understood theoretically, empirical confirmation of their effectiveness is generally scarce. This paper pres ..."
Abstract - Cited by 8 (0 self) - Add to MetaCart
The stochastic multi-armed bandit problem is an important model for studying the exploration-exploitation tradeoff in reinforcement learning. Although many algorithms for the problem are well-understood theoretically, empirical confirmation of their effectiveness is generally scarce. This paper

Multi-armed Bandit Problems with History

by Pannagadatta Shivaswamy, Thorsten Joachims
"... ..."
Abstract - Cited by 1 (0 self) - Add to MetaCart
Abstract not found

Multi-armed Bandit Problem with Lock-up Periods

by Junpei Komiyama, Issei Sato, Soon Ong, Tu Bao Ho
"... We investigate a stochastic multi-armed bandit problem in which the forecaster’s choice is restricted. In this problem, rounds are divided into lock-up periods and the forecaster must select the same arm throughout a period. While there has been much work on finding optimal algorithms for the stocha ..."
Abstract - Cited by 1 (0 self) - Add to MetaCart
We investigate a stochastic multi-armed bandit problem in which the forecaster’s choice is restricted. In this problem, rounds are divided into lock-up periods and the forecaster must select the same arm throughout a period. While there has been much work on finding optimal algorithms

Pure exploration in multi-armed bandits problems

by Sébastien Bubeck, Rémi Munos , Gilles Stoltz - IN PROCEEDINGS OF THE TWENTIETH INTERNATIONAL CONFERENCE ON ALGORITHMIC LEARNING THEORY (ALT 2009 , 2009
"... We consider the framework of stochastic multi-armed bandit problems and study the possibilities and limitations of strategies that explore sequentially the arms. The strategies are assessed not in terms of their cumulative regrets, as is usually the case, but through quantities referred to as simpl ..."
Abstract - Cited by 80 (13 self) - Add to MetaCart
We consider the framework of stochastic multi-armed bandit problems and study the possibilities and limitations of strategies that explore sequentially the arms. The strategies are assessed not in terms of their cumulative regrets, as is usually the case, but through quantities referred

Analysis of Thompson sampling for the multi-armed bandit problem

by Shipra Agrawal, Navin Goyal - In COLT 2012 , 2012
"... The multi-armed bandit problem is a popular model for studying exploration/exploitation trade-off in sequential decision problems. Many algorithms are now available for this well-studied problem. One of the earliest algorithms, given by W. R. Thompson, dates back to 1933. This algorithm, referred to ..."
Abstract - Cited by 53 (4 self) - Add to MetaCart
The multi-armed bandit problem is a popular model for studying exploration/exploitation trade-off in sequential decision problems. Many algorithms are now available for this well-studied problem. One of the earliest algorithms, given by W. R. Thompson, dates back to 1933. This algorithm, referred
Next 10 →
Results 1 - 10 of 1,247
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University