• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations

Tools

Sorted by:
Try your query at:
Semantic Scholar Scholar Academic
Google Bing DBLP
Results 1 - 10 of 10,011
Next 10 →

Observable Markov Decision Process

by Robert J. Wood, Mark P. Woodward
"... (Article begins on next page) The Harvard community has made this article openly available. Please share how this access benefits you. Your story matters. ..."
Abstract - Add to MetaCart
(Article begins on next page) The Harvard community has made this article openly available. Please share how this access benefits you. Your story matters.

Partially Observable Markov Decision Processes

by Trung H. Bui, Boris Van Schooten, Dennis Hofs
"... Partially Observable Markov Decision Processes (POMDPs) are attractive for dialogue management because they are made to deal with noise and partial information. This paper addresses the problem of using them in a practical development cycle. We apply factored POMDP models to three applications. We e ..."
Abstract - Add to MetaCart
Partially Observable Markov Decision Processes (POMDPs) are attractive for dialogue management because they are made to deal with noise and partial information. This paper addresses the problem of using them in a practical development cycle. We apply factored POMDP models to three applications. We

Algorithms for Partially Observable Markov Decision Processes

by Weihong Zhang - HONG KONG UNIVERSITY OF SCIENCE AND TECHNOLOGY , 2001
"... Partially Observable Markov Decision Process (POMDP) is a general sequential decision-making model where the effects of actions are... ..."
Abstract - Cited by 21 (1 self) - Add to MetaCart
Partially Observable Markov Decision Process (POMDP) is a general sequential decision-making model where the effects of actions are...

Partially-Observable Markov Decision Processes

by Tom Erez, William D. Smart
"... Partially-Observable Markov Decision Processes (POMDPs) are typically solved by finding an approximate global solution to a corresponding belief-MDP. In this paper, we offer a new method to solve POMDPs with continuous state, action and observation spaces. Since such domains have an inherent notion ..."
Abstract - Add to MetaCart
Partially-Observable Markov Decision Processes (POMDPs) are typically solved by finding an approximate global solution to a corresponding belief-MDP. In this paper, we offer a new method to solve POMDPs with continuous state, action and observation spaces. Since such domains have an inherent notion

♯-acyclic Partially Observable Markov Decision Processes ⋆

by Hugo Gimbert, Youssouf Oualhadj , 2013
"... Abstract. The value 1 problem is a natural decision problem in algorithmic game theory. For partially observable Markov decision processes with reachability objective, this problem is defined as follows: are there observational strategies that achieve the reachability objective with probability arbi ..."
Abstract - Add to MetaCart
Abstract. The value 1 problem is a natural decision problem in algorithmic game theory. For partially observable Markov decision processes with reachability objective, this problem is defined as follows: are there observational strategies that achieve the reachability objective with probability

Transition Entropy in Partially Observable Markov Decision Processes

by Francisco S. Melo, M. Isabel Ribeiro - Proceedings of the 9th International Conference on Intelligent Autonomous Systems (IAS-9) , 2006
"... This paper proposes a new heuristic algorithm suitable for real-time applications using partially observable Markov decision processes (POMDP). The algorithm is based in a reward shaping strategy which includes entropy information in the reward structure of a fully observable Markov decision process ..."
Abstract - Cited by 7 (3 self) - Add to MetaCart
This paper proposes a new heuristic algorithm suitable for real-time applications using partially observable Markov decision processes (POMDP). The algorithm is based in a reward shaping strategy which includes entropy information in the reward structure of a fully observable Markov decision

Partially Observable Markov Decision Processes

by Kyle Hollins Wray, Shlomo Zilberstein
"... We parallelize the Point-Based Value Iteration (PBVI) algo-rithm, which approximates the solution to Partially Observ-able Markov Decision Processes (POMDPs), using a Graph-ics Processing Unit (GPU). We detail additional optimiza-tions, such as leveraging the bounded size of non-zero values over all ..."
Abstract - Add to MetaCart
We parallelize the Point-Based Value Iteration (PBVI) algo-rithm, which approximates the solution to Partially Observ-able Markov Decision Processes (POMDPs), using a Graph-ics Processing Unit (GPU). We detail additional optimiza-tions, such as leveraging the bounded size of non-zero values over

Partially Observable Markov Decision Processes

by Yu Nishiyama, Abdeslam Boularias, Arthur Gretton, Kenji Fukumizu
"... A nonparametric approach for policy learn-ing for POMDPs is proposed. The approach represents distributions over the states, ob-servations, and actions as embeddings in feature spaces, which are reproducing ker-nel Hilbert spaces. Distributions over states given the observations are obtained by ap-p ..."
Abstract - Add to MetaCart
A nonparametric approach for policy learn-ing for POMDPs is proposed. The approach represents distributions over the states, ob-servations, and actions as embeddings in feature spaces, which are reproducing ker-nel Hilbert spaces. Distributions over states given the observations are obtained by ap

Partially Observable Markov Decision Processes

by unknown authors
"... ..."
Abstract - Add to MetaCart
Abstract not found

Partially-Observable Markov Decision Processes

by unknown authors
"... ..."
Abstract - Add to MetaCart
Abstract not found
Next 10 →
Results 1 - 10 of 10,011
Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University