Have a personal or library account? Click to login
Bottom-up learning of hierarchical models in a class of deterministic POMDP environments Cover

Bottom-up learning of hierarchical models in a class of deterministic POMDP environments

Open Access
|Sep 2015

Abstract

The theory of partially observable Markov decision processes (POMDPs) is a useful tool for developing various intelligent agents, and learning hierarchical POMDP models is one of the key approaches for building such agents when the environments of the agents are unknown and large. To learn hierarchical models, bottom-up learning methods in which learning takes place in a layer-by-layer manner from the lowest to the highest layer are already extensively used in some research fields such as hidden Markov models and neural networks. However, little attention has been paid to bottom-up approaches for learning POMDP models. In this paper, we present a novel bottom-up learning algorithm for hierarchical POMDP models and prove that, by using this algorithm, a perfect model (i.e., a model that can perfectly predict future observations) can be learned at least in a class of deterministic POMDP environments

DOI: https://doi.org/10.1515/amcs-2015-0044 | Journal eISSN: 2083-8492 | Journal ISSN: 1641-876X
Language: English
Page range: 597 - 615
Submitted on: Apr 24, 2014
Published on: Sep 30, 2015
Published by: University of Zielona Góra
In partnership with: Paradigm Publishing Services
Publication frequency: 4 issues per year

© 2015 Hideaki Itoh, Hisao Fukumoto, Hiroshi Wakuya, Tatsuya Furukawa, published by University of Zielona Góra
This work is licensed under the Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License.