FCM-type co-clustering transfer reinforcement learning for non-Markov processes

0Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In applying reinforcement learning to continuous space problems, discretization or redefinition of the learning space can be a promising approach. Several methods and algorithms have been introduced to learning agents to respond to this problem. In our previous study, we introduced an FCCM clustering technique into Q-learning (called QLFCCM) and its transfer learning in the Markov process. Since we could not respond to complicated environments like a non-Markov process, in this study, we propose a method in which an agent updates his Q-table by changing the trade-off ratio, Q-learning and QL-FCCM, based on the damping ratio. We conducted numerical experiments of the single pendulum standing problem and our model resulted in a smooth learning process.

Cite

CITATION STYLE

APA

Notsu, A., Ueno, T., Hattori, Y., Ubukata, S., & Honda, K. (2015). FCM-type co-clustering transfer reinforcement learning for non-Markov processes. In Lecture Notes in Artificial Intelligence (Subseries of Lecture Notes in Computer Science) (Vol. 9376, pp. 214–225). Springer Verlag. https://doi.org/10.1007/978-3-319-25135-6_21

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free