In applying reinforcement learning to continuous space problems, discretization or redefinition of the learning space can be a promising approach. Several methods and algorithms have been introduced to learning agents to respond to this problem. In our previous study, we introduced an FCCM clustering technique into Q-learning (called QLFCCM) and its transfer learning in the Markov process. Since we could not respond to complicated environments like a non-Markov process, in this study, we propose a method in which an agent updates his Q-table by changing the trade-off ratio, Q-learning and QL-FCCM, based on the damping ratio. We conducted numerical experiments of the single pendulum standing problem and our model resulted in a smooth learning process.
CITATION STYLE
Notsu, A., Ueno, T., Hattori, Y., Ubukata, S., & Honda, K. (2015). FCM-type co-clustering transfer reinforcement learning for non-Markov processes. In Lecture Notes in Artificial Intelligence (Subseries of Lecture Notes in Computer Science) (Vol. 9376, pp. 214–225). Springer Verlag. https://doi.org/10.1007/978-3-319-25135-6_21
Mendeley helps you to discover research relevant for your work.