Reinforcement Learning Using State Space Compression


The Transactions of the Korea Information Processing Society (1994 ~ 2000), Vol. 6, No. 3, pp. 633-640, Mar. 1999
10.3745/KIPSTE.1999.6.3.633,   PDF Download:

Abstract

Reinforcement learning performs learning through interacting with trial-and-error in dynamic environment. Therefore, in dynamic environment, reinforcement learning method like Q-learning and TD(Temporal Difference)-learning are faster in learning than the conventional stochastic learning method. However, because many of the proposed reinforcement learning algorithms are given the reinforcement value only when the learning agent has reached its goals state, most of the reinforcement algorithms are given the reinforcement value only when the learning agent has reached its goal state, most of the reinforcement algorithms converge to the optimal solution too slowly. In this paper, we present COMREL(COMpressed REinforcement Learning) algorithm for finding the shortest path fast in a maze environment. COMREL compress the given maze environment, select the candidate states that can guide the shortest path in compressed maze environment, and learn only the candidate states to find the shortest path. After comparing COMREL algorithm with the already existing Q-learning and Priortized Sweeping algorithm, we could see that the learning time shortened very much.


Statistics
Show / Hide Statistics

Statistics (Cumulative Counts from September 1st, 2017)
Multiple requests among the same browser session are counted as one view.
If you mouse over a chart, the values of data points will be shown.


Cite this article
[IEEE Style]
K. B. Cheon and Y. B. Joo, "Reinforcement Learning Using State Space Compression," The Transactions of the Korea Information Processing Society (1994 ~ 2000), vol. 6, no. 3, pp. 633-640, 1999. DOI: 10.3745/KIPSTE.1999.6.3.633.

[ACM Style]
Kim Byung Cheon and Yoon Byung Joo. 1999. Reinforcement Learning Using State Space Compression. The Transactions of the Korea Information Processing Society (1994 ~ 2000), 6, 3, (1999), 633-640. DOI: 10.3745/KIPSTE.1999.6.3.633.