Accession Number : AD1024580


Title :   Characterizing Reinforcement Learning Methods through Parameterized Learning Problems


Descriptive Note : Journal Article - Open Access


Corporate Author : University of Texas at Austin Austin United States


Personal Author(s) : Kalyanakrishnan,Shivaram ; Stone,Peter


Full Text : https://apps.dtic.mil/dtic/tr/fulltext/u2/1024580.pdf


Report Date : 03 Jun 2011


Pagination or Media Count : 47


Abstract : The field of reinforcement learning (RL) has been energized in the past few decades by elegant theoretical results indicating under what conditions, and how quickly, certain algorithms are guaranteed to converge to optimal policies. However, in practical problems, these conditions are seldom met. When we cannot achieve optimality, the performance of RL algorithms must be measured empirically. Consequently, in order to meaningfully differentiate learning methods, it becomes necessary to characterize their performance on different problems, taking into account factors such as state estimation, exploration, function approximation, and constraints on computation and memory. To this end, we propose parameterized learning problems, in which such factors can be controlled systematically and their effects on learning methods characterized through targeted studies. Apart from providing very precise control of the parameters that affect learning, our parameterized learning problems enable benchmarking against optimal behavior; their relatively small sizes facilitate extensive experimentation. Based on a survey of existing RL applications, in this article, we focus our attention on two predominant, first order factors: partial observability and function approximation. We design an appropriate parameterized learning problem, through which we compare two qualitatively distinct classes of algorithms: on-line value function-based methods and policy search methods. Empirical comparisons among various methods within each of these classes project Sarsa() and Q-learning() as winners among the former, and CMA-ES as the winner in the latter. Comparing Sarsa() and CMA-ES further on relevant problem instances, our study highlights regions of the problem space favoring their contrasting approaches. Short run-times for our experiments allow for an extensive search procedure that provides additional insights on relationships between method-specific parameterssuch as eligibility traces.


Descriptors :   machine learning , algorithms , computer programming , optimization , artificial neural networks , artificial intelligence


Subject Categories : Cybernetics


Distribution Statement : APPROVED FOR PUBLIC RELEASE