6533b82dfe1ef96bd1290919
RESEARCH PRODUCT
A Conclusive Analysis of the Finite-Time Behavior of the Discretized Pursuit Learning Automaton
B. John OommenXuan ZhangLei JiaoOle-christoffer Granmosubject
Property (philosophy)Learning automataDiscretizationMarkov chainComputer Networks and CommunicationsComputer scienceMarkov processMonotonic function02 engineering and technologyVDP::Matematikk og Naturvitenskap: 400::Informasjons- og kommunikasjonsvitenskap: 420Computer Science ApplicationsAutomatonsymbols.namesakeArtificial Intelligence0202 electrical engineering electronic engineering information engineeringsymbols020201 artificial intelligence & image processingMathematical economicsSoftwaredescription
Author's accepted version (post-print). © 20XX IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. Available from 20/03/2021. This paper deals with the finite-time analysis (FTA) of learning automata (LA), which is a topic for which very little work has been reported in the literature. This is as opposed to the asymptotic steady-state analysis for which there are, probably, scores of papers. As clarified later, unarguably, the FTA of Markov chains, in general, and of LA, in particular, is far more complex than the asymptotic steady-state analysis. Such an FTA provides rigid bounds for the time required for the LA to attain to a given convergence accuracy. We concentrate on the FTA of the Discretized Pursuit Automaton (DPA), which is probably one of the fastest and most accurate reported LA. Although such an analysis was carried out many years ago, we record that the previous work is flawed. More specifically, in all brevity, the flaw lies in the wrongly ``derived'' monotonic behavior of the LA after a certain number of iterations. Rather, we claim that the property should be invoked is the submartingale property. This renders the proof to be much more involved and deep. In this paper, we rectify the flaw and reestablish the FTA based on such a submartingale phenomenon. More importantly, from the derived analysis, we are able to discover and clarify, for the first time, the underlying dilemma between the DPA's exploitation and exploration properties. We also nontrivially confirm the existence of the optimal learning rate, which yields a better comprehension of the DPA itself.
year | journal | country | edition | language |
---|---|---|---|---|
2019-03-26 | IEEE Transactions on Neural Networks and Learning Systems |