Reinforcement learning via kernel temporal difference.
Publication
, Conference
Bae, J; Chhatbar, P; Francis, JT; Sanchez, JC; Principe, JC
Published in: Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference
January 2011
This paper introduces a kernel adaptive filter implemented with stochastic gradient on temporal differences, kernel Temporal Difference (TD)(λ), to estimate the state-action value function in reinforcement learning. The case λ=0 will be studied in this paper. Experimental results show the method's applicability for learning motor state decoding during a center-out reaching task performed by a monkey. The results are compared to the implementation of a time delay neural network (TDNN) trained with backpropagation of the temporal difference error. From the experiments, it is observed that kernel TD(0) allows faster convergence and a better solution than the neural network.
Duke Scholars
Published In
Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference
DOI
EISSN
2694-0604
ISSN
2375-7477
Publication Date
January 2011
Volume
2011
Start / End Page
5662 / 5665
Related Subject Headings
- User-Computer Interface
- Reinforcement, Psychology
- Pattern Recognition, Automated
- Humans
- Electroencephalography
- Brain
- Biomimetics
- Artificial Intelligence
- Algorithms
Citation
APA
Chicago
ICMJE
MLA
NLM
Bae, J., Chhatbar, P., Francis, J. T., Sanchez, J. C., & Principe, J. C. (2011). Reinforcement learning via kernel temporal difference. In Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference (Vol. 2011, pp. 5662–5665). https://doi.org/10.1109/iembs.2011.6091370
Bae, Jihye, Pratik Chhatbar, Joseph T. Francis, Justin C. Sanchez, and Jose C. Principe. “Reinforcement learning via kernel temporal difference.” In Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference, 2011:5662–65, 2011. https://doi.org/10.1109/iembs.2011.6091370.
Bae J, Chhatbar P, Francis JT, Sanchez JC, Principe JC. Reinforcement learning via kernel temporal difference. In: Annual International Conference of the IEEE Engineering in Medicine and Biology Society IEEE Engineering in Medicine and Biology Society Annual International Conference. 2011. p. 5662–5.
Bae, Jihye, et al. “Reinforcement learning via kernel temporal difference.” Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference, vol. 2011, 2011, pp. 5662–65. Epmc, doi:10.1109/iembs.2011.6091370.
Bae J, Chhatbar P, Francis JT, Sanchez JC, Principe JC. Reinforcement learning via kernel temporal difference. Annual International Conference of the IEEE Engineering in Medicine and Biology Society IEEE Engineering in Medicine and Biology Society Annual International Conference. 2011. p. 5662–5665.
Published In
Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE Engineering in Medicine and Biology Society. Annual International Conference
DOI
EISSN
2694-0604
ISSN
2375-7477
Publication Date
January 2011
Volume
2011
Start / End Page
5662 / 5665
Related Subject Headings
- User-Computer Interface
- Reinforcement, Psychology
- Pattern Recognition, Automated
- Humans
- Electroencephalography
- Brain
- Biomimetics
- Artificial Intelligence
- Algorithms