default search action
ADPRL 2013: Singapore
- Proceedings of the 2013 IEEE Symposium on Adaptive Dynamic Programming and Reinforcement Learning, ADPRL 2013, IEEE Symposium Series on Computational Intelligence (SSCI), 16-19 April 2013, Singapore. IEEE 2013, ISBN 978-1-4673-5925-2
- Yujiao Huang, Huaguang Zhang, Dongsheng Yang:
Local stability analysis of high-order recurrent neural networks with multi-step piecewise linear activation functions. 1-5 - Qiming Zhao, Hao Xu, Sarangapani Jagannathan:
Finite-horizon optimal control design for uncertain linear discrete-time systems. 6-12 - Chunbin Qin, Huaguang Zhang, Yanhong Luo:
Adaptive optimal control for nonlinear discrete-time systems. 13-18 - Ruizhuo Song, Wendong Xiao, Yanhong Luo:
Optimal control for a class of nonlinear system with controller constraints based on finite-approximation-errors ADP algorithm. 19-23 - Hao Xu, Sarangapani Jagannathan:
Finite horizon stochastic optimal control of uncertain linear networked control system. 24-30 - Mostafa D. Awheda, Howard M. Schwartz:
Exponential moving average Q-learning algorithm. 31-38 - Zhen Ni, Xiao Fang, Haibo He, Dongbin Zhao, Xin Xu:
Real-time tracking on adaptive critic design with uniformly ultimately bounded condition. 39-46 - Jian Wang, Zhenhua Huang, Xin Xu:
A novel approach for constructing basis functions in approximate dynamic programming for feedback control. 47-51 - Yifan Cai, Simon X. Yang, Xin Xu:
A combined hierarchical reinforcement learning based approach for multi-robot cooperative target searching in complex unknown environments. 52-59 - Qi-ming Fu, Quan Liu, Fei Xiao, Guixin Chen:
The second order temporal difference error for Sarsa(λ). 60-68 - Lucian Busoniu, Alexander Daniels, Rémi Munos, Robert Babuska:
Optimistic planning for continuous-action deterministic systems. 69-76 - Raphaël Fonteneau, Lucian Busoniu, Rémi Munos:
Optimistic planning for belief-augmented Markov Decision Processes. 77-84 - Tobias Jung, Damien Ernst, Francis Maes:
Optimized look-ahead trees: Extensions to large and continuous action spaces. 85-92 - Donghun Lee, Boris Defourny, Warren B. Powell:
Bias-corrected Q-learning to control max-operator bias in Q-learning. 93-99 - Mingyuan Zhong, M. Johnson, Yuval Tassa, Tom Erez, Emo Todorov:
Value function approximation and model predictive control. 100-107 - Michiel van der Ree, Marco A. Wiering:
Reinforcement learning in the game of Othello: Learning against a fixed opponent and learning from self-play. 108-115 - A. Y. F. Lau, Dipti Srinivasan, Thomas Reindl:
A reinforcement learning algorithm developed to model GenCo strategic bidding behavior in multidimensional and continuous state and action spaces. 116-123 - Evangelos A. Theodorou, Jiri Najemnik, Emanuel Todorov:
Free energy based policy gradients. 124-131 - Teck-Hou Teng, Ah-Hwee Tan:
Delayed insertion and rule effect moderation of domain knowledge for reinforcement learning. 132-139 - Robert Lowe, Tom Ziemke:
Exploring the relationship of reward and punishment in reinforcement learning. 140-147 - Toshiyuki Yasuda, Nanami Wada, Kazuhiro Ohkura, Yoshiyuki Matsumura:
Analyzing collective behavior in evolutionary swarm robotic systems based on an ethological approach. 148-155 - Luuk Bom, Ruud Henken, Marco A. Wiering:
Reinforcement learning to train Ms. Pac-Man using higher-order action-relative inputs. 156-163 - Sachiko Soga, Ichiro Kobayashi:
A study on the efficiency of learning a robot controller in various environments. 164-169 - Hisashi Handa:
On the coordination system for the dimensionality-reduced inputs of mario. 170-176 - Xiaofeng Lin, Nuyun Cao, Yuzhang Lin:
Optimal control for a class of nonlinear systems with state delay based on Adaptive Dynamic Programming with ε-error bound. 177-182 - Xiong Luo, Jennie Si, Yuchao Zhou:
An integrated design for intensified direct heuristic dynamic programming. 183-190 - Kristof Van Moffaert, Madalina M. Drugan, Ann Nowé:
Scalarized multi-objective reinforcement learning: Novel design techniques. 191-199 - Zhanshan Wang, Fufei Chu, Hongjing Liang, Huaguang Zhang:
Fault accommodation for complete synchronization of complex neural networks. 200-205
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.