[1]
|
Matsui M. A generalized model of convey-serviced production station (CSPS). Journal of Japan Industrial Management Association, 1993, 44(1): 25-32
|
[2]
|
Matsui M. CSPS model: look-ahead controls and physics. International Journal of Production Research, 2005, 43(10): 2001-2025
|
[3]
|
Hao T, Tamio A. Look-ahead control of conveyor-serviced production station by using potential-based online policy iteration. International Journal of Control, 2009, 82(10): 1917-1928
|
[4]
|
Yamada T, Satomi K, Matsui M. Strategic selection of assembly systems under viable demands. Assembly Automation, 2006, 26(4): 335-342
|
[5]
|
Nakase N, Yamada T, Matsui M. A management design approach to a simple flexible assembly system. International Journal of Production Economics, 2002, 76(3): 281-292
|
[6]
|
Feyzbakhsh S A, Matsui M. Adam-eve-like genetic algorithm: a methodology for optimal design of a simple flexible assembly system. Computers & Industrial Engineering, 1999, 36(2): 233-258
|
[7]
|
Tang Hao, Wan Hai-Feng, Han Jiang-Hong, Zhou Lei. Coordinated look-ahead control of multiple CSPS system by multi-agent reinforcement learning. Acta Automatica Sinica, 2010, 36(2): 289-296(唐昊, 万海峰, 韩江洪, 周雷. 基于多Agent 强化学习的多站点CSPS 系统的协作 Look-ahead 控制. 自动化学报, 2010, 36(2): 289-296)
|
[8]
|
Yan Q C, Liu Q, Hu D J. A hierarchical reinforcement learning algorithm based on heuristic reward function. In: Proceedings of the 2nd IEEE International Conference on Advanced Computer Control. Shenyang, China: IEEE, 2010. 371-376
|
[9]
|
Botvinick M M. Hierarchical reinforcement learning and decision making. Current Opinion in Neurobiology, 2012, 22(6): 956-962
|
[10]
|
Jia Q S. Event-based optimization with lagged state information. In: Proceedings of the 31st Chinese Control Conference. Hefei, China: IEEE, 2012. 2055-2060
|
[11]
|
Yuasa H, Ito M. Self-organizing system theory by use of reaction-diffusion equation on a graph with boundary. In: Proceedings of the 1999 IEEE International Conference on Systems, Man, and Cybernetics. Tokyo, Japan: IEEE, 1999. 211-216
|
[12]
|
Wright R, Lin S. Evolutionary tile coding: an automated state abstraction algorithm for reinforcement learning. In: Proceedings of the the 2010 Abstraction, Reformulation, and Approximation. Atlanta, Georgia, USA: the Association for the Advancement of Artificial Intelligence Workshops, 2010
|
[13]
|
Li L H, Walsh T J, Littman M L. Towards a unified theory of state abstraction for MDPs. In: Proceedings of the 9th International Symposium on Artificial Intelligence and Mathematics. Fort Lauderdale, Florida, USA: Kluwer AcademicPublishers, 2006. 531-539
|
[14]
|
Singh S P, Jaakkola T, Jordan M I. Reinforcement learning with soft state aggregation. In: Proceedings of the 1995 Conference on Neural Information Processing Systems. Denver, CO, USA: MIT, 1995. 361-368
|
[15]
|
Gunady M K, Gomaa W. Reinforcement learning generalization using state aggregation with a maze-solving problem. In: Proceedings of the 2012 Japan-Egypt Conference on Electronics, Communication and Computers. Alexandria, Egypt: IEEE, 2012. 157-162
|
[16]
|
Cao X R. Semi-Markov decision problems and performance sensitivity analysis. IEEE Transaction on Automatic Control, 2003, 48(5): 758-769
|