M. Brunner, T. Fiolka, D. Schulz, and C. M. Schlick, Design and comparative evaluation of an iterative contact point estimation method for static stability estimation of mobile actively reconfigurable robots, Robotics and Autonomous Systems, vol.63, pp.89-107, 2015.

K. I. Chatzilygeroudis, V. Vassiliades, F. Stulp, S. Calinon, and J. Mouret, A survey on policy search algorithms for learning robot controllers in a handful of trials, 2018.

K. Chua, R. Calandra, R. Mcallister, and S. Levine, Deep reinforcement learning in a handful of trials using probabilistic dynamics models, 2018.

H. Chung, C. Hou, Y. Chen, and C. Chao, An intelligent service robot for transporting object, IEEE Int. Symp. on Industrial Electronics, pp.1-6, 2013.

Y. Cong, X. Li, J. Liu, and Y. Tang, A stairway detection algorithm based on vision for ugv stair climbing, IEEE Int. Conf. on Networking, Sensing and Control, 2008.

K. Doelling, J. Shin, and D. O. Popa, Service robotics for the home: A state of the art review, Int. Conf. on Pervasive Technologies Related to Assistive Environments, 2014.

M. Ester, H. Kriegel, J. Sander, and X. Xu, A density-based algorithm for discovering clusters a density-based algorithm for discovering clusters in large spatial databases with noise, Proceedings of the Second International Conference on Knowledge Discovery and Data Mining, KDD'96, pp.226-231, 1996.

G. Genki, K. Nagatani, T. Hashimot, and K. Fujino, Slip-compensated odometry for tracked vehicle on loose and weak slope, ROBOMECH Journal, vol.4, issue.1, p.27, 2017.

M. Gianni, F. Ferri, M. Menna, and F. Pirri, Adaptive robust three-dimensional trajectory tracking for actively articulated tracked vehicles, Journal of Field Robotics, vol.33, issue.7, pp.901-930, 2016.

G. Grisetti, C. Stachniss, and W. Burgard, Improved techniques for grid mapping with rao-blackwellized particle filters, IEEE Transactions on Robotics, vol.23, issue.1, pp.34-46, 2007.

G. Grisetti, G. D. Tipaldi, C. Stachniss, W. Burgard, and D. Nardi, Fast and accurate slam with rao-blackwellized particle filters, vol.55, pp.30-38, 2007.

D. M. Helmick, S. I. Roumeliotis, M. C. Mchenry, and L. Matthies, Multi-sensor, high speed autonomous stair climbing, IEEE/RSJ Int. Conf. on Intelligent Robots and Systems, 2002.

J. A. Hesch, G. L. Mariottini, and S. I. , Roumeliotis. Descending-stair detection, approach, and traversal with an autonomous tracked vehicle, 2010.

. Ieee/rsj and . Int, Conf. on Intelligent Robots and Systems

S. Hirose, H. Tsukagoshi, and K. Yoneda, Normalized energy stability margin and its contour of walking vehicles on rough terrain, ICRA. IEEE Int. Conf. on Robotics and Automation, 2001.

J. A. Ijspeert, J. Nakanishi, and S. Schaal, Learning attractor landscapes for learning motor primitives, Proceedings of the 15th International Conference on Neural Information Processing Systems, NIPS'02, pp.1547-1554, 2002.

J. Kober and J. Peters, Policy search for motor primitives in robotics, Mach. Learn, vol.84, issue.1-2, pp.171-203, 2011.

J. Kober and J. Peters, Reinforcement Learning in Robotics: A Survey, pp.9-67, 2014.

S. Koenig and M. Likhachev, Eighteenth National Conference on Artificial Intelligence, pp.476-483, 2002.

A. Kupcsik, M. P. Deisenroth, J. Peters, A. P. Loh, P. Vadakkepat et al., Model-based contextual policy search for data-efficient generalization of robot skills, Special Issue on AI and Robotics, vol.247, pp.415-439, 2017.

A. G. Kupcsik, M. P. Deisenroth, J. Peters, and G. Neumann, Data-efficient generalization of robot skills with contextual policy search, Proceedings of the Twenty-Seventh AAAI Conference on Artificial Intelligence, 2013.

A. Laflaquière, J. K. O'regan, S. Argentieri, B. Gas, and A. V. Terekhov, Learning agent's spatial configuration from sensorimotor invariants, 2018.

M. Menna, M. Gianni, F. Ferri, and F. Pirri, Realtime autonomous 3d navigation for tracked vehicles in rescue environments, 2014 IEEE/RSJ Int. Conf. on Intelligent Robots and Systems

K. Nagatani, A. Yamasaki, K. Yoshida, T. Yoshida, and . Koyanagi, Semi-autonomous traversal on uneven terrain for a tracked vehicle using autonomous control of active flippers, IEEE/RSJ Int. Conf. on Intelligent Robots and Systems, 2008.

G. Neumann, W. Maass, and J. Peters, Learning complex motions by sequencing simpler motion templates, vol.382, p.95, 2009.

G. Paolo, L. Tai, and M. Liu, Towards continuous control of flippers for a multi-terrain robot using deep reinforcement learning, 2017.

M. Pecka, S. Valansky, K. Zimmermann, and T. Svoboda, Autonomous flipper control with safety constraints, 2016 IEEE/RSJ Int. Conf. on Intelligent Robots and Systems

J. Peters, K. Muelling, and Y. Altun, Relative entropy policy search, AAAI, 2010.

J. Peters, K. Mülling, and Y. Altun, Reinforcement learning by relative entropy policy search, 30th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Engineering, vol.30, p.69, 2010.

M. Quigley, K. Conley, B. P. Gerkey, J. Faust, T. Foote et al., Ros: an open-source robot operating system, ICRA Workshop on Open Source Software, 2009.

S. Soichiro, H. Satoshi, and O. Masayuki, Remote control system of disaster response robot with passive sub-crawlers considering falling down avoidance, ROBOMECH Journal, vol.1, issue.1, p.20, 2014.

R. S. Sutton and A. G. Barto, Introduction to Reinforcement Learning, 1998.

L. Zhang, K. Thurow, H. Liu, J. Huang, N. Stoll et al., Multi-floor laboratory transportation technologies based on intelligent mobile robots, Transportation Safety and Environment, 2019.