Adaptive behavior navigation of a mobile robot
2002, IEEE Transactions on Systems, Man, and Cybernetics
https://doi.org/10.1109/3468.995537Abstract
Describes a neural network model for the reactive behavioral navigation of a mobile robot. From the information received through the sensors the robot can elicit one of several behaviors (e.g., stop, avoid, stroll, wall following), through a competitive neural network. The robot is able to develop a control strategy depending on sensor information and learning operation. Reinforcement learning improves the navigation of the robot by adapting the eligibility of the behaviors and determining the linear and angular robot velocities
References (23)
- Richard S. Sutton and Andrew G. Barto, Toward a modern theory of adaptive networks: Expectation and prediction," Psychological R eview, v ol. 88, pp. 135 170, 1981.
- E. Thorndike, Animal Intelligence, Hafner, Darien, C.T., 1911.
- P. Maes, Robotics and autonomous systems," Situated A gents Can Have Goals, v ol. 6, pp. 49 70, 1990.
- S. Mahadevan and J. Connell, Automatic programing of behavior-based robots using reinforcement learning," in Proceedings of the Ninth National Conference o n A rti cial Intelligence AAAI'91, pp. 768 773. Ananheim, CA, 1991.
- M. Asada, S. Noda, S. Tawaratsumida, and K. Hosoda, Vision-based reinforcement learning for purposive behavior acquisition," in Proceedings of the IEEE International Conference o n R obotics and Automation, pp. 146 153. 1995.
- T Kohonen, The self-organizing map," in Proc. of the IEEE, 78, 9. 1990, pp. 1464 1480, Piscataway, NJ: IEEE Service Center.
- S. Grossberg, Contour enhancement, short-term memory, and constancies in reverberating neural networks," Studies in Applied Mathematics, v ol. 52, pp. 217 257, 1973.
- P. Gaudiano, E. Zalama, Carolina Chang, and J. L opez Coronado, A model of operant conditioning for adaptive obstacle avoidance," in From Animals to Animats 4: Proceedings of the Fourth International Con- ference on Simulation of Adaptive Behavior, The MIT Press Bradford Books, Ed., Cape Cod, Masachussets, USA, 1996.
- D. Bullock and S. Grossberg, The VITE model: A neural command circuit for generating arm and articulator trajectories," Dynamic Patterns In Complex Systems, pp. 305 326, 1988.
- A. Baloch and A. Waxman, Visual learning, adaptive expectations, and behavioral conditioning of the mobile robot MAVIN," Neural Networks, v ol. 4, pp. 271 302, 1991.
- E. Zalama, P. Gaudiano, and J. L opez-Coronado, A real-time, unsupervised neural network for the low level control of a mobile robot in a nonstationary environment," Neural Networks, v ol. 8, pp. 103 123, 1995.
- A. Bl uhlmeier, H.R. Everett, and L. Feng, Operant conditioning in robotics," in Neural Systems for Robotics, O. Omidvar and P. V and der Smagt, Eds., pp. 195 225. 1997.
- P.F.M.J. Verschure and A.C.C. Coolen, Adaptive elds: Distributed representations of clasically conditioned associations," Network, v ol. 2, pp. 189 217, 1991.
- P.F.M.J. Verschure and R. Pfeifer, Categorization, representations, and the dynamics of system-environment interaction," in Proceedings of the Second International Conference on Simulation of Adaptive Behavior, J.A. Meyer, H. Roitblat, and S. Wilson, Eds., pp. 210 217. Cambridge, MA: MIT Press, 1992.
- R. Pfeifer and P. V erschure, Distributed adaptive control: a paradigm for designing autonomous agents," in Toward a p r actice of autonomous systems, F.J. Varela and P. Bourgine, Eds., pp. 21 30. 1992.
- U. Nehmozow, Flexible control of mobile robot through autonomous competence adquisition," Mesurement and Control, v ol. 282, 1995.
- DRAFT April 25, 2001
- A. Kurz, Constructing maps for mobile robot navigation based on ultrasonic range data," IEEE Transaction on System Man and Cybernetics, v ol. 26, no. 2, pp. 433 441, 1996.
- H.R. Beom and H.S. Cho, A sensor based navigation for a mobile robot using fuzzy logic and reinforcing learning," IEEE Transaction on System Man and Cybernetics, v ol. 25, no. 3, pp. 464 477, 1995.
- D. Gachet, M. Moreno, and J. Pimentel, Learning emergent tasks for an autonomous mobile robot," in Pro- ceedings of the International Conference on Intelligent Robots and Systems IROS'94, pp. 290 297. Munich, Germany, 1994.
- J. Millan, Learning e cient reactive behavioral sequences from basic re exes in a goal-directed autonomous robot," in Proceedings of the Third Conference on Simulation of Adaptive Behavior, pp. 266 274. 1994.
- R. Sutton, Learning to predict by the methods of temporal di erences," Machine Learning, v ol. 3, pp. 9 44, 1988.
- A. Fagg, D. Lotspeich, and G. Bekey, A reinforcement-learning approach to reactive control policy design for autonomous robots," in Proceedings of the IEEE International Conference o n R obotics and Automation, pp. 39 44. 1994. April 25, 2001 DRAFT