@ARTICLE{10.3389/fnbot.2017.00020, AUTHOR={Goldschmidt, Dennis and Manoonpong, Poramate and Dasgupta, Sakyasingha}, TITLE={A Neurocomputational Model of Goal-Directed Navigation in Insect-Inspired Artificial Agents}, JOURNAL={Frontiers in Neurorobotics}, VOLUME={11}, YEAR={2017}, URL={https://www.frontiersin.org/articles/10.3389/fnbot.2017.00020}, DOI={10.3389/fnbot.2017.00020}, ISSN={1662-5218}, ABSTRACT={Despite their small size, insect brains are able to produce robust and efficient navigation in complex environments. Specifically in social insects, such as ants and bees, these navigational capabilities are guided by orientation directing vectors generated by a process called path integration. During this process, they integrate compass and odometric cues to estimate their current location as a vector, called the home vector for guiding them back home on a straight path. They further acquire and retrieve path integration-based vector memories globally to the nest or based on visual landmarks. Although existing computational models reproduced similar behaviors, a neurocomputational model of vector navigation including the acquisition of vector representations has not been described before. Here we present a model of neural mechanisms in a modular closed-loop control—enabling vector navigation in artificial agents. The model consists of a path integration mechanism, reward-modulated global learning, random search, and action selection. The path integration mechanism integrates compass and odometric cues to compute a vectorial representation of the agent's current location as neural activity patterns in circular arrays. A reward-modulated learning rule enables the acquisition of vector memories by associating the local food reward with the path integration state. A motor output is computed based on the combination of vector memories and random exploration. In simulation, we show that the neural mechanisms enable robust homing and localization, even in the presence of external sensory noise. The proposed learning rules lead to goal-directed navigation and route formation performed under realistic conditions. Consequently, we provide a novel approach for vector learning and navigation in a simulated, situated agent linking behavioral observations to their possible underlying neural substrates.} }