Next Article in Journal
A Novel Artificial Intelligence Technique to Estimate the Gross Calorific Value of Coal Based on Meta-Heuristic and Support Vector Regression Algorithms
Next Article in Special Issue
Detection of Sensitive Data to Counter Global Terrorism
Previous Article in Journal
Mechanical Analysis of Grouted Sleeve Lapping Connector
Previous Article in Special Issue
iHealthcare: Predictive Model Analysis Concerning Big Data Applications for Interactive Healthcare Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Brain-Inspired Goal-Oriented Robot Navigation System

College of Automation, Harbin Engineering University, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Submission received: 10 September 2019 / Revised: 11 November 2019 / Accepted: 11 November 2019 / Published: 14 November 2019

Abstract

:
Autonomous navigation in unknown environments is still a challenge for robotics. Many efforts have been exerted to develop truly autonomous goal-oriented robot navigation models based on the neural mechanism of spatial cognition and mapping in animals’ brains. Inspired by the Semantic Pointer Architecture Unified Network (SPAUN) neural model and neural navigation mechanism, we developed a brain-like biologically plausible mathematical model and applied it to robotic spatial navigation tasks. The proposed cognitive navigation framework adopts a one-dimensional ring attractor to model the head-direction cells, uses the sinusoidal interference model to obtain the grid-like activity pattern, and gets optimal movement direction based on the entire set of activities. The application of adaptive resonance theory (ART) could effectively reduce resource consumption and solve the problem of stability and plasticity in the dynamic adjustment network. This brain-like system model broadens the perspective to develop more powerful autonomous robotic navigation systems. The proposed model was tested under different conditions and exhibited superior navigation performance, proving its effectiveness and reliability.

1. Introduction

Animals have innate skills to navigate in unfamiliar environments [1]. It seems effortless for animals such as rats and primates but has been a fundamental long-term challenge for robots. Developing a brain-like navigation model based on the brain navigation mechanism of animals has attracted increasing attention.
The hippocampus, one of the most important brain regions, which codes for direction, distance, and speed signals of movement, is thought to play a pivotal role in place learning, navigation, and exploration. Place cells in the CA1 field of the hippocampus have a conspicuous functional characteristic of location-specific activity. They have a high firing rate when animals are in a particular location and may form a map-like representation of the environment, called grid cells space. Animals can use grid cells space for efficient navigation [2]. The grid cells found by Moser are also key cells for navigation [3,4,5]. Grid cells in the entorhinal cortex (EC) are arranged in strikingly periodic firing fields. A regularly tessellating triangular grid-like pattern was found to cover an entire environment, regardless of external cues. This suggests that grid cells have the ability to take in path integrals by internally generating a mapping of the spatial environment, which has been proved by an attractor network structure [6,7]. Head-direction (HD) cells exist in the postsubiculum and the EC [8]. Like a compass, they provide the input of biological plasticity for the brain-like navigation system [9]. The discharge frequency of speed cells found in the medial entorhinal cortex (MEC) is regulated by running speed, which could dynamically represent the animal’s location [10]. Integrating HD cells and speed cells, animals have knowledge of their direction and speed of motion and can track their idiothetic motion [11].
Navigation also involves memory and retrieval of environmental information, which is considered a cognitive mechanism for storing and processing information. A navigation system with spatial cognition and mapping can better achieve rapid and stable navigation. Anatomically, the neural responses of working memory are found in large areas of the parietal and frontal cortex. Accumulating evidence suggests that the posterior parietal cortex (PPC) and the dorsolateral prefrontal cortex (DLPFC) in the cerebral cortex are closely involved in working memory [12,13,14], temporary storage, and manipulation of data related to cognitive control [15]. Adaptive resonance theory (ART) is a cognitive and neural theory of how the brain autonomously learns to categorize and recognize events in a changing world. In this paper, the ART mechanism was used to deal with the retrieval and storage of environmental information. The rewards are a positive incentive for the desired action. The rewards will increase the selected probability of the previous action and guide the animal to easily reach the final goal. Anatomically, the orbitofrontal cortex (OFC) is involved in the process of reward signals [16]. In this paper, the rewards are represented by value functions of Q-learning to aid navigation.
Numerous computational models have been proposed based on place cells and grid cells. Each place cell has its own firing field and, taken together, the firing fields of place cells tend to cover the entire surface of the environment. The firing fields of place cells have mainly been modeled as Gaussian functions [17,18,19]. A large majority of grid cells have been modeled as three classes: oscillatory interference [20,21,22], continuous attractor network [6,7,23], and self-organizing [24,25,26]. These models concentrate only on the formation of place cells and grid cells, which does not really apply to navigation and is not instructive for autonomous robotics navigation.
Many robot navigation systems use a brain-like computational mechanism to complete navigation tasks. Milford developed a bionic navigation algorithm called RatSLAM [27,28]. It combines head cells and place cells to form an abstract pose cell, which is mainly used for accurate positioning. Zeng and Si [29] proposed a cognitive mapping model in which HD cells and grid cells use the same working mechanism; however, a place cell network is not included, and their model could not provide metric mapping. Tang [30] developed a hierarchical brain-like cognitive model with a particular focus on the spatial coding properties of the EC, which does not include navigation strategy or cognitive mapping. Moreover, none of the above models contains the biological mechanism of memory storage and retrieval, which has an important effect on effective navigation. Adaptive resonance theory (ATR) is a cognitive and neural theory of how the brain autonomously learns to categorize, recognize, and predict objects and events in a changing world. In this paper, ART is adapted to deal with working memory. ART based on matching learning is an effective approach to handle the dynamic stability of the network and get better navigation performance.
Eliasmith [31] proposed a unified set of neural mechanisms, Semantic Pointer Architecture Unified Network (SPAUN), which has performed plenty of tasks. The structure of SPAUN is divided into three layers: input information processing, working memory, and action execution. Inspired by the framework of SPAUN’s neural structure and the neural mechanisms of biological brain navigation, we proposed a novel three-layer brain-like overall structure of cognitive navigation, which was a supplement to SPAUN. The proposed navigation model can autonomously complete navigation tasks.

2. Mathematical Model

2.1. System Structure

Inspired by the SPAUN neural structure and the mechanism of neural navigation in animals’ brains, a brain-like biologically plausible cognitive model was proposed and applied to robotic spatial navigation tasks. The model’s architecture is shown in Figure 1.
The model includes three layers: cerebral cortex, hippocampal region and basal ganglia. The basal ganglia determine and perform the most highly recommended behavior, in this case, leading the robot to the target. The cerebral cortex mainly handles the working memory and reward function. When new knowledge is learned, the memory capacity of ART can increase adaptively without destroying the previous knowledge of the network. The ART method can build environmental information dynamically and stably. In this paper we used the ART network to dynamically process information, helping to achieve rapid and stable navigation.
Sitting between the cerebral cortex and basal ganglia is the hippocampal region, which includes the hippocampus and the EC. This plays a crucial role in spatial learning and navigation [32], where external information (HD and speed information) are converted to expressions of the internal environment. The hippocampal region is an area where grid cells, HD cells, and speed cells converge to transfer place information to the place cells.
Initially, speed cells and HD cells are inputs to layer III of the EC, which handles the ring attractor pre-processing. Then, they go through layer II of the EC, where they obtain the grid cells’ activities. The grid cells in EC II pass through the perforant pathway into the dentate gyrus (DG). Then, the granular cells in the DG are associated with the CA3 of the hippocampus by moss fibers. CA3 projects to the CA1 by Schaffer collaterals. The deep V/VI layer performs top-down feedback from the hippocampus to the EC to help stabilize memories.

2.2. Model Description

Motion information is obtained by integrating direction and speed signals, then updating the self-position. This paper uses a 1D ring attractor to model the HD cells:
HD i = [ cos ( ϕ i , 1 ) sin ( ϕ i , 1 ) cos ( ϕ i , m ) sin ( ϕ i , m ) ]
where ϕ i , j is the preferred direction of the jth cell in the ith ring attractor.
ϕ i , j = 2 π j m
j = 1, …, m, m = 6 is the number of neurons in a single attractor. Different rings have different preferred directions. There are three ring attractors. The angle difference is π/3.
Suppose that at time t, the robot is moving with speed V(t). The velocity projected to different ring attractors is:
v i ( t ) = HD i * V ( t )
Given the velocity, we can figure out the wavelength of sine function:
λ = v f
If two sinusoidal functions of different frequencies interfere with each other, it will generate beat frequency oscillations with a fixed spatial distance (as shown in Figure 2b). The beat frequency is equal to the difference between the two frequencies:
f b = | f 0 f i |
Then the wavelength of the periodic oscillation is:
λ i = v f b = v | f 0 f i |
λi a fixed value and represents the distance between vertices in the grid cells space. According to Equation (6), we derive:
f i = f 0 + v λ i
Set f0 = 7 Hz as a reference frequency. Beat frequency oscillation is integrated into the corresponding phase in different ring attractors (as shown in Figure 2a):
α i ( t ) = 2 π 0 t f i ( T ) d T
Converting phase-coded into rate-coded to express the position signal, the activity of the jth cell in the ith ring is:
S i , j ( t ) = sin ( α i ( t ) + ϕ i , j ) + 1 2
With comprehensive input of ring attractors with different preferred orientations, the amplitude of the beat frequency oscillations will be recognized at each x ± i (x = V(t) × t). This means that when the robot is moving linearly, it will produce grid-like activity with vertex spacing λi (as shown in Figure 2c). The activity of grid cells is:
g i , j ( t ) = S ( t )
The place cells accept the inputs of the grid cells’ population with different activities. The place cells are activated when the inputs are greater than the threshold h:
ψ ( g ) = { 1 g > h 0 otherwise

2.3. Navigation Strategy

There are two ways to guide the robot to the goal in an unknown environment: select the direction according to the reward signal or the Q-value of the action cells. Initially, the robot randomly explores the environment and will get a reward signal after reaching the target. If there are no reward signals within the search scope, it then selects a direction according to the Q-values; otherwise, the robot follows the reward signals.
If the reward signals can be detected in the step range, the value of the reward signals will increase according to Equation (12). With each step, all nodes update their reward signals. The direction with the maximum sum of reward signals will be chosen.
r ( t ) = r ( t 1 ) + Δ r
If the reward signal is zero at the present position, the robot selects the direction according to the Q-values. In our system, there are eight action directions (at 45° intervals). Each action cell receives the input from all activated place cells. According to a winner-takes-all mechanism, the direction with the maximum mean of Q-values is the optimal output direction.
Q o = i w i , o ψ i ( g ) i ψ i ( g )
where w i , o expresses the weight of place cell i in action direction o. The weights of the winning neuron are modified according to the following equation:
Δ w i , o = α ( r + β max ( Q o ) w i , o )
If both the Q-value and the reward signal are zero at the present position, the robot will move randomly. In this case, the robot keeps the direction with a probability of 1 − ξ. With a probability of ξ, it randomly chooses a new direction with the aim of finding other, shorter routes.

2.4. Network Dynamic Adjustment

The stability-plasticity dilemma problem, whereby the brain learns quickly and stably without catastrophically forgetting its past knowledge [33], exerts a significant influence on stable robot navigation. In our system, nodes were sparsely added to the cognitive map during robot exploration. We combined the ART network [34] with the robot navigation system not only to solve the problem of stable plasticity of the network dynamic adjustment, but also to enable the robot to complete navigation tasks autonomously.
In this paper, ART (simultaneously working and learning) was slightly modified to evaluate whether the nodes should be added to the map. The fundamental structure of ART is shown in Figure 3. The algorithm flow is as follows: (1) Network receives input xi; (2) Compute the match degree.
Calculate the distance between xi and other points in map bj.
n e t j = j = 1 n b j x i
(3) Choose the best match neuron.
Obtain the winning neuron by competition:
n e t j * = min { n e t j } = min { j = 1 n b j x i }
(4) Inspect warning threshold.
If n e t j * < ρ , compare the rewards of two nodes, adding the node with bigger reward to the map, then update the map connection.
Else, if random probability >0.25, then add the winning neuron to the map.

3. Results

3.1. Simulation Setup

Numerous researchers have used the Morris water maze experiment [35,36,37,38], in which a rat must find a hidden platform in its environment. This has become the standard paradigm for measuring navigation tasks.
A dry variant of the Morris water maze experiment was adopted to verify our system model. The navigation task was completed in a 100 × 100 virtual environment. The robot was placed in the lower left corner to find the target as shown in Figure 4. Initially, the robot randomly explored the surroundings, and the reward value was set to 1 when it found the goal. Table 1 summarizes the parameters used in the model.

3.2. Simulation Result

The robot started random exploration with an empty cognitive map and continued to update it using the ART process. We report three types of simulation environments to determine the effectiveness of our system model: an ideal situation with no obstacles in the environment (Figure 5), simple obstacles set in the environment (Figure 6), and changes the start position in the simple obstacles situation (Figure 7).
In the ideal situation with no obstacles, the robot explored until it found the target 30 times. The early path was relatively random due to the absence of reward signals, but with our navigation strategy, it could quickly search a shorter path of an approximately straight route to the goal after five trials. Moreover, the probability of 0.1 for random exploration was used to increase the opportunity of finding other shorter paths. For example, on the 8th and 16th paths, it tried new directions and avoided getting into the local optimum.
To further test the performance of the model, we added simple obstacles to the environment (Figure 6). In this situation, the robot must search for the goal and, at the same time, effectively avoid obstacles. Compared to the ideal situation with no obstacles, more exploration steps were needed in the early stage, but with the diffusion of rewards, it could quickly find a better path to the goal. In this paper, we aim to outline the general navigation principles in different environments, not provide a specific obstacle-avoidance algorithm.
The main challenge of robot navigation in an unfamiliar environment is simultaneous localization and mapping (SLAM). After establishing the cognitive map, navigation becomes path planning. The A * algorithm, one of the most popular heuristic search algorithms, is widely used in path optimization. According to the obtained cognitive map shown in Figure 6c, we easily obtained a relatively shorter path using the A* algorithm, before (Figure 7a) and after (Figure 7b) changing the start point.

3.3. Contrastive Analysis

The steps used to complete each trial are an important indicator of the navigation system. Kulvicius [39] added four pieces of olfactory information to aid navigation. Compared with that, the number of steps required to reach the goal are presented in Figure 8. In our system, having no olfactory cues to guide navigation, the first few trials needed more exploratory steps, but we could quickly search the relatively shorter path. This indicates that with the same environmental cues, we could provide better performance. Unlike in [39], the proposed model was consistent with the biological mechanism, indicating that it could effectively navigate in the experiment with immobile obstacles.
In terms of the map build, nodes were sparsely added to the cognitive map during robot exploration. However, in [35] there was no judgment when new nodes were recruited, which would lead to high node randomness and not help further navigation. We added the ART network to solve the problem of stable plasticity of network dynamic adjustment. Based on the matching learning method, the ART process can effectively save resources, dynamically express the environment, and autonomously complete navigation tasks.

4. Conclusions

Due to the flexibility and autonomy of biological navigation, many efforts have been exerted to apply biological mechanisms to developing autonomous robot navigation systems. Although a variety of computational models have been proposed according to animals’ goal-oriented behavior, rarely have they included the structure of multiple navigation cells. We propose a general structure for robot positioning, navigation, and memory mapping. We combined it with ART to form a new intelligent navigation system. The application of ART in robot technology promotes its cognitive ability. The proposed model was simulated and verified with three different conditions. The results show that the model enables the robot to maintain the basic ability of building an overall topological and metric map of an unknown environment. It also has the following additional capabilities and advantages:
Under the guidance of population activities and reward signals, the robot can complete autonomous navigation tasks in an unknown environment and converge to an optimal path faster.
By using the mechanism of forced exploration, new paths can be explored so as to effectively avoid falling into the local optimum.
In conclusion, there are three main contributions of this paper. First, we constructed a unified overall navigation mathematical model that combines the biological plasticity of brain-like learning processes with working memory. This provides a valuable feedback tool for neuroscience. Second, we applied ART to aid in navigation, which reduced resource consumption and broadened the perspective to develop more powerful autonomous robotic navigation systems. Third, detecting the potential mechanism of biological space navigation, a hypothesis is proposed that there may be a region in the brain corresponding to the ART function waiting for biological neurologists to explore.

Author Contributions

Methodology, Q.C.; Software, Q.C.; Supervision, H.M.; Writing—original draft, Q.C.; Writing—review & editing, H.M.

Funding

This research was funded by National key research and development program of China [grant number 2018AAA0102700].

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Geva-Sagiv, M.; Las, L.; Yovel, Y.; Ulanovsky, N. Spatial cognition in bats and rats: From sensory acquisition to multiscale maps and navigation. Nat. Rev. Neurosci. 2015, 16, 94–108. [Google Scholar] [CrossRef] [PubMed]
  2. O’Keefe, J.; Dostrovsky, J. The hippocampus as a spatial map. Preliminary evidence from unit activity in the freely-moving rat. Brain Res. 1971, 34, 171–175. [Google Scholar] [CrossRef]
  3. Giocomo, L.M.; Moser, M.B.; Moser, E.I. Computational models of grid cells. Neuron 2011, 71, 589–603. [Google Scholar] [CrossRef]
  4. Hafting, T.; Fyhn, M.; Molden, S.; Moser, M.B.; Moser, E.I. Microstructure of a spatial map in the entorhinal cortex. Nature 2005, 436, 801–806. [Google Scholar] [CrossRef] [PubMed]
  5. Stensola, H.; Stensola, T.; Solstad, T.; Frøland, K.; Moser, M.B.; Moser, E.I. The entorhinal grid map is discretized. Nature 2012, 492, 72–78. [Google Scholar] [CrossRef] [PubMed]
  6. Burak, Y.; Fiete, I.R. Accurate path integration in continuous attractor network models of grid cells. PLoS Comput. Biol. 2009, 5, 16. [Google Scholar] [CrossRef] [PubMed]
  7. McNaughton, B.L.; Battaglia, F.P.; Jensen, O.; Moser, M.B.; Moser, E.I. Path integration and the neural basis of the ‘cognitive map’. Nat. Rev. Neurosci. 2006, 7, 663–678. [Google Scholar] [CrossRef] [PubMed]
  8. Ranck, J.B. Head-direction cells in the deep cell layers of dorsal presubiculum in freely moving rats. Soc. Neurosci. Abstr. 1984, 10, 599. [Google Scholar]
  9. Taube, J.S.; Bassett, J.P. Persistent neural activity in head direction cells. Cereb. Cortex 2003, 13, 1162–1172. [Google Scholar] [CrossRef]
  10. Kropff, E.; Carmichael, J.E.; Moser, M.B.; Moser, E.I. Speed cells in the medial entorhinal cortex. Nature 2015, 523, 419–478. [Google Scholar] [CrossRef]
  11. Valerio, S.; Taube, J.S. Path integration: How the head direction signal maintains and corrects spatial orientation. Nat. Neurosci. 2012, 15, 1445–1453. [Google Scholar] [CrossRef]
  12. Allen, T.A.; Fortin, N.J. The evolution of episodic memory. Proc. Natl. Acad. Sci. USA 2013, 110, 10379–10386. [Google Scholar] [CrossRef] [PubMed]
  13. Liu, D.; Gu, X.; Zhu, J.; Zhang, X.; Han, Z.; Yan, W.; Cheng, Q.; Hao, J.; Fan, H.; Hou, R.; et al. Medial prefrontal activity during delay period contributes to learning of a working memory task. Science 2014, 346, 458–463. [Google Scholar] [CrossRef] [PubMed]
  14. Silva, M.A.D.; Huston, J.P.; Wang, A.; Petri, D.; Chao, O.Y.H. Evidence for a specific integrative mechanism for episodic memory mediated by AMPA/kainate receptors in a circuit involving medial prefrontal cortex and hippocampal CA3 region. Cereb. Cortex 2016, 26, 3000–3009. [Google Scholar] [CrossRef] [PubMed]
  15. Hoshi, E. Functional specialization within the dorsolateral prefrontal cortex: A review of anatomical and physiological studies of non-human primates. Neurosci. Res. 2006, 54, 73–84. [Google Scholar] [CrossRef]
  16. Kringelbach, M.L. The human orbitofrontal cortex: Linking reward to hedonic experience. Nat. Rev. Neurosci. 2005, 6, 691–702. [Google Scholar] [CrossRef]
  17. Arleo, A.; Gerstner, W. Spatial cognition and neuro-mimetic navigation: A model of hippocampal place cell activity. Biol. Cybern. 2000, 83, 287–299. [Google Scholar] [CrossRef]
  18. Hartley, T.; Burgess, N.; Lever, C.; Cacucci, F.; O’Keefe, J. Modeling place fields in terms of the cortical inputs to the hippocampus. Hippocampus 2000, 10, 369–379. [Google Scholar] [CrossRef]
  19. Solstad, T.; Moser, E.I.; Einevoll, G.T. From grid cells to place cells: A mathematical model. Hippocampus 2006, 16, 1026–1031. [Google Scholar] [CrossRef]
  20. Brandon, M.P.; Bogaard, A.R.; Libby, C.P.; Connerney, M.A.; Gupta, K.; Hasselmo, M.E. Reduction of theta rhythm dissociates grid cell spatial periodicity from directional tuning. Science 2011, 332, 595–599. [Google Scholar] [CrossRef]
  21. Burgess, N.; Barry, C.; O’Keefe, J. An oscillatory interference model of grid cell firing. Hippocampus 2007, 17, 801–812. [Google Scholar] [CrossRef]
  22. Giocomo, L.M.; Zilli, E.A.; Fransen, E.; Hasselmo, M.E. Temporal frequency of subthreshold oscillations scales with entorhinal grid cell field spacing. Science 2007, 315, 1719–1722. [Google Scholar] [CrossRef] [PubMed]
  23. Fuhs, M.C.; Touretzky, D.S. A spin glass model of path integration in rat medial entorhinal cortex. J. Neurosci. 2006, 26, 4266–4276. [Google Scholar] [CrossRef] [PubMed]
  24. Grossberg, S.; Pilly, P.K. How entorhinal grid cells may learn multiple spatial scales from a dorsoventral gradient of cell response rates in a self-organizing map. Plos Comput. Biol. 2012, 8, 31. [Google Scholar] [CrossRef] [PubMed]
  25. Mhatre, H.; Gorchetchnikov, A.; Grossberg, S. Grid cell hexagonal patterns formed by fast self-organized learning within entorhinal cortex. Hippocampus 2012, 22, 320–334. [Google Scholar] [CrossRef] [PubMed]
  26. Pilly, P.K.; Grossberg, S. Spiking neurons in a hierarchical self-organizing map model can learn to develop spatial and temporal properties of entorhinal grid cells and hippocampal place cells. PLoS ONE 2013, 8, 22. [Google Scholar] [CrossRef] [PubMed]
  27. Milford, M.J.; Wyeth, G.F. Mapping a suburb with a single camera using a biologically inspired SLAM system. IEEE Trans. Robot. 2008, 24, 1038–1053. [Google Scholar] [CrossRef]
  28. Milford, M.J.; Wyeth, G.F. Persistent navigation and mapping using a biologically inspired SLAM system. Int. J. Robot. Res. 2010, 29, 1131–1153. [Google Scholar] [CrossRef]
  29. Zeng, T.P.; Si, B.L. Cognitive mapping based on conjunctive representations of space and movement. Front. Neurorobot. 2017, 11, 16. [Google Scholar] [CrossRef]
  30. Tang, H.J.; Huang, W.; Narayanamoorthy, A.; Yan, R. Cognitive memory and mapping in a brain-like system for robotic navigation. Neural Netw. 2017, 87, 27–37. [Google Scholar] [CrossRef]
  31. Eliasmith, C.; Stewart, T.C.; Choo, X.; Bekolay, T.; DeWolf, T.; Tang, Y.; Tang, C.; Rasmussen, D. A Large-scale model of the functioning brain. Science 2012, 338, 1202–1205. [Google Scholar] [CrossRef]
  32. Carpenter, F.; Manson, D.; Jeffery, K.; Burgess, N.; Barry, C. Grid cells form a global representation of connected environments. Curr. Biol. 2015, 25, 1176–1182. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Grossberg, S. How does a brain build a cognitive code? Psychol. Rev. 1980, 87, 1–51. [Google Scholar] [CrossRef] [PubMed]
  34. Grossberg, S. Adaptive resonance theory: How a brain learns to consciously attend, learn, and recognize a changing world. Neural Netw. 2013, 37, 1–47. [Google Scholar] [CrossRef] [PubMed]
  35. Erdem, U.M.; Hasselmo, M. A goal-directed spatial navigation model using forward trajectory planning based on grid cells. Eur. J. Neurosci 2012, 35, 916–931. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Llofriu, M.; Tejera, G.; Contreras, M.; Pelc, T.; Fellous, J.M.; Weitzenfeld, A. Goal-oriented robot navigation learning using a multi-scale space representation. Neural Netw. 2015, 72, 62–74. [Google Scholar] [CrossRef] [PubMed]
  37. Rogers, J.; Churilov, L.; Hannan, A.J.; Renoir, T. Search strategy selection in the Morris water maze indicates allocentric map formation during learning that underpins spatial memory formation. Neurobiol. Learn. Mem. 2017, 139, 37–49. [Google Scholar] [CrossRef]
  38. Erdem, U.M.; Milford, M.J.; Hasselmo, M.E. A hierarchical model of goal directed navigation selects trajectories in a visual environment. Neurobiol. Learn. Mem. 2015, 117, 109–121. [Google Scholar] [CrossRef] [Green Version]
  39. Kulvicius, T.; Tamosiunaite, M.; Ainge, J.; Dudchenko, P.; Wörgötter, F. Odor supported place cell model and goal navigation in rodents. J. Comput. Neurosci. 2008, 25, 481–500. [Google Scholar] [CrossRef] [Green Version]
Figure 1. System diagram of navigation model based on Semantic Pointer Architecture Unified Network (SPAUN) and brain neural system. This framework deals with data from input signals to final action generation to achieve autonomous navigation. The model consists of three layers: hippocampal region, dealing with input signals (head direction and instantaneous speed signals) by path integration to expression position; cerebral cortex, creating and modifying the topological map of environmental information and passing the reward information to the decision-making layer; and basal ganglia, determining the final action by winner-take-all rule. Meaning of abbreviations: DG, dentate gyrus; CA3 and CA1, internal area of hippocampus; EC, entorhinal cortex; PPC, posterior parietal cortex; DLPFC, dorsolateral prefrontal cortex.
Figure 1. System diagram of navigation model based on Semantic Pointer Architecture Unified Network (SPAUN) and brain neural system. This framework deals with data from input signals to final action generation to achieve autonomous navigation. The model consists of three layers: hippocampal region, dealing with input signals (head direction and instantaneous speed signals) by path integration to expression position; cerebral cortex, creating and modifying the topological map of environmental information and passing the reward information to the decision-making layer; and basal ganglia, determining the final action by winner-take-all rule. Meaning of abbreviations: DG, dentate gyrus; CA3 and CA1, internal area of hippocampus; EC, entorhinal cortex; PPC, posterior parietal cortex; DLPFC, dorsolateral prefrontal cortex.
Applsci 09 04869 g001
Figure 2. (a) Illustration of ring attractor mechanism for path integration. The attractor model was simplified into a one-dimensional linear ring attractor model. Suppose a mouse moves in one direction along a straight path. With different distances from the apex, each attractor can be expressed by the activity of a ring attractor. (b) Schematic diagram of periodic oscillation obtained by interference of different frequency waveforms. y1 = sin(14πx + π/2); y2 = sin(16πx + 3π/2); y3 = y1 + y2; The θ oscillating beat interference will produce a fixed interval beat y4, and the amplitude of the beat oscillation will be identified at each attractor. (c) Illustration of how hexagon-like firing mode was generated. Trajectory of random movement of robot (blue) and grid-like firing mode generated by this paper’s method (red).
Figure 2. (a) Illustration of ring attractor mechanism for path integration. The attractor model was simplified into a one-dimensional linear ring attractor model. Suppose a mouse moves in one direction along a straight path. With different distances from the apex, each attractor can be expressed by the activity of a ring attractor. (b) Schematic diagram of periodic oscillation obtained by interference of different frequency waveforms. y1 = sin(14πx + π/2); y2 = sin(16πx + 3π/2); y3 = y1 + y2; The θ oscillating beat interference will produce a fixed interval beat y4, and the amplitude of the beat oscillation will be identified at each attractor. (c) Illustration of how hexagon-like firing mode was generated. Trajectory of random movement of robot (blue) and grid-like firing mode generated by this paper’s method (red).
Applsci 09 04869 g002
Figure 3. The fundamental structure of adaptive resonance theory.
Figure 3. The fundamental structure of adaptive resonance theory.
Applsci 09 04869 g003
Figure 4. Environmental setup. The N/E at the top left is environmental direction, N = north, E = east.
Figure 4. Environmental setup. The N/E at the top left is environmental direction, N = north, E = east.
Applsci 09 04869 g004
Figure 5. Results of navigation in ideal no-obstacle condition: (a) Navigation trajectories of 30 runs; first row is 1–6 runs, and so on. (b) Number of steps needed to reach the goal. (c) Map nodes (blue circle) were established during the exploration.
Figure 5. Results of navigation in ideal no-obstacle condition: (a) Navigation trajectories of 30 runs; first row is 1–6 runs, and so on. (b) Number of steps needed to reach the goal. (c) Map nodes (blue circle) were established during the exploration.
Applsci 09 04869 g005
Figure 6. Results of navigation with immobile obstacle. Red square represents the target, blue line is the path, and obstacles are shown by “***”: (a) Navigation trajectories results of 30 runs; first row is 1–6 runs, and so on. (b) Number of steps needed to reach the goal. (c) Map nodes (blue circle) were established during the exploration.
Figure 6. Results of navigation with immobile obstacle. Red square represents the target, blue line is the path, and obstacles are shown by “***”: (a) Navigation trajectories results of 30 runs; first row is 1–6 runs, and so on. (b) Number of steps needed to reach the goal. (c) Map nodes (blue circle) were established during the exploration.
Applsci 09 04869 g006
Figure 7. Results of navigation according to the cognitive map, obstacles are shown by “***”: (a) We got a shorter path (red line) from the start point to the target. (b) After changing the start point, we quickly got a shorter path (red line) to the target.
Figure 7. Results of navigation according to the cognitive map, obstacles are shown by “***”: (a) We got a shorter path (red line) from the start point to the target. (b) After changing the start point, we quickly got a shorter path (red line) to the target.
Applsci 09 04869 g007
Figure 8. Comparison of steps required to reach the goal. Bar: Running steps of our model with immobile obstacle. Nobar: Running steps of our model with no obstacle. Vis+Olf: Running steps of [39] based on visual and olfactory cues. Vis+Olf+Mark: Running steps of [39] based on visual, olfactory, and self-marking cues.
Figure 8. Comparison of steps required to reach the goal. Bar: Running steps of our model with immobile obstacle. Nobar: Running steps of our model with no obstacle. Vis+Olf: Running steps of [39] based on visual and olfactory cues. Vis+Olf+Mark: Running steps of [39] based on visual, olfactory, and self-marking cues.
Applsci 09 04869 g008
Table 1. Parameter setting.
Table 1. Parameter setting.
ParameterαβλξhΔrρ
Value0.70.720 cm0.10.50.00512

Share and Cite

MDPI and ACS Style

Chen, Q.; Mo, H. A Brain-Inspired Goal-Oriented Robot Navigation System. Appl. Sci. 2019, 9, 4869. https://0-doi-org.brum.beds.ac.uk/10.3390/app9224869

AMA Style

Chen Q, Mo H. A Brain-Inspired Goal-Oriented Robot Navigation System. Applied Sciences. 2019; 9(22):4869. https://0-doi-org.brum.beds.ac.uk/10.3390/app9224869

Chicago/Turabian Style

Chen, Qiuying, and Hongwei Mo. 2019. "A Brain-Inspired Goal-Oriented Robot Navigation System" Applied Sciences 9, no. 22: 4869. https://0-doi-org.brum.beds.ac.uk/10.3390/app9224869

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop