Sie sind auf Seite 1von 9

Importing the Computational Neuroscience Toolbox into

Neuro-Evolution—Application to Basal Ganglia


Jean-Baptiste Mouret, Stéphane Doncieux, Benoît Girard

To cite this version:


Jean-Baptiste Mouret, Stéphane Doncieux, Benoît Girard. Importing the Computational Neuroscience
Toolbox into Neuro-Evolution—Application to Basal Ganglia. GECCO’10, 2010, Portland, United
States. pp.587-594, �10.1145/1830483.1830592�. �hal-00687639v2�

HAL Id: hal-00687639


https://hal.archives-ouvertes.fr/hal-00687639v2
Submitted on 26 Apr 2019

HAL is a multi-disciplinary open access L’archive ouverte pluridisciplinaire HAL, est


archive for the deposit and dissemination of sci- destinée au dépôt et à la diffusion de documents
entific research documents, whether they are pub- scientifiques de niveau recherche, publiés ou non,
lished or not. The documents may come from émanant des établissements d’enseignement et de
teaching and research institutions in France or recherche français ou étrangers, des laboratoires
abroad, or from public or private research centers. publics ou privés.
Importing the Computational Neuroscience Toolbox into
Neuro-Evolution—Application to Basal Ganglia

Jean-Baptiste Mouret Stéphane Doncieux Benoît Girard


mouret@isir.upmc.fr doncieux@isir.upmc.fr girard@isir.upmc.fr
ISIR, Universit Pierre et Marie Curie-Paris 6, CNRS UMR 7222
4 place Jussieu, F-75252, Paris Cedex 05, France

ABSTRACT
Neuro-evolution and computational neuroscience are two sci-
entific domains that produce surprisingly different artificial
neural networks. Inspired by the “toolbox” used by neu-
roscientists to create their models, this paper argues two
main points: (1) neural maps (spatially-organized identi-
cal neurons) should be the building blocks to evolve neural
networks able to perform cognitive functions and (2) well-
identified modules of the brain for which there exists com-
putational neuroscience models provide well-defined bench-
marks for neuro-evolution.
To support these claims, a method to evolve networks
of neural maps is introduced then applied to evolve neu-
ral networks with a similar functionality to basal ganglia in
animals (i.e. action selection). Results show that: (1) the
map-based encoding easily achieves this task while a direct
encoding never solves it; (2) this encoding is independent of
the size of maps and can therefore be used to evolve large
and brain-like neural networks; (3) the failure of direct en-
coding to solve the task validates the relevance of action
selection as a benchmark for neuro-evolution.

Categories and Subject Descriptors


Figure 1: Contracting basal ganglia model for action
I.2.6 [Artificial intelligence]: Learning—Connectionism
selection. D1/D2 Str: striatal neurons with D1 or
and neural nets
D2 dopamine receptors; FC: frontal cortex; FS: fast
spiking interneurons; GPe/GPi: external and inter-
General Terms nal part of the globus pallidus; S: input saliences;
Algorithms SNr: substancia nigra reticulata; STN: subthalamic
nucleus; TH: thalamus; TRN: reticular thalamic nu-
cleus.
Keywords
evolutionary algorithms; neural networks; computational neu-
roscience; basal ganglia; neuro-evolution.
based generative systems [16, 20, 30, 31] and development-
inspired approaches [4, 43]. Resulting neural networks can
1. INTRODUCTION then control robots or artificial agents [28, 33, 31].
Evolution has been one of the prominent processes to Concurrently with these researches, the field of computa-
shape the animal brain; imitating this process by employing tional neuroscience (see e.g. [6]) designs models of parts of
evolution-inspired algorithms to design “artificial nervous the nervous system based on experimental data from biol-
systems” therefore drew considerable attention in artificial ogy. Among the published models, some of them are effi-
intelligence over the last two decades. This line of thought cient enough to be useful in robots [23, 22, 47, 29, 25, 35,
drives researchers to propose many methods to exploit evo- 15] and could provide the basic blocks to build a complete
lutionary algorithms (EA) to design neural networks (neuro- bio-inspired “artificial brain” for artificial intelligence.
evolution), from efficient direct encodings [44] to grammar- Computational neuroscience and neuro-evolution there-
fore share a characteristic—creating artificial nervous sys-
tems; however, the comparison of the resulting neural net-
Author-generated version.
GECCO’10, July 7–11, 2010, Portland, Oregon, USA. works is surprising: models produced in computational neu-
(c) The authors. roscience display almost no similarities with evolved neural
networks. The latter usually involve up to a few dozens of trol problems, direct encodings fail to scale up to large net-
neurons and do not scale up to many more neurons, whereas works, supposedly because they cannot capture the regu-
computational neuroscience often manipulate hundreds of larity of the search space by using several times the same
neurons in a very organized fashion. A detailed investi- sub-networks.
gation of published models (especially [17, 18, 15, 45, 46, Noticing that biological systems widely rely on the rep-
40, 39]) reveals that the chosen building blocks constitute etition and combination of hierarchically organized mod-
at least one fundamental difference; evolutionary methods ules [20, 19, 30, 31], several researchers proposed to encode
mostly use individual neurons, ideally organized in modu- neural networks indirectly by employing a compact repre-
lar fashion [20, 38, 8, 30, 31], but many neuroscience mod- sentation that is then developed into a neural network. One
els (e.g. figure 1) rely on the concepts of maps (a N × M of the straightforward way to implement this idea is to use
grid of neurons, in which N and M are free variables of the a list of modules as a genotype, as done in modular encod-
model) connected by regular connection schemes (either one ings [8, 38, 31]. A more complex but potentially more pow-
to one connections or one to all with a regular assignation erful approach is to evolve “construction programs” whose
of weights). This allows such neural networks to scale up instructions are interpreted to build neural networks. This
to larger maps (e.g. to handle higher-dimensional inputs) modus operandi was employed in cellular encoding [16] to de-
while maintaining the same overall structure. This descrip- sign highly repetitive neural networks and in several other
tion of neural networks as connected maps can be seen as grammar-based generative systems [20, 30, 31].
the result of a developmental process in which a network of A last method is to evolve chemical gradients in Cartesian
maps is developed to form a neural network. space that can then be used to compute synaptic weights.
Besides this analysis, it appears that computational neu- Provided that these gradients displays some regularities (e.g.
roscience can also provide an efficient approach to bench- repetitive patterns or symmetry), the synaptic weights can
mark neuro-evolution methods in the context of cognitive mirror them. This process is currently instantiated in Hy-
functions. On one hand, neuro-evolution ended up with perNEAT [43], one of the rare neuro-evolution framework
substantially good results to control non-linear systems [16, able to evolve networks of several hundred of neurons. How-
44, 8], but despite Beer’s preliminary proposals [2, 41], no ever, the published method is not designed to evolve the
clear benchmark has been widely accepted to evaluate their topology of neural networks1 and requires the experimenter
potential for generating more cognitive functions. On the to fix the number of neurons. Moreover, despite its inspi-
other hand, experimental neuroscience isolated several mod- ration from biological development, it remains to be proved
ules of the nervous systems, and precisely described their that HyperNEAT can generate neural networks similar to
inputs and outputs such that computational neuroscience those observed in real animals.
can model them. On the road to automatically design ar-
tificial nervous systems, the minimum benchmark for any 2.2 EAs and Computational Neuroscience
neuro-evolution method should be to reproduce the func- Current neuro-evolution methods mainly draw from com-
tions modeled by neuro-scientists and to equal the efficiency puter science (L-systems [20, 19], formal grammars [16, 30],
of hand-designed neural networks. graph theory, machine learning, control theory[44]) and more
Building on this inspiring parallel, this paper argues two rarely biological development [4, 43]. To our knowledge, only
main points: a few attempts to explicitly link this field to computational
neuroscience have been published.
• Maps (and not individual neurons) should be the build- The main import from neuroscience to neuro-evolution
ing blocks to evolve neural networks able to perform is undoubtedly the neuron formulation. Beer et al. [3] ana-
cognitive functions. lyzed the evolution of networks of leaky integrators, a neuron
• Well-identified modules of the brain for which there ex- model initially introduced in computational neuroscience,
ists computational neuroscience models provide well- for “minimal cognitive behaviors”. Models at other abstrac-
defined benchmarks for neuro-evolution. tion levels, such as spiking neurons [10], nonlinear oscilla-
tors [27] and neural fields [21], have also been considered
To support these claims, a method to evolve networks of as the building blocks of evolved neural networks. Another
maps is described then applied to evolve neural networks important inspiration from neuroscience to neuro-evolution
with a similar functionality to basal ganglia (winner-takes- is on-line learning schemes such as Hebbian-like rules [12]
all selection by disinhibition). Results are compared with an and neuro-modulation [24, 42]. It should be emphasized
optimized basal ganglia model from neuroscience and net- that these papers mostly copied a brick from neuroscience
works evolved with a classic direct encoding. and then evolved networks without any biological inspira-
tion about the topology. As a consequence, the resulting
2. RELATED WORK neural networks are topologically different from those hand-
designed by neuro-scientists.
2.1 Evolving Neural Networks Following a reasoning closer to the one argued here, Ijspeert
et al. [23] imported from Ekeberg et al. [9] the overall orga-
Most of the work on the structural evolution of neural
nization and the neuron parameters—both extracted from
networks focuses on the definition of efficient encodings with
animal experiments—of the neural network that generates
their associated genetic operators. In direct encodings, ge-
the locomotion patterns of lampreys (the Central Pattern
netic operators directly manipulate connections and neu-
Generator, CPG). In a multi-stage evolutionary process, the
rons. The most successful direct encoding is undoubtedly
authors first optimized the connectivity and the synaptic
NEAT [44], which is based on “innovation numbers” to im-
prove cross-over and structural diversity preservation. De- 1
HyperNEAT only evolves synaptic weights but connections
spite their simplicity and their efficiency to solve some con- with near zero weights can be removed.
the activation level of neuron j and dt the integration step
(e.g. 0.001 s).
The main building blocks of the considered computational
neuroscience models are either N × M maps of neurons or
individual neurons. Maps are defined as spatially organized
grids of identical neurons (same time constant, same thresh-
old, same inhibitory status). Many models employ only
maps with the same dimension, arbitrary fixed to the di-
Figure 2: Three ways to connect two maps of neu- mension of the input. At any rate, the number of map di-
rons. (a) one to one; (b) one to all, with the same mensions in a given model is very restricted as it typically
synaptic weight; (c) one to all with a Gaussian dis- covers only the dimension of the inputs and those of the
tribution of synaptic weights. outputs.
Connection schemes between maps are restricted to three
cases (figure 2) in most models: (1) one to one connection
weights of segmental oscillators and then the connectivity with constant weights (neuron i of map M1 is connected to
between these oscillators. The obtained network oscillates neuron j of map M2 , with a positive weight identical for
over a wider range of frequency and amplitude than the each connection), (2) one to all connections with constant
hand-designed neural network [23] and, although they have weights (neuron i of map M1 is connected to each neuron
only been investigated in simulation, they could be embed- of map M2 , with identical weights for all connections) and
ded in snake-like robots [22]. By constraining the evolu- (3) one to all connections with weights following a Gaussian
tionary process with biological data, Ijspeert et al. man- distribution computed as follows:
aged to automatically design a large but regular neural net-
work more efficient than a hand-designed controller. This ||i − j||2
di,j = , i ∈ M1 , j ∈ M2 (4)
result concurs with the approach presented here: using data N
from neuroscience (connectivity patterns, overall organiza- −di,j
wi,j = Γ exp( 2 ), i ∈ M1 , j ∈ M2 (5)
tion, common parameters) could lead neuro-evolution to a σ
new level of functionality.
where N is the map size2 , i is the i-th neuron of map M1 , j
the j-th of M2 and wi,j the synaptic weight between i and
3. MAP-BASED ENCODING j. σ (standard deviation) and Γ (weight amplitude) are the
only parameters that can be changed.
3.1 Computational Neuroscience Toolbox This regular assignation of weights (either constant or fol-
A careful examination of the published neuroscience mod- lowing a simple distribution) contrasts with typical neuro-
els, and especially of those that could be employed in artifi- evolution methods in which each synaptic weight is set-up
cial intelligence, allows to extract some regularities in model separately. In effect, this adds numerous regularities in neu-
descriptions. Neuro-scientists converged to a limited set of ral networks that are similar to those sought by modular
structures that can describe a wide range of biological neural and developmental encodings.
networks while being simple enough to be analyzed and well This computational neuroscience toolbox is sufficient to
understood. We list here the main features of this “compu- describe a wide range of models, for instance models of basal
tational neuroscience toolbox”. ganglia [17, 18, 15], of colliculus [45] and of visual attention
Most neuroscience models use either spiking neurons [14] [46, 40, 39]. Additions such as Hebbian learning and neural
or leaky integrators, which only simulate the overall dynamic fields would allow to describe many other models.
of a population of neurons [3]. This latter simple dynamic
neuron has also been widely used in evolutionary robotics 3.2 Evolving a Graph of Neural Maps
(see [11]) as it shows many different temporal behaviors but Most computational neuroscience models can be described
remains computationally inexpensive to simulate. Further- as a graph of neural maps in which each map and each con-
more, in contrast to neurons employed in neuro-evolution or nection is described by a set of parameters (time constant of
in machine learning, neurons manipulated in computational neurons, connection scheme, synaptic weights, etc.). Such
neuroscience are either inhibitory or excitatory. a labeled graph can be modified structurally (add/remove a
In this work, we opt for the lPDS-based (locally Projected connection or a node) and parametrically (change of a la-
Dynamic System) neuron model [15], a variant of the clas- bel). We first describe a basic method to evolve a generic
sic leaky integrator with similar dynamics but which verify labeled graph then we introduce the set of labels employed
the dynamic property of contraction [26]. Using the Euler to describe a graph of neural maps.
integration method, the output yt+dt of a lPDS neuron i at
time t + dt is computed as follows: 3.2.1 Evolving a Labeled Graph
The graph is represented as a classic adjacency list for
(i) P (j)
pt = j∈C wi,j yt (1) which mutation operators have been designed; cross-over is
(i) (i) (i) (i) not used.
at+dt = max(0, min(1, at + τ1 (pt − at + Ti ) · dt)) (2)
( In this work, we employ straightforward mutation oper-
(i) ators that apply each of the possible changes with a user-
(i) at+dt if i is excitatory
yt+dt = (i) (3) defined probability (see appendix to know the chosen val-
−at+dt otherwise
2
It is here assumed that all the maps have the same size.
where τ is the time-constant of neuron i, Ti a threshold This formula is easily generalized to the case of two maps
(j)
value, wi,j the synaptic weight between neuron i and j, at with different sizes.
Figure 3: Overview of the development process. From left to right: (1) the genotype is a labeled graph with
evolvable labels; (2) the labels are interpreted to a neuroscience-inspired description of the neural network;
(3) for a given size of maps, this neural network can be fully developed into a neural network (for instance
to evaluate its fitness).

ues). Three structural mutation operators, inspired by those Similarly, connections are labeled with four evolved real
defined in NEAT [44], have been designed: numbers {c, k, Γ, σ} ∈ [0, 1)4 that are interpreted as follows:
• addition of a node on an existent connection, with ran- 
1 to 1, if c < 0.5
dom labels; the connection is split in two and the two Scheme: (10)
1 to all, otherwise
parts keep the same labels; 
Gaussian if c > 0.5 and k < 0.5
• removal of a random node and its associated connec- Function: (11)
constant otherwise
tions; 
Γ if c > 0.5 and k < 0.5
• addition/removal of a connection between two random Parameter 1: 1 (12)
5
w (synaptic weight) otherwise.
neurons. 
σ if c > 0.5 and k < 0.5
Nodes and connections can be labeled by a list of real pa- Parameter 2: (13)
otherwise: ignored
rameters that represent weights, threshold, neuron type, ...
These parameters are mutated using polynomial mutation [7]. 3.3 Classic Direct Encoding
To initialize a random graph, a random number of nodes
The previously described evolvable labeled graph can also
(with random labels) are first created then connected by a
be employed in a more classic fashion to directly define a
random number of connections (also with random labels).
neural network. This leads to a direct encoding of neural
Connections cannot be doubled. Last, graphs are simplified
networks. In this case, each node describes a neuron (instead
by removing each sub-graph not connected to both an input
of a map). Labels used to describe neurons are the same as
and an output.
those used to describe maps (time constant, etc.). Connec-
3.2.2 Map-based Encoding tions are labeled with a single real number interpreted as
the synaptic weight. This direct encoding will be used as a
To evolve a network of maps inspired by the neuroscience
reference encoding in the next section.
toolbox, the previously defined operators are employed with
a particular labeling of nodes and connections (figure 3).
This representation is independent of the size of maps, that 4. EXPERIMENTS
is the same genotype can be developed into a neural network
with an arbitrary number of neurons. 4.1 Action Selection in Basal Ganglia
In the current implementation, nodes are labeled with four Having introduced the map-based encoding, we now focus
evolved numbers {e, v, T, a} ∈ [0, 1)4 that are scaled or bi- on evaluating its ability to generate brain-like structures; to
narized to obtain nodes’ parameters. They are interpreted that aim, we chose to reproduce the function performed by
as follows3 : the basal ganglia (BG).
 More precisely, the basal ganglia are a set of intercon-
inhibitory if e < 0.5 nected subcortical nuclei [36], that are thought to be in-
Type: (6)
excitatory otherwise. volved in action selection [37, 34], i.e. the problem, for an
Time constant (τ ): τ = vbτ 0 ×4c (7) agent, of choosing which action to perform within a reper-
−3 −3 −3
toire, given internal and external sensory information, in
where v = [5 · 10 , 10 , 20 · 10 , order to achieve its goals. Solving this generic resource al-
40 · 10−3 ] location problem seems to be a central cognitive function as
Threshold: 10 · (T − 0.5) (8) the BG circuitry is common to all vertebrates, and as it con-
 tains duplicated circuits [1] dedicated to many critical func-
isolated neuron if a > 0.75 tions (skeleton movements, eye movements, sequence learn-
Map type: (9)
standard map otherwise. ing, working memory, navigation, planning, etc.). The BG
3
bxc denotes the floor of x. performs the two main aspects of action selection: the cen-
tral process of selection of one action only among conflicting 4.2.1 Objective Function
ones, similar to a winner-takes-all (WTA), and the learn- The main objective function (fitness) aims at checking
ing process necessary to bias the selection process towards that the maximum salience corresponds to the minimum
the most profitable option. We focus here on the WTA-like activation in the corresponding output neuron (and conse-
process. quently to the less inhibited action). Furthermore, we are
The BG circuits are subdivided in parallel channels, which interested in the best contrast possible, that is the selected
are supposed to be associated to the actions of the reper- action should be as little inhibited as possible and the other
toire, and which receive convergent inputs from the cortex. actions should be as inhibited as possible. To formulate this
At rest, the BG output is tonically active and inhibits its tar- idea, let first define min (v) (the index of the maximum value
gets. The selection of an action, caused by a strong conver- of v) and mout (x, v) (the index of the minimum value of the
gent input on the associated channel signaling its urgency, output vector, for the input v):
causes a pause of the channel output inhibition, allowing the
activation of the targeted brain region [5]. Various models of
this WTA process have been proposed; we consider here the min (v) = arg max(vi ) (18)
i∈1,··· ,k
CBG model proposed in [15], which derives from the new
functional interpretation of the circuitry proposed by [17], mout (x, v) = arg min(γc (x, v)i ) (19)
i∈1,··· ,k
but adds some previously unused connections (figure 1).
In the following, we test the ability of our new map-based We also need the index of the second minimum value of the
(2)
encoding to evolve circuits exhibiting the WTA ability ex- output vector, mout (x, v):
pected from BG models, and compare the obtained results
(2)
with CBG models whose parameters4 are optimized by a mout (x, v) = arg min (x, v)(γc (x, v)i ) (20)
i∈1,··· ,k,i6=mout
basic evolutionary algorithm.
Let I a set of N random salience vectors in [0.1, 1], uni-
4.2 Fitness formly distributed. Given that an action will be selected if
In the remaining text, the following notations are used: the corresponding output channel is below γ0 (x), we define
• x: a developed individual (a neural network); the fitness to be maximized as the average selection success
(the correct channel was disinhibited) weighted by the differ-
• k: number of inputs (also the number of outputs); ence between the minimum output (the selected action) and
the second minimum output (the first inhibited action). If
• T : the maximum simulation duration;
all outputs except the minimum one are above γ0 (x), γ0 (x)
• γ(x, v, t)i : activation level of the output neuron i (i ∈ is used to compute the difference. We thus foster the maxi-
{1, ..., k}) at time t (t ∈ [0, T ]), given the input vector mum inhibition of the selected channel but let the network
v of saliences (v ∈ Rk ); free to use any value above γ0 (x) for the other channels. The
fitness is normalized such that the maximum value is 1.
• Tc (x, v): the duration of the simulation for the input
vector v (see K(x) below); 1 X
F (x) = δmin ,mout (x, v) · γc (x, v)mout ... (21)
|I| · γ0 (x) v∈I
• γc (x, v)i : activation level of the output neuron i (i ∈
(2)

{1, ..., k}) at the end of the simulation (i.e. t = Tc (x, v)); ... − min(mout (x, v), γ0 (x))
• γ0 (x) the average activation level of the output neurons where δmin ,mout (x, v) denotes the Kronecker’s delta:
when the inputs are null, at the end of the experiment: 
1 if mout (x, v) = min (x, v)
k δmin ,mout (x, v) = (22)
1X 0 otherwise
γ0 (x) = γc (x, [0, · · · , 0], 1)i (14)
k i=1
In these experiments, N was fixed to N = 1000 and the
same vectors were employed to evaluate all individuals.
Each individual is simulated until its output converges to a
constant vector or until it reaches the maximum number of 4.2.2 Constraints
time-steps (t = T ). From a practical viewpoint, a neural
The search is restricted to networks that converge during
network is considered to have converged when 10 successive
the simulation time and that do not select anything if all the
outputs have a difference of less than ε (in these experiments,
input saliences are low. These restrictions are implemented
ε = 10−6 ). To compute Ts , we first define the “convergence
using three constraints.
function” K(x, t, v):
The first constraint C1 (x) ensures that γ0 (x) is greater
 than a minimum value γmin , fixed to 0.09 in this work (the
 0 if γ(x, v, t)i − γ(x, v, t − n)i < ε, resting activity of the CBG model is 0.096). This means
K(x, t, v) = ∀n ∈ {1, · · · , 10}, ∀i ∈ {1, · · · , k} that all the actions should be sufficiently inhibited if all the
1 otherwise saliences are null.

(15)
Ts can now be defined as: 
0 if γ0 (x) > γmin
C1 (x) = (23)
Ts (x, v) = t such as K(x, t, v) = 1, t ∈ [0, · · · , T ](16) 1 otherwise
Tc = min(T, Ts (x, v)) (17) The second constraint C2 (x) ensures that small values are
4 filtered (this behavior is observed in animals’ basal ganglia)
the CBG is parameterized by 22 synaptic weights and 3
neuron thresholds by sampling random saliences in [0, 0.1] and checking that
Table 1: Summary of launched experiments.
Search space Genotype Channels
1 synaptic weights of CBG 25 real numbers 6
2 topology and param. direct encoding 6
3 topology and param. map-based encoding 6
4 synaptic weights of CBG 25 real numbers 15
5 topology and param. direct encoding 15
6 topology and param. map-based encoding 15

Figure 4: Median fitness with regards to genera-


Table 2: Parameters used in the experiments. tion number (200 evaluations for each generation).
Parameter / Genotype Map-based Direct enc.
min./max. nb. of nodes (rand. gen.) 3/5 18 / 25
The map-based encoding achieve the highest fitness
min/max. nb. of links (rand. gen.) 3/5 18 / 25 while the direct encoding does not solve the task
prob. to add/remove a node 0.05 / 0.05 0.05 / 0.05 (median fitness is negative for direct encoding with
prob. to add/remove link 0.05 / 0.05 0.05 / 0.05 15 channels, therefore it does not appear on the
prob. to change each label 0.1 0.1
σ for gaussian mutation 0.05 0.05 graph).

1
0.8
the outputs are greater than γ0 (x): 0.6
 0.4
 0 if γc (x, v)i > γ0 (x),
0.2
C2 (x) = ∀v ∈ [0, 0.1]k , ∀i ∈ {1, · · · , k} (24)
0

1 otherwise CB ma dir CB ma dir
G  p e G p e
(6) -bas ct en  (15 -bas ct en
ed c.  ) ed c. 
Last, C3 (x) checks that the tested neural network con-  (6 (  (1 (15
) 6) 5) )
verges to a constant output vector before the end of the ex-
fit > 0 mean fit.
periment, for all the tests performed on the neural network
(C1 (x), C2 (x) and F (x)):
C3 (x) = max(K(x), v) (25) Figure 5: Proportion of runs that satisfy the con-
v∈I straints (fitness > 0) after 200, 000 evaluations and
These constraints are enforced with the penalty method average fitness of those runs. All differences are sta-
[7]: an arbitrary large penalty is added to the fitness each tistically significant (Student T-test, p < 0.01) except
time a constraint is violated. Instead of maximizing F (x), between map-based (6) and map-based (15).
we thus maximize Fc (x):
Fc (x) = −K(C1 (x) + C2 (x) + C1 (x)) + F (x) (26)
10
where K is an arbitrary large constant (e.g. 10 ).

4.3 Experiments
To evaluate the relevance of the proposed approach to
solve the action selection task, we ask three questions: (1)
how well does the proposed map-based encoding (section
3.2.2) solve this task compared to an optimized CBG ([15],
section 4.1)? (2) compared to a basic direct encoding (sec-
tion 3.3)? and (3) does it scale up to high numbers of neu-
rons?
To that aim, six different experiments have been launched
(table 1), each of those consisting of 10 independent evolu-
tionary runs, each of them with a budget of 200, 000 evalu-
ations. The same evolutionary algorithm, the same fitness
and the same model of neurons were employed in all experi-
ments; only the genotype/phenotype mapping was changed.
The chosen evolutionary algorithm is a single-objective im-
plementation of NSGA-2 [7], an elitist tournament-based
evolutionary algorithm. Parameters are provided in table 2
and the source code is available online5 . Experiments were
carried in the Sferesv2 framework [32].

Figure 6: Example of a neural network obtained


5. RESULTS with the map-based encoding for 6 channels (fit-
5 ness: 1.0).
http://www.isir.fr/evorob_db
For six channels, the optimized CBGs reach a fitness of action selection task whereas a basic direct encoding
about 0.8 (after 200, 000 evaluations, median: 0.83, mean: never solves it;
0.84, s.d.: 0.035; see figures 4 and 5). This means that it
performs well, but some saliences are not well selected or • this description is independent of the size of maps and
the contrast is not maximum. Surprisingly, the map-based can therefore be used to evolve large and brain-like
encoding leads to better results than this reference point by neural networks;
achieving a fitness of almost 1 with a low standard devia-
• the failure of direct encoding to solve the task validates
tion (median: 0.951, mean: 0.950, s.d: 0.013). Overall, the
the relevance of action selection as a benchmark.
map-based encoding easily and reliably solves the task of
action selection. The discrepancy between map-based en- The described map-based encoding could easily be im-
coding and the reference point (optimized CBG) probably proved by adding more “tools” from the computational neu-
stems from the many biological constraints on the design of roscience literature. Learning rules [12, 24, 42] are probably
the CBG. the most useful addition. Moreover, other basic tasks of
With the basic direct encoding, about 600 generations the brain have to be investigated with the same encoding
(120, 000 evaluations) were required to obtain a positive me- such as auto-calibration, working memory [13] or selective
dian fitness (i.e. more than half of the runs obtained at least attention [41, 46, 40].
one individual that satisfies the constraint). This contrasts
with the map-based encoding runs, in which only a few gen-
erations were necessary to satisfy the same constraints. Af-
7. ACKNOWLEDGMENTS
ter 200, 000 evaluations (1000 generations), 80% of the direct This project was funded by the ANR EvoNeuro project,
encoding runs satisfy the constraints (figure 5); however, the ANR-09-EMER-005-01.
mean fitness is low (0.3) and corresponds to individuals that
do not solve the task. 8. REFERENCES
A typical neural network obtained with the map-based [1] G. E. Alexander, M. R. DeLong, and P. L. Strick.
encoding is drawn on figure 6. In addition to the input and Parallel organization of functionally segregated
output maps, this network is made of only two hidden maps. circuits linking basal ganglia and cortex. Annual
They are linked using a mix of 8 “one to one” and 2 “one Review of Neuroscience, 9:357–381, 1986.
to all” connections (no Gaussian “one to all” connections are [2] R. Beer. Toward the evolution of dynamical neural
employed; as in the CBG model, they are useless to solve networks for minimally cognitive behavior. In From
this problem). This neural network is therefore described
Animals to Animats 4, pages 421–429, 1996.
by 4 × 4 + 10 × 4 = 56 real numbers and graph topology.
[3] R. Beer and J. Gallagher. Evolving dynamical neural
The same network described using a direct encoding would
networks for adaptive behavior. Adaptive behavior,
require (for six channels) 3+8×6+2×6 = 132 real numbers
1(1):91, 1992.
in addition to graph topology. This illustrates the intuitive
difference in size of the search spaces using map-based and [4] A. Cangelosi, D. Parisi, and S. Nolfi. Cell division and
direct encoding. The low score achieved by direct encoding migration in a ’genotype’ for neural networks.
suggests that the subspace explored with the map-based en- Network: Computation in Neural Systems,
coding is a useful restriction to solve the targeted cognitive 5(4):497–515, 1994.
function, in this case action selection. [5] G. Chevalier and M. Deniau. Disinhibition as a basic
The 15 channels experiments demonstrate how well the process of striatal functions. Trends in Neurosciences,
map-based encoding scales up. Since the neural network 13:277–280, 1990.
description is independent of the size of the map, results [6] P. Dayan, L. Abbott, and L. Abbott. Theoretical
obtained with 6 or 15 channels are statistically not different. neuroscience: Computational and mathematical
No runs managed to satisfy the constraints using the direct modeling of neural systems. MIT Press, 2001.
encoding. Surprisingly, although the description of the CBG [7] K. Deb. Multi-objectives optimization using
is also based on maps, the CBG obtained lower fitness values evolutionnary algorithms. Wiley, 2001.
with 15 channels than with 6 channels. We are investigating [8] S. Doncieux and J.-A. Meyer. Evolving modular
this issue. neural networks to solve challenging control problems.
In Proc. of EIS 2004, 2004.
6. CONCLUSIONS AND FUTURE WORK [9] O. Ekeberg. A combined neuronal and mechanical
model of fish swimming. Biological cybernetics,
A new encoding to evolve neural networks has been in- 69(5):363–374, 1993.
troduced in this paper. It is based on the “computational
[10] D. Floreano, Y. Epars, J. Zufferey, and C. Mattiussi.
neuroscience toolbox” and especially on maps of neurons
Evolution of spiking neural circuits in autonomous
connected with regular connections schemes. This encod-
mobile robots. International Journal of Intelligent
ing is designed to explore with evolutionary algorithms a
Systems, 21(9):1005–1024, 2006.
subspace of possible neural networks similar to the one ex-
plored in computational neuroscience. The proposed encod- [11] D. Floreano and C. Mattiussi. Bio-Inspired Artificial
ing has been compared to a basic direct encoding and to a Intelligence: Theories, Methods, and Technologies.
hand-designed model on the task of action selection, an im- MIT Press, 2008.
portant function of basal ganglia in animals. Results show [12] D. Floreano and J. Urzelai. Evolution of plastic control
that: networks. Autonomous Robots, 11(3):311–317, 2001.
[13] M. J. Frank, B. Loughry, and R. C. O’Reilly.
• the map-based encoding easily and reliably solves the Interactions between frontal cortex and basal ganglia
in working memory: a computational model. [30] J.-B. Mouret and S. Doncieux. MENNAG: a modular,
Cognitive, Affective, and Behavioral Neuroscience, regular and hierarchical encoding for neural-networks
1(2):137–160, 2001. based on attribute grammars. Evolutionary
[14] W. Gerstner and W. Kistler. Spiking Neuron Models: Intelligence, 1:187–207, 2008.
An Introduction. Cambridge University Press, 2002. [31] J.-B. Mouret and S. Doncieux. Evolving modular
[15] B. Girard, N. Tabareau, Q. C. Pham, A. Berthoz, and neural-networks through exaptation. In Proc. of
J.-J. Slotine. Where neuroscience and dynamic system IEEE-CEC, 2009.
theory meet autonomous robotics: a contracting basal [32] J.-B. Mouret and S. Doncieux. Sferesv2 : evolvin’ in
ganglia model for action selection. Neural Networks, the multicore world. In Proc. of IEEE-CEC, 2010.
21(4):628–641, 2008. [33] S. Nolfi and D. Floreano. Evolutionary robotics: The
[16] F. Gruau. Automatic definition of modular neural biology, intelligence, and technology of self-organizing
networks. Adaptive Behaviour, 3(2):151–183, 1995. machines. Bradford Book, 2004.
[17] K. Gurney, T. J. Prescott, and P. Redgrave. A [34] T. Prescott. Action selection. Scholarpedia, 3(2):2705,
computational model of action selection in the basal 2008. doi:10.4249/scholarpedia.2705.
ganglia. I. A new functional anatomy. Biological [35] T. J. Prescott, F. Montes-Gonzalez, K. Gurney, M. D.
cybernetics, 84(6):401–410, 2001. Humphries, and P. Redgrave. A robot model of the
[18] K. Gurney, T. J. Prescott, and P. Redgrave. A basal ganglia: Behavior and intrinsic processing.
computational model of action selection in the basal Neural Networks, 19:31–61, 2006.
ganglia. II. Analysis and simulation of behaviour. [36] P. Redgrave. Basal ganglia. Scholarpedia, 2(6):1825,
Biological cybernetics, 84(6):411–423, 2001. 2007.
[19] G. S. Hornby. Measuring, enabling and comparing [37] P. Redgrave, T. Prescott, and K. N. Gurney. The
modularity, regularity and hierarchy in evolutionary basal ganglia: A vertebrate solution to the selection
design. In Proc. of GECCO, pages 1729–1736, 2005. problem? Neuroscience, 89:1009–1023, 1999.
[20] G. S. Hornby and J. B. Pollack. Creating high-level [38] J. Reisinger, K. O. Stanley, and R. Miikkulainen.
components with a generative representation for Evolving reusable neural modules. In Proc. of
body-brain evolution. Artificial Life, 8(3):223–246, GECCO. Springer, 2004.
2002. [39] N. Rougier. Dynamic neural field with local inhibition.
[21] C. Igel, W. Erlhagen, and D. Jancke. Optimization of Biological cybernetics, 94(3):169–179, 2006.
neural field models. Neurocomputing, 36(1-4):225–233, [40] N. Rougier and J. Vitay. Emergence of attention
2001. within a neural population. Neural Networks,
[22] A. Ijspeert, A. Crespi, D. Ryczko, and J. Cabelguen. 19(5):573–581, 2006.
From swimming to walking with a salamander robot [41] A. Slocum, D. Downey, and R. Beer. Further
driven by a spinal cord model. Science, experiments in the evolution of minimally cognitive
315(5817):1416–1420, 2007. behavior: From perceiving affordances to selective
[23] A. Ijspeert, J. Hallam, and D. Willshaw. Evolving attention. In From Animals to Animats 6, pages
swimming controllers for a simulated lamprey with 430–439, 2000.
inspiration from neurobiology. Adaptive Behavior, [42] A. Soltoggio, J. A. Bullinaria, C. Mattiussi, P. Dürr,
7(2):151–172, 1999. and D. Floreano. Evolutionary Advantages of
[24] T. Kondo. Evolutionary design and behavior analysis Neuromodulated Plasticity in Dynamic, Reward-based
of neuromodulatory neural networks for mobile robots Scenarios. Artificial Life, 11:569, 2008.
control. Applied Soft Computing Journal, [43] K. Stanley, D. D’Ambrosio, and J. Gauci. A
7(1):189–202, 2007. Hypercube-Based encoding for evolving Large-Scale
[25] J. Krichmar and G. Edelman. Brain-based devices for neural networks. Artificial Life, 15(2):185–212, 2009.
the study of nervous systems and the development of [44] K. O. Stanley and R. Miikkulainen. Evolving neural
intelligent machines. Artificial Life, 11(1-2):63–77, networks through augmenting topologies. Evolutionary
2005. Computation, 10(2), 2002.
[26] W. Lohmiller and J.-J. E. Slotine. Contraction [45] T. Trappenberg, M. Dorris, D. Munoz, and R. Klein.
analysis for nonlinear systems. Automatica, A model of saccade initiation based on the
34(6):683–696, 1998. competitive integration of exogenous and endogenous
[27] D. Marbach and A. J. Ijspeert. Online optimization of signals in the superior colliculus. Journal of Cognitive
modular robot locomotion. In Proc. of ICMA’2005, Neuroscience, 13(2):256–271, 2001.
pages 248–253, 2005. [46] J. Vitay, N. Rougier, and F. Alexandre. A distributed
[28] J.-A. Meyer, S. Doncieux, D. Filliat, and A. Guillot. model of spatial visual attention. In Biomimetic
Biologically Inspired Robot Behavior Engineering, Neural Learning for Intelligent Robots, pages 54–72.
chapter Evolutionary Approaches to Neural Control of Springer, 2005.
Rolling, Walking, Swimming and Flying Animats or [47] B. Webb. Robots in invertebrate neuroscience. Nature,
Robots, pages 1–43. Springer-Verlag, 2002. 417(6886):359–363, 2002.
[29] J.-A. Meyer, A. Guillot, B. Girard, M. Khamassi,
P. Pirim, and A. Berthoz. The Psikharpax project:
Towards building an artificial rat. Robotics and
Autonomous Systems, 50(4):211–223, 2005.

Das könnte Ihnen auch gefallen