Vous êtes sur la page 1sur 11

Biologically Inspired Cognitive Architectures 19 (2017) 21–31

Contents lists available at ScienceDirect

Biologically Inspired Cognitive Architectures


journal homepage: www.elsevier.com/locate/bica

Behavior planning of intelligent agent with sign world model


Aleksandr I. Panov
Federal Research Center ‘‘Computer Science and Control” of RAS, Moscow, Russia
National Research University Higher School of Economics, Moscow, Russia

a r t i c l e i n f o a b s t r a c t

Article history: Behavior planning is an important function of any complex technical facility intelligent control system.
Received 25 October 2016 Presently, a symbol paradigm of artificial intelligence offers a variety of planning algorithms, including
Revised 21 December 2016 those that use precedent information, i.e. algorithms based on acquired knowledge. A symbol grounding
Accepted 23 December 2016
problem within the exiting approaches of knowledge representation does not allow effective use the
developed algorithms together with learning mechanisms for the purpose of solving a wide variety of
applied problems by actual intelligent agents (robotics systems). This article presents the original plan-
Keywords:
ning algorithm (MAP Planner), which uses a sign world model as the basis for acquisition and mainte-
Behavior planning
Sign world model
nance of knowledge for future use in behavior planning. the sign problem approach describes planning
Sign image as a cognitive function actualized by the world model of a subject of activity. Apart from solving symbol
Sign significance grounding problems and ensuring psychological and biological plausibility, a sign planning process
Sign personal meaning model allows interaction of an intelligent agent with other participants in solving a cooperative task.
Causal matrix The article presents the description of the knowledge representation method used, a MAP planning algo-
Semiotic network rithm, and a model experiment in a ‘‘block world”.
MAP algorithm Ó 2016 Elsevier B.V. All rights reserved.
Case-based planning

Introduction multi-agent applications, which suggests that agents possess both


different action sets and different knowledge of the environment
The issue of behavior planning of a complex technical or virtual Brafman (2015). In case of cooperative interaction, it is also neces-
subject has a long history and is mainly associated with the suc- sary to ensure non-discretionary incorporation of learning ele-
cesses in specific area of the Artificial Intelligence discipline — ments to augment the database of an agent with information
automatic planning. In this sphere, considerable success has been supplied by other participants of the group.
achieved and a number of sign-based planning methods were In more recent times, researchers in the fields of control and
proposed — both for classical problem definition, where actions planning theory have focused on psychologically and biologically
are deterministic (these are such planning algorithms as FF inspired models and architectures of agent control Kelley (2006)
Hoffmann & Nebel (2001), FD Helmert (2006), LAMA Richter & and Sun and Hélie (2012). The use of different types of memory
Westphal (2010)), and in non-deterministic definition, which takes (episodic, procedural, etc.) in cognitive architectures is aimed
into consideration the nonzero probabilities of non-appliance of exactly at solving the task of reproducing biological and psycholog-
actions and probabilistic environment reaction (algorithms based ical methods of information interchange and organization to solve
on Markov processes and dynamic programming Barto, Bradtke, such problems as behavior control and planning. This is primarily
& Singh (1995) and Bonet & Geffner (2009)). However, the develop- driven by the fact that the increasing complexity of tasks per-
ment of effective and fast planning algorithms is based on the formed by robotics systems (agents) requires their higher self-
preset heuristic graph search principles and on the assumption sufficiency, versatility, and flexibility, which the existing methods
that a set of actions is known in advance, which makes a planning and algorithms are unable to provide. Researchers in the field of
systems automatic adaptation to new problems with a new list of Artificial Intelligence once again turn to the natural examples of
actions impossible. This implies that classical approaches do not such problem solving — to the research of human and animal
offer a carry-over of planning experience or abstract actions, which behavior Redko and Burtsev (2016) and Panov and Yakovlev
may have varying realizations in different situations. Substantial (2016b). Psychologically inspired models of cognitive functions
challenges arise when the existing algorithms are adapted for (including planning) are focused both on reproducing human
behavior in complex, specifically cooperative conditions, and at
complying, as fully as possible, with the existing psychological
E-mail address: pan@isa.ru

http://dx.doi.org/10.1016/j.bica.2016.12.001
2212-683X/Ó 2016 Elsevier B.V. All rights reserved.
22 A.I. Panov / Biologically Inspired Cognitive Architectures 19 (2017) 21–31

concepts of human mind functioning. On the one hand, this may This article is further organized as follows: Section ‘Sign world
result in an increased resource intensity of the proposed algo- model’ introduces the main concepts used in the article: a world
rithms, but, on the other hand, it will allow a realization of new model, as well as a sign and its components are defined and
possibilities, which previously had been left out of the scope of substantiated from psychological and biological standpoint.
problems tackled by planning specialists, such as goal-setting or Section ‘Sign elements’ introduces the concept of a causal matrix
role designation capacity. In the past, cognitive psychology con- as a mathematical structure for the description of sign components
cepts have been also used in classical planning; however, mainly and considers its main characteristics. Section ‘Causal network’
in behavioristic agendas. For instance, the concept of dividing the discusses the networks, which are formed on the basis of sets of
entire multiplicity of actions into automatic, fast actions, specific, causal matrices and which represent relations of sign components.
voluntary actions and generalized actions predicted by psycholog- Section ‘Semiotic network’ introduces the concept of semiotic net-
ical theory Kahneman (2011) was implemented by hierarchical work as a model of world model and discusses the main types of
planning and the concept of planning experience maintenance — processes of activity propagation within a semiotic network. Sec-
in precedent planning Hammond (1990), De La Rosa, Garcia- tion ‘Planning in sign world model’ presents the description of a
Olaya, and Borrajo (2013) and Borrajo, Roubíčková, and Serina MAP algorithm of behavior planning in a sign world model (in a
(2015). semiotic network). Section ‘A model experiment: cube world’ con-
Cognitive psychology has a number of branches that study the cludes with a model example of operation of the presented MAP
phenomenon of planning, within which three main areas should Planner.
be mentioned: planning as part of a cognitive scheme Neisser
(1976), planning as a meta-process Flavell (1979) and Sternberg
et al. (2000) and planning as part of an activity Leontyev (2009). Sign world model
The first branch uses cognitive schemes to describe behavior of
humans. For example, a perceptive scheme is a program of gather- In this article, the method of knowledge representation is
ing information about objects and events, as well as acquisition of based on a sign world model Osipov et al. (2014), Osipov,
new information to provide its consistent interpretation. The Panov, and Chudova (2015) and Osipov (2015), which both stores
scheme simultaneously incorporates a plan and its implementa- knowledge about objects, processes and relations of external
tion; it is both an actionable structure and a structure of actions. environment and represents the internal parameters of the intel-
The second approach provides for the existence of metacognitive ligent agent that determine its motivational constituent and
processes allowing a person to control his/her cognitive processes activity experience. The world model also includes the proce-
and knowledge. From Sternberg’s point of view, one may talk about dures of operation with knowledge, its acquisition and its use
global (strategic) and local (tactical) planning. Global planning in various processes, such as perception, reasoning, goal-setting
requires more time, but this is compensated by the reduction of and behavior planning Osipov et al. (2015). The representation
time dedicated to local, tactical planning. Finally, the third of the world model is based on psychological concepts of human
approach, which is one of the most general concepts, considers brain functioning, in particular on the concepts of cultural and
hierarchical activity theory. This theory is used in this article and historical approach Vygotsky (1986), activity theory Leontyev
is described in the following section. (2009), Verenikina and Gould (1998) and Igira and Gregory
It is also worth mentioning that psychologically and biologically (2009) and dual systems Evans and Stanovich (2013) and
inspired control and planning models provide a new perspective to Stanovich (2009). According to psychological views, a world
the symbol grounding problem Harnad (1990), Barsalou (1999), model component is a four-element structure: a sign, which
Chella, Frixione, and Gaglio (2003) and Besold and Kuhnberger represents all entities of external environment and inner space
(2015). Neurophysiological models of brain cortex sensor region for the subject (in our case, an intelligent agent); objects and
functioning together with psychological categorization and per- their properties; processes; and relations between objects and
ception theory form the basis for the development of new consis- processes. It should be noted that a sign is a product of interac-
tent models of association of symbols and sensor data. Success in tion between several subjects of activity forming a certain group
this field has made it possible to implement certain models in (a cultural environment), thus, the concept of a sign inherently
robotic systems Heintz, Kvarnstrom, and Doherty (2010). assumes that an individual’s world model interacts with the
This article will present a new psychologically and biologically world models of other individuals.
inspired method of behavior planning based on sign theory of Below we will give informal definitions and examples of a sign
activity and structural models of cortical-thalamic regions of brain. and each its components and then introduce a formal structure
Apart from its value in terms of modeling of human cognitive func- (causal matrix) (see Section ‘Sign elements’) to define all of them
tions, sign approach may be used to solve a number of cooperative more precisely.
robotics problems (e.g., for intelligent movement problems Panov An image element of a sign holds specific attributes of the rep-
& Yakovlev (2016b, 2016a)), which cannot be solved by classical resented entity and, at the same time, is a function of representa-
or other psychology-oriented methods (such as BDI Sardina, tion of the entity on the basis of the stream of data drawn from
Silva, & Padgham (2006)). both external and internal sensors, in which key attributes are dis-
The main purpose of the article is to demonstrate the sign tinguished. An image element is individual for each bearer of a
approach for modeling of such an important cognitive process as world model and is formed as the result of observation and gener-
a behavior planning. The proposed planning method (MAP- alization Osipov et al. (2015) and Skrynnik et al. (2016).
algorithm) does not build more efficient plans than other existing An element of significance for a sign represents the generalized
planners and is faced with the same problems (such as a Sussman conceptual knowledge of a subject about the entities of the exter-
anomaly). Also considered is that MAP-algorithm does not use all nal environment, as well as about the internal space, both its own
components of a sign leading to a simplified model similar to a and that of other participants of the group. This knowledge is
frame approach and rule systems. However, inclusion of the learn- coherent, that is, similar for all representatives of the group. Com-
ing process and another components of the sign enables us to munication processes occurring in a group of subjects (intelligent
describe the goal setting process and models of coalition formation agents) are based on the messages built with signs having common
Skrynnik, Petrov, and Panov (2016) and Osipov, Panov, and significance, which in such way determine the syntax of the com-
Chudova (2014). munication protocol.
A.I. Panov / Biologically Inspired Cognitive Architectures 19 (2017) 21–31 23

A sign’s personal meaning element contains the individual per- (actualization) of the sign, or, in other words, in the formation of
sonal experience of the subject’s interaction with the external the concept of a mediated object, or process, based on perceptual
environment with regard to the attitude to this experience — information coming from the external environment and registered
whether it served to achieve a certain goal (satisfy a certain need) by internal motor-information sensors. Before giving it a name, let
or, conversely, was unsuccessful. The personal meaning of a sign is us call the sign proto-sign or feature.
its dynamic characteristic, which is being constantly shaped and Let us assume that within the incoming data flow a sequence
updated as the result of certain cognitive processes (planning or ðx1 ; x2 ; . . . ; xh Þ is isolated, having the length of h and consisting of
goal-setting). It is a personal meaning element, which is deter- real number vectors from 0 to 1, which we will call events. Each
mined by the inner characteristics of a subject and its sphere of event xt with the length of q represents a record of output of q
needs and motivations. sensors and each component of the event means a certainty of trig-
Finally, the fourth component — a name — serves to identify a gering of the sensor. As an example, event ð0:1; 0:9; 0:9Þ arrives
sign both for communication processes and for voluntary processes from three sensors — red, blue and green light transducers, and sig-
of planning and reasoning. A sign’s name, as well as its significance, nifies that the certainty of the red light transducer’s triggering is
is an established element, which, within a group of subjects, is 10%, and of blue and green light transducers, 90% each.
altered in the slowest manner out of all the signs elements. An image element of the sign shall determine, based on the
Within a subject’s world model, signs represent the static input sequence, whether the mediated object or process is present
objects and the properties of the external environment, as well (coded) in this sequence. To do this we will code characteristic
as its dynamic components: processes, situations and internal attributes of the object or process in a special structure — the cau-
characteristics of the agent: actions, objects and properties sal matrix z ¼ ðe1 ; e2 ; . . . ; eh Þ having the scale of q by h where q is
‘‘of the internal environment”. Let us assume that we have an dimension of input events and h is the length of sequence of the
object of external environment — lemon. Within the world model input events. Here every column et of the causal matrix is a bit vec-
of a subject, it may be represented by a sign having the name tor having the length of q coding characteristics (to which 1 corre-
‘‘lemon” whose image also includes such attributes as yellow color, sponds) that are necessary for the input event at the moment of t,
oval form and acid flavor. These attributes may also be represented so that the mediated object or process may be recognized within
in the world model by signs or may be the information received input data flow, that is, that determine a collection of simultaneous
directly from sensors. Generalized actions and processes, in which characteristics. As an example, the image of sign s, representing
(according to a common shared opinion of a group to which the ‘‘square”, may be represented by a causal matrix
subject belongs) lemon participates are the significance of the
2 3
‘‘lemon” sign. Lemon is usually eaten, is used as a sauce for a fish 1 1 1 1
course or is used for the prevention of diseases. The personal meaning 6 7
60 1 1 07
of a lemon for a subject includes specific personal actions and pro- 6 7
cesses, in which the subject had the experience of using a lemon
z ¼ 61 0 0 17
6
7;
6 7
for solving a certain problem. I threw a lemon at my classmate or I 41 1 0 05
ate a whole lemon without making a wry face. All actions and pro- 0 0 1 1
cesses may also be represented by certain signs or may not be
brought to a sign’s voluntary level and may be some undesignated where the first line represents the characteristic vector of informa-
operations. tion received by an image angle sensor, the second one line by the
Apart from the psychological basis of the four-element sign position transducer of the visual sensor (upper position), the third
structure, there are neurophysiological evidences proving the exis- line by the lowest position of the sensor, the fourth line by the left
tence of such a structure for the storage and activation of compo- position of the sensor and the fifth line by the right position (see
nents of individual experience Edelman (1987) and Ivanitsky Fig. 1).
(1997). Besides, neurophysiological data serve as the basis for
developing sign element models and certain functions, such as per-
ception and recognition George and Hawkins (2009) and Osipov
et al. (2015). Neurophysiological evidence proving high uniformity
of structures in various regions of the cerebral cortex, as well as
participation of the thalamus in the formation and remembering
of chronological sequences Buxhoeveden and Casanova (2002)
and Constantinople and Bruno (2013), leads to the mathematical
structure of a causal matrix Osipov et al. (2015), employed to
describe the structure of sign elements used in this article.
Symbolic approach to representation of knowledge and descrip-
tion of processes occurring in a sign world model makes it possible
to solve a number of important problems in the area of situational
control Osipov (1997) and Pospelov and Osipov (1997) and control
of complex technical facilities Emel’yanov, Makarov, Panov, and
Yakovlev (2016). The use of sign world model to implement strate-
gic functions of robotics systems Emel’yanov et al. (2016) demon-
strates the applicability of the used approach not only to represent
knowledge, but also to solve cooperative planning and role distri-
bution problems.

Sign elements

Let us discuss the structure of sign elements using the example


of an image element, which participates in the recognition Fig. 1. Visual interpretation of the causal matrix.
24 A.I. Panov / Biologically Inspired Cognitive Architectures 19 (2017) 21–31

Fig. 2. Schema of the causal matrix.

Several causal matrices determining various manifestations of events and may be implemented by various means, including Nor-
the represented object or process (e.g. another order of sensor trig- ris algorithm, FcbO, or AddIntent (Krajca, Outrata, & Vychodil,
gering) may correspond to the image of each sign. We will desig- 2010; Merwe, Obiedkov, & Kourie, 2004; Norris, 1978).
nate the complete sequence of causal matrices of the sign s In the case when for matrices Z p ðsÞ of the sign s image, the set of
image as Z p ðsÞ. effect columns is empty, Ie ¼ £, that is, when it is impossible to
The case, in which a sign’s characteristics are data received by unambiguously determine what events always precede others we
sensors, is a special one. In a more general statement, a sign’s will assume that a cause-and-effect relationship was not
image attributes are other signs representing these specific charac- established and the sign mediates a certain object or situation.
teristics. Thus, we are capable of associating the image of sign s Otherwise, we will assume that the sign mediates a certain action
with set Sp ðsÞ, having the potency of q. Each element of the set cor- or process, the result of which is coded in effect columns, and the
responds to the number of lines of causal matrix z, having the condition is coded in condition columns.
dimensions of q by h, that is, each feature si 2 Sp ðsÞ corresponds The following statements in respect to properties of procedure
to an attribute bit vector setting at position 1 moments in time, Kp are valid:
when this attribute should be present in input data to actualize
the sign s (recognize the image of the sign).  Ic \ Ie ¼ £ — a column of the causal matrix may not be simulta-
To refine the definition of the set Sp ðsÞ, let us introduce a class of neously condition and effect;
binary relations f@p ; @1p ; @2p ; . . .g determined by Cartesian product  jIc [ Ie j ¼ h — a column of the causal matrix is either a condition
S  S. Let us suppose that the sign si is an image component of sign or an effect;
s; ðsi ; sÞ 2 @p or si @p s if si 2 Sp ðsÞ. If it is known that the sign si cor-  Ic : ¼ £ — among causal matrix columns there must be at least
responds to 1 in tth column of a certain causal matrix z 2 Z p ðsÞ of one condition column, whereas there may be no effects (in the
sign s, we shall use nested relation @tp  @p . case of object features);
 8i 2 Ie ; j 2 Ic i > j — all conditions precede effects.
Causal network
Taking the above into consideration, a causal matrix diagram is
Let us introduce a special procedure Kp : 2 Z ! 2N  2N , which shown at Fig. 2.
associates to every list of causal matrices Z p ðsÞ  Z of the sign s Let us now introduce the concept of a causal network, which
image two non-overlapping subsets of indexes of its own columns will determine the heterarchy of a set of images. Causal network
Ic  N; 8i 2 Ic i 6 h (indexes of condition columns) and Ie  N; W p ¼ hV p ; Ep i is a marked orgraph in which
8i 2 Ie i 6 h (indexes of effect columns). Kp ðZ p ðsÞÞ ¼
c e c e  every node v 2 V p is associated with a sequence of causal matri-
ðI ; I Þ; I \ I ¼ £. For example, if for a set of matrices
ces Z p ðsÞ of a certain sign s image, which we will designate as
Z ¼ fðð1; 0Þ; ð0; 1ÞÞg procedure Kp returns two sets f1g and f2g this
means that the occurrence of the feature corresponding to the first v ! Zp ðsÞ;
line of the matrix causes an occurrence of the feature correspond-  edge e ¼ ðv 1 ; v 2 Þ belongs to a set of graph edges E if
ing to the second line. In essence, the procedure Kp is a function of v 1 ! Z p ðs1 Þ; v 2 ! Zp ðs2 Þ and s1 2 Sp ðs2 Þ, i.e. if the sign s1 is an
establishing a cause-and-effect relationship at the set of input element of sign image s2 ;
A.I. Panov / Biologically Inspired Cognitive Architectures 19 (2017) 21–31 25

 each graph edge e ¼ ðv 1 ; v 2 Þ; v 1 ! Z p ðs1 Þ; v 2 ! Z p ðs2 Þ is associ- causal matrices hzm 1 ðsÞ; z2 ðsÞ; . . .i corresponding to node wm ðsÞ of
m

ated with the label  ¼ ð1 ; 2 ; 3 Þ, a tuple of three natural the causal matrix for significances; and a is the personal meaning
numbers: of the sign, a sequence of causal matrices hza1 ðsÞ; za2 ðsÞ; . . .i corre-
– 1 — index of the initial matrix of the sequence Z p ðs1 Þ may sponding to node wa ðsÞ of the causal matrix for meanings.
possess a special value of 0 if any matrices of the sequence Let us define semiotic network as the quintuple
may serve as source matrices; X ¼ hW p ; W m ; W a ; Rn ; Hi where.
– 2 is the target matrix in the sequence Z p ðs2 Þ the line of
which is associated with feature s1 ;  W p ; W m ; W a are causal networks within the set of images, sig-
– 3 is the index of a column in the target matrix in which the nificances and personal meanings, respectively;
value of the line corresponding to feature s1 is 1, and the  Rn is a class of relations within the set of signs generated, based
index may possess positive values (condition columns) or on the three causal networks, that is Rn ¼ fRp ; Rm ; Ra g;
negative values (effect columns).  H is the class of operations of a set of signs, which are generated
based on three types of causal networks to which respective
A causal network represents a certain set of overlapping hierar- sign elements belong (see Osipov et al., 2014 for more detail).
chies of signs. Every sign is represented by a set of causal matrices
determining the image of the sign, and hierarchy represents hier- We must emphasize one more time that a sign represents not
archical relations between images. This relation may be read as fol- only the objects of the external world, but also the processes occur-
lows: ‘‘sign x participates in the formation of image of sign y”. Here ring in it, realizable actions, as well as the situations observed in
we specify for what specific matrix of sign y and for what specific the external environment. The three types of causal networks com-
column of this matrix sign x is required (labels 2 and 3 accord- posing a semiotic network are not independent from each other.
ingly). In some cases, we may also state sign matrix x (label 1 ) par- The following unambiguous correspondence is established among
ticipating in the process. The example of such a network is shown the nodes of each network: for each node wx of network W x , such
in Fig. 3. unique nodes wy and wz may be found within networks W y ; W z
Similarly, causal networks are determined for the remaining (x; y; z 2 fp; m; ag) so that all three nodes correspond to the same
elements of the sign — for significance and personal meaning. For sign s ¼ hwp ; wm ; wa i. A sign’s name is the label of nodes in each
each sign s, sets Sm ðsÞ and Sa ðsÞ are determined, that is, classes of of the networks: in the causal network, there may be only one node
relations f@m ; @1m ; @2m ; . . .g and f@a ; @1a ; @2a ; . . .g are determined. with this label and nodes of all networks with the same labels com-
Set Sm ðsÞ is interpreted as the role composition of sign s, for exam- pose elements of the sign. To associate causal matrices having dif-
ple, a subset element or action role. Set Sa ðsÞ is interpreted as the ferent types of nodes within one sign, special binding functions are
immediate element composition of a certain situation observed used: Wm p ; Wm ; Wa , and functions opposite to them Wm ; Wa ; Wp
a p p m a

and experienced presently by a subject, which is a bearer of a Osipov et al. (2014). Every binding function associates a causal
world model. Similarly, sets Z m ðsÞ and Z a ðsÞ, as well as procedures matrix of one type with a causal matrix of another type or gener-
Km and Ka are determined. ates this matrix in case it is absent in the respective node of the
network.
Semiotic network Let us introduce the concepts of activity in a semiotic network
and of its propagation process. Let us introduce a label of activity
Let us define sign s as a quadruple hn; p; m; ai where n is the for causal matrices of network W x (x 2 fp; m; ag) and call matrix
name of the sign; p is the image of the sign, a sequence of causal set Z x , having the label ‘‘active”. An activity propagation process
matrices hzp1 ðsÞ; zp2 ðsÞ; . . .i corresponding to node wp ðsÞ of the causal is an alteration of the composition of set Z x over time (at each dis-
matrix for images; m is the significance of the sign, a sequence of crete moment of time) and is described for each type of causal net-
work with the following intrinsic function: ua ; um ; up . An activity
propagation process is iterative, i.e., at each step a new composi-
tion of active matrices is initiated based on the previous composi-
tion and it depends on the matrices included in it. In the simplest
case, we will discuss the process in which each matrix does not
influence the course of activity propagation from another matrix,
and consequently, we will assume that functions ux accept one
active matrix as an input and generate a new subset of active
matrices.
Due to the fact that the edges of causal networks have direc-
tions, we will distinguish the propagation of activity upward
within a network, if edges outbound from a node (u"x functions)
are used, and the propagation of activity downward within a net-
work, if edges inbound to a node (u#x functions) are used. Further,
describing a planning algorithm we will need only the functions of
significances and personal meaning networks. We will parameter-
ize each function u"x ; u#x by the depth of activity propagation dx ,
which shows at which depth edges can be viewed in a given direc-
tion (upward or downward).
Further, when describing a planning algorithm we will use the
Fig. 3. Schema of the causal network. Causal matrices are displayed as squares, concept of causal network fragment. We will define a fragment F
condition columns are the left white part of the square, effect columns - the right as a certain set of nodes V of network W x ¼ hV x ; Ex i together with
black part of the square. Label 1 is placed at the beginning of each arrow, label 2 is
all the edges E connecting them: F ¼ hV; Ei : V # V x and
defined as square number in the oval node, label 3 is placed at the end of each
arrow. 8e ¼ ðv 1 ; v 2 Þ 2 Ev 1 2 V; v 2 2 V.
26 A.I. Panov / Biologically Inspired Cognitive Architectures 19 (2017) 21–31

Planning in sign world model In brief, MAP algorithm performs an iterative generation of cau-
sal matrices znext of personal meanings based on the current active
The process of planning in a sign world model is implemented matrix zcur until the maximum number of steps imax is reached (step
with the use of the MAP algorithm and is implemented inversely 10), or the initial matrix zstart (step 41) corresponding to the per-
- from the final to the initial situation. Let us briefly describe the sonal meaning astart of the initial situation is completely activated.
main stages of the algorithm’s operation. A problem definition is A matrix corresponding to the personal meaning of target situation
at the input, zagoal (step 6) serves as an active causal matrix for the first iteration.
Upon completion of all iterations, the plans found are sorted by
T ¼ hNT ; S; Sit start ; Sit goal i; length (step 7), and the shortest of them represent the solution
of the planning problem in the sign world model (step 8).
where NT is the problem identifier, S is the set of signs of a semiotic
The first stage of MAP iteration is S-stage. Its essence is that the
network, Sitstart ¼ hidstart ; £; £; astart i is the initial state having the
search for precedents is performed within the world model of an
meaning of astart ¼ fzastart g and Sitgoal ¼ hidgoal ; £; £; agoal i is the target
intelligent agent, i.e., a search for actions that were performed in
situation agoal ¼ fzagoal g where zai is the causal matrix of the corre-
current conditions zcur . To achieve that, all signs in world model S
sponding personal meaning. In this case significances and images are reviewed, as well as their personal meanings aðsÞ (steps 13–14).
of situations are empty because we don’t consider recognition and If current conditions zcur are satisfied with matrix za , the list of
communication part of the task; we use only personal information ^ case is supplemented by the activity in the personal
precedents A
about situations. In general, problem T is the result of a ‘‘designa-
meaning networks of sign s for the distance of da (step 16).
tion” procedure, that is, the result of forming a world model using
the initial descriptions of planning domain D, which sets the lists
of possible actions and object types, as well as planning problems 9: function MAP_ITERATION(zcur ; zstart ; Plancur ; i)
P. The world model includes the definitions of the starting 10: if i P imax then
conditions and of the final goal (step 1). 11: return £
The result of MAP planning is a plan. 12: A ^ case :¼ £ // List of precedents
Plan ¼ fhzas1 ; zap1 i; hzas2 ; zap2 i; . . . ; hzasn ; zapn ig, a sequence with the length // S-stage
of n pairs hzasi ; zapi i, where zasi is the causal matrix of a certain network // Search of action cases for current conditions
13: for all s 2 S do
node for personal meanings representing a ith planning situation,
14: for all za 2 aðsÞ do
and zapi is the causal matrix of certain personal meaning represent-
15: if za P zcur then
ing the action used in situation zasi . Here, the situation zasiþ1 is the
16: A ^ case [ u" ðs; da Þ
^ case ¼ A
a
result of the performance of action zapi in the meaning that will
be elaborated further during the discussion of the algorithm,
zas1 :¼ zastart is the causal matrix corresponding to the meaning of
After that, MAP algorithm switches to M-stage, at which the
the initial situation, and zasn ¼ zagoal Þ is the causal matrix correspond-
propagation of activity through personal meaning is performed
ing to the meaning of the target situation.
over the distance of da for the purpose of activating all signs related
to the current situation (step 17). The components of the obtained
Input: planning domain D, planning task P, maximal iteration set of causal matrices A serve as the starting points for activity
deep imax propagation over the network of significances: using binding func-
Output: plan Plan tion za , for each matrix Wm
a , the required node is determined in the
1: T = hN T ; S; Sitstart ; Sit goal i :¼ GROUND(P) causal network of significances, from which activity propagates
// N T - task id, S - set of signs, Sitstart ¼ hidstart £; £; fzastart gi over the distance of dm (step 20). If the activated matrices are cau-
- start situation with meaning astart ; Sitgoal ¼ sal, they are added to the set of active significances M  (step 22).
hidgoal £; £; fzagoal gi - goal situation with meaning agoal
// M-stage
2: Plan :¼ MAP_SEARCH(T)
// Spreading of activity downward in the personal
3: function MAP_SEARCH(T)
meaning network
4: zcur :¼ zagoal
17: A ¼ u#a ðzcur ; da Þ
5: zstart :¼ zastart
18: M ¼ £
6: Plans :¼ MAP_ITERATION(zcur ; zstart ; £, 0)
19: for all za 2 A do
7: fPlan0 ; Plan1 ; . . .g = SORT(Plans)
// Spreading of activity upward in the significance
8: return Plan0
network
20: for all zm 2 u"m ðsðza Þ; dm Þ do
21: if Ie ðzm Þ – £ then
The planning process is hierarchical and consists in the repeti- 22: M :¼ M  [ fzm g
tion of MAP iterations, including four stages (see Fig. 4):

 S-stage is the search for a precedent for performing actions in Next, we switch to A-stage, at which a causal matrix generation
the current situation. is performed in the personal meaning network, the former repre-
 M-stage is the search for applicable actions in the network of senting actions zcur specified in relation to current conditions
significances. determined by the active significances of the set M  . For this pur-
 A-stage is the generation of personal meanings, which corre- pose, steps 25–27 are performed, during which the propagation
spond to the significances found. of activity in the causal network of significances over the distance
 P-stage is the building of a new situation using the feature set of of dm leads to the activation of a set of significances M  of signs
conditions for the actions found. related to role structure of procedure matrix zm . After that, using
A.I. Panov / Biologically Inspired Cognitive Architectures 19 (2017) 21–31 27

Fig. 4. Schema of the behavior planning process.

binding function Wam , a new causal matrix is generated in personal ei 2 fek jek 2 za ; k 2 Ic ðza Þ or which belong to the current active cau-
meaning network copying values zm , replacing abstract signs-roles sal matrix and are not condition columns of action
with object signs associated with roles by a set-subset relation. ei 2 zcur ^ ei R fej jej 2 za ; j 2 Ie ðza Þg. In the current plan Plancur , cur-
Further, at A-stage, selection of causal matrices is performed that rent conditions are applicable to action pair hzcur ; za i. If the new
represent the actions in current conditions zcur (steps 30–32) is situation does not overlap the starting situation (step 41), itera-
performed. For this purpose, all causal matrices, the effects of tions continue with the new current situation thus supplementing
which are not included in the current situation, are deleted (it all sets of generated plans Plansfin .
should be borne in mind that planning is performed inversely). Constants da ; dm , which determine the depth of activity propa-
Finally, at A-stage, one of the operations within world model ha gation within causal networks, are parameters of the algorithm
is performed, which provides, in this case, meta-adjustment — and determine internal characteristics of the world model bearer,
checking for a certain heuristic rule, which may mean, for example, varying from agent to agent. Usually, in model experiments these
that no action may be repeated, or that it is better to perform the parameters do not exceed 5.
action that advances fastest to the initial conditions zstart
(step 33). Any heuristic rule can also be represented by a causal
matrix of the sign’s personal meaning representing the internal //P-stage
strategy of its behavior planning. 36: Plansfin :¼ £
37: for all za 2 A ^
//A-stage 38: Plancur ¼ Plancur [ fhzcur ; za ig
23: ^ gen ¼ £
A //Generation of new situation - action application
24: for all zm 2 M  do 39: znext :¼ ðei jðei 2 zcur ^ ei R fej jej 2 za ; j 2 Ie ðza ÞgÞ_
//Spreading activity downward in the significance ei 2 fek jek 2 za ; k 2 Ic ðza ÞgÞ
network 40: Sitnext ¼ hidnext ; £; £; fznext gi
25: M  ¼ u#m ðzm ; dm Þ 41: if znext P zstart
26: for all zm 2 M  do 42: Plansfin ¼ Plansfin [ fPlancur g
27: ^ gen :¼ A
A ^ gen [ fWa ðz Þg 43: else
m m
//Merging of activity of formed meanings and meaning of 44: Plansrec :¼ MAP_ITERATION(znext ; zstart ; Plancur ; i þ 1)
the current situation 45: Plansfin ¼ Plansfin [ Plansrec
28: A ^¼A ^ gen [ A
^ case 46: return Plansfin
29: for all za 2 A ^ do
30: zshift ¼ ðei ji 2 Ie Þ
A model experiment: cube world
31: if zcur : P zshift then
32: ^¼A
A ^ n fza g Let us demonstrate the work of the presented behavior plan-
// Checking of metacognitive heuristic ning algorithm by a model experiment in which the planning
33: A ^
^ ¼ fha ðza Þjza 2 Ag domain is a ‘‘block world” example, well known in the field of
^
34: if A ¼ £ then automation planning Gupta and Nau (1992). The description of
35: return £ the domain in PDDL language Gerevini, Haslum, Long, Saetti, and
Dimopoulos (2009) consists of type definition of (blocks), four
predicates (ontable, clear, handempty, holding) and four actions
(pick-up, put-down, stack, unstack) (see Table 1).
MAP algorithm concludes with P-stage. Here, for every gener- Using MAP algorithm, let us make an example of a solution to a
ated causal matrix za , which represents a certain action, new situ- planning problem — building a tower of four blocks, which lie on
ation Sit next is generated, which is the result of reverse application the table (Table 1). A fragment of causal network of personal mean-
of action in the current conditions zcur . Reverse application (step ings, which determines the causal matrix of meaning named start
39) includes the generation of a causal matrix znext consisting of is shown at Fig. 5. Each separate block (a, b, c, d) has one causal
events, which are condition columns of action matrix situated in a network node, whereas each of the predicates
28 A.I. Panov / Biologically Inspired Cognitive Architectures 19 (2017) 21–31

Table 1
Description of the planning domain ‘‘block world” and a task of tower building (‘‘BLOCKS-4-0”).

(Define (domain BLOCKS) (:requirements: (:Action pick-up:parameters (?x - block): (:Action put-down:parameters (?x - block):
strips:typing) (:types block) (:predicates (on?x - precondition (and (clear?x) (ontable?x) precondition (holding?x):effect (and (not (holding?
block?y - block) (ontable?x - block) (clear?x - (handempty)):effect (and (not (ontable?x)) (not x)) (clear?x) (handempty) (ontable?x)))
block) (handempty) (holding?x - block) ) (clear?x)) (not (handempty)) (holding?x)))

(:Action stack:parameters (?x - block?y - block): (:Action unstack:parameters (?x - block?y - block): (Define (problem BLOCKS-4-0) (:domain BLOCKS) (:
precondition (and (holding?x) (clear?y)):effect precondition (and (on?x?y) (clear?x) objects D B A C - block) (:INIT (CLEAR C) (CLEAR A)
(and (not (holding?x)) (not (clear?y)) (clear?x) (handempty)):effect (and (holding?x) (clear?y) (not (CLEAR B) (CLEAR D) (ONTABLE C) (ONTABLE A)
(handempty) (on?x?y))) (clear?x)) (not (handempty)) (not (on?x?y))))) (ONTABLE B) (ONTABLE D) (HANDEMPTY)) (:goal
(AND (ON D C) (ON C B) (ON B A))) )

Fig. 5. Start situation: all blocks are on the table.

Fig. 6. Goal situation: a tower built from four blocks.

clear and ontable have four matrices in a node, since they partici- defines the relation ‘‘located on” may be represented in the form
pate in events with different blocks. For example, matrices of sign of a procedural causal matrix, so as to demonstrate the nonsymme-
clear are preset in the 1st, 3rd, 5th and 7th columns of the sign try of this relation distinctly, notwithstanding the fact that use of
matrix start simultaneously, with blocks a, b, c and d respectively, the object matrix has no influence on the result. Here, one causal
which means that no blocks are located on any of the blocks. matrix of each block also participates in the situation, and the
Fig. 6 shows the target situation in which all four blocks are predicate on is represented as a node with three causal matrices,
stacked in a tower: block d is located at the table, block c — on d, since it participates in a causal matrix of the target situation goal
b — on c, and finally block a is at the top. Predicate on, which in different columns with three different blocks.
A.I. Panov / Biologically Inspired Cognitive Architectures 19 (2017) 21–31 29

Fig. 7. Fragment of the causal network: representation of the action stack.

Fig. 8. Spreading activity in the significance network, generation of personal meaning of the sign unstack.

Fig. 7 shows a fragment of a causal network for significances causal matrix unstack by matrix a, it is enough to use the value
representing components of a procedural causal matrix of the sign of three edges as constant dm .
stack and relations‘‘set-subset” of objects (blocks), of class block ^ gen occurs in
At A-stage, the generation of new causal matrices A
and roles in action stack: block?x (analog of semantic role ‘‘object”) the personal meaning network by propagating activity down the
block?y (analog of semantic role‘‘directive”). It should be noted network of significances. An example of this propagation, which
here that in the case of ‘‘set-subset” relation (a ! block, block ! results in a new causal matrix of sign unstack, is shown at the right
block?x) 1 edge label v (index of the original matrix of the node, of Fig. 8. The new causal matrix of the personal meaning matrix is a
from which the edge is outbound v) possesses a special zero value copy of a corresponding matrix for the network of significances,
indicating that any causal matrix of this node may be original. In with references pointing to role signs replaced with references
other words, any of the blocks a, b, c or d may play the role of pointing to object non-abstract signs representing blocks. In our
block?x. example, a matrix corresponding to action unstack(a, b) — remove
Let us discuss the stages of MAP algorithm, S, M, A and P-stages, block a from block b will be generated. At this stage, four matrices
at the first iteration of the algorithm. Let us discuss the simplest for each single-place actions will be generated and twelve — for
case when our intelligent agent has not accumulated any experi- two-place actions.
ence of acting in the context of the presented problem. Due to this, A-stage is completed by testing the effects of the generated pro-
at S-stage the set of precedents A^ case will be empty. Taking into con- cedural matrices for applicability in the current situation, and
sideration the fact that planning is performed inversely, at the first among applicable actions those satisfying a certain meta-
M-stage we view the target situation as the current active predic- cognitive (heuristic) rule ha are selected. In our example, of all
tion matrix zcur and propagation of activity downward from it the the generated variants the action unstack(a, b) will be the only
personal meaning network will activate set A , which coincides applicable action. As a heuristic rule, the following greedy rule
with the fragment shown at Fig. 5. The set of significances M  may be used: select actions that bring success the fastest (the
will include the significances of the signs representing blocks a, new situation has more common features with the target
b, c, d — these are all procedural signs associated with them within situation).
meaning network stack, unstack, pick-up, put-down. The left part of At the end of the iteration, at P-stage, new causal matrix znext is
Fig. 8 shows a fragment of a causal network for significances, generated in the personal meaning network of the sign represent-
which includes the procedural causal matrix unstack. To activate ing the next planning situation. In our example, the new current
30 A.I. Panov / Biologically Inspired Cognitive Architectures 19 (2017) 21–31

situation will coincide with the previous with the exception that Besold, T. R., & Kuhnberger, K. U. (2015). Towards integrated neural-symbolic
systems for human-level AI: Two research programs helping to bridge the gaps.
block a is now held by a manipulator and there is nothing on block
Biologically Inspired Cognitive Architectures, 14, 97–110. http://dx.doi.org/
b. A pair hzcur ; za i of the current situation’s causal matrices and the 10.1016/j.bica.2015.09.003.
selected action are added to the current Plancur plan. Since the new bibitemEvans2013 Evans, J., & Stanovich, K. E. (2013). Dual-process theories of
higher cognition: Advancing the debate. Perspectives on Psychological Science, 8,
situation does not include starting situation, we will begin a new
223–241. http://dx.doi.org/10.1177/1745691612460685.
iteration. Bonet, B., & Geffner, H. (2009). Solving POMDPs: RTDP-Bel vs. point-based
In our example, as the result of activity from MAP algorithm the algorithms. In IJCAI international joint conference on artificial intelligence
following plan of 6 actions will be generated: pick-up(c), stack(c, d), (pp. 1641–1646).
Borrajo, D., Roubíčková, A., & Serina, I. (2015). Progress in case-based planning. ACM
pick-up(b), stack(b, c), pick-up(a), stack(a, b). After the agent com- Computing Surveys, 47, 1–39. http://dx.doi.org/10.1093/brain/awf110 <http://
pletes working on this problem, it saves the planning precedent brain.oxfordjournals.org/content/125/5/935.long>.
in its world model: it saves the initial and the final situation in Brafman, R. I. (2015). A privacy preserving algorithm for multi-agent planning and
search. In Proceedings of the twenty-fourth international joint conference on
the form of new signs and generates a new procedural sign, which artificial intelligence (IJCAI 2015) (pp. 1530–1536).
could be called ‘‘build a tower”. The initial situation will be the Buxhoeveden, D. P., & Casanova, M. (2002). The minicolumn hypothesis in
only feature in the condition column of this sign, and the target sit- neuroscience. Brain, 125, 935–951. http://dx.doi.org/10.1093/brain/awf110
<http://brain.oxfordjournals.org/content/125/5/935.long>.
uation will be the only attribute in the effect column. After that, the Chella, A., Frixione, M., & Gaglio, S. (2003). Anchoring symbols to conceptual spaces:
intelligent agent will be able to solve the same problem by finding The case of dynamic scenarios. Robotics and Autonomous Systems, 43, 175–188.
the required action leading to the immediate solution at S-stage. http://dx.doi.org/10.1016/S0921-8890(02)00358-5.
Constantinople, C. M., & Bruno, R. M. (2013). Deep cortical layers are activated
The same situation may occur while solving another problem, lead-
directly by thalamus. Science, 39, 1591–1594. http://dx.doi.org/
ing to a reduction of applicable action search space. 10.1126/science.5801236425.
De La Rosa, T., Garcia-Olaya, A., & Borrajo, D. (2013). A case-based approach to
heuristic planning. Applied Intelligence, 39, 184–201. http://dx.doi.org/10.1007/
Conclusion s10489-012-0404-6.
Edelman, G. M. (1987). Neural darwinism: The theory of neuronal group selection. New
In the classical symbol planning problem definition, Artificial York: Basic Books.
Emel’yanov, S., Makarov, D., Panov, A. I., & Yakovlev, K. (2016). Multilayer cognitive
Intelligence faces the problem of combining symbol planning algo- architecture for UAV control. Cognitive Systems Research, 39, 58–72. http://dx.
rithms with the methods of learning, which allow both to preserve doi.org/10.1016/j.cogsys.2015.12.008 <http://linkinghub.elsevier.com/retrieve/
planning experience and adapt to new conditions. This problem pii/S1389041716000048>.
Flavell, J. H. (1979). Metacognition and cognitive monitoring: A new area of
overlaps with the symbol grounding problem — the problem of cognitivedevelopmental inquiry. American Psychologist, 34, 906–911. http://dx.
associating symbols used in the classical method of knowledge doi.org/10.1037/0003-066x.34.10.906.
representation with real objects, processes, and properties of the George, D., & Hawkins, J. (2009). Towards a mathematical theory of cortical micro-
circuits. PLoS Computational Biology, 5, e1000532. http://dx.doi.org/10.1371/
external environment. These problems are very vividly manifested journal.pcbi.1000532 <http://www.pubmedcentral.nih.gov/articlerender.fcgi?
in the area of implementing learning robotics systems, for which it artid=2749218&tool=pmcentrez&rendertype=abstract>.
is important to associate the symbols used in conceptual planning Gerevini, A. E., Haslum, P., Long, D., Saetti, A., & Dimopoulos, Y. (2009).
Deterministic planning in the fifth international planning competition: PDDL3
and the data obtained by sensors. It should be noted that when a and experimental evaluation of the planners. Artificial Intelligence, 173,
complex technical system is presented with a planning problem 619–668. http://dx.doi.org/10.1016/j.artint.2008.10.012.
in a broad spectrum of conditions, including cooperative ones, Gupta, N., & Nau, D. S. (1992). On the complexity of Blocks-World planning. Artificial
Intelligence, 56, 223–254.
approaches using a preset, albeit replenished knowledge base
Hammond, K. J. (1990). Case-based planning: A framework for planning from
prove to be inefficient. A method of representing knowledge serv- experience. Cognitive Science, 14, 385–443. http://dx.doi.org/10.1016/0364-
ing as the basis for the control functions of an intelligent agent, 0213(90)90018- R.
should inherently support the possibility to associate symbols with Harnad, S. (1990). Symbol grounding problem. Physica, 42, 335–346. http://dx.doi.
org/10.4249/scholarpedia.2373. Available from arXiv:9906002 <http://
sensor data, as well as support the representation of both internal eprints.soton.ac.uk/271345/5/Harnad-CangelelosiComm.rtf>.
information and generalized information coordinated between Heintz, F., Kvarnstrom, J., & Doherty, P. (2010). Bridging the sense-reasoning gap:
other participants of the group. This article solves the said prob- DyKnow - Stream-based middleware for knowledge processing. Advanced
Engineering Informatics, 24, 14–26. http://dx.doi.org/10.1016/j.aei.2009.08.007.
lems by using a sign world model. We present an original planning Helmert, M. (2006). The fast downward planning system. Journal of Artificial
method (MAP algorithm), which uses and maintains the precedent Intelligence Research, 26, 191–246. http://dx.doi.org/10.1613/jair.1705.
information in the process of plan generation. The four-component Available from arXiv:1109.6051v1.
Hoffmann, J., & Nebel, B. (2001). The FF planning system: Fast plan generation
world model component (sign) used allows coding both the through heuristic search. Journal of Artificial Intelligence Research, 14, 253–302.
information of the external environment and internal attributes, Igira, F. T., & Gregory, J. (2009). Cultural historical activity theory. Handbook of
motivation and need properties, as well as general collective research on contemporary theoretical models in information systems, 434–454.
http://dx.doi.org/10.4018/978-1-60566-659-4.ch025.
knowledge. The algorithm presented may also be used for the gen- Ivanitsky, A. M. (1997). Information synthesis in key parts of the cerebral cortex as
eration of cooperative plans. To demonstrate MAP Planner, a model the basis of subjective experience. Neuroscience and Behavioral Physiology, 27,
example of the generation of the plan for one of the ‘‘block world” 414–426.
Kahneman, D. (2011). Thinking fast and slow. New York: Penguin.
problems is presented. Software support and model experiments
Kelley, T. D. (2006). Developing a psychologically inspired cognitive architecture for
may as well be found at https://github.com/cog-isa/map-planner. robotic control: The Symbolic and Subsymbolic Robotic Intelligence Control
System (SS-RICS). International Journal of Advanced Robotic Systems, 3, 219–222.
http://dx.doi.org/10.5772/5736.
Acknowledgment Krajca, P., Outrata, J., & Vychodil, V. (2010). Advances in algorithms based on CbO. In
M. Kryszkiewicz, & S. Obiedkov (Eds.), Proceedings of the 7th international
This work was supported by the Russian Foundation for Basic conference on concept lattices and their applications (pp. 325–337). CEUR.
Leontyev, A. N. (2009). The development of mind. kettering: Erythros press and
Research (Project No. 16-37-60055).
media. <http://marxists.org/archive/leontev/works/development-mind.pdf>.
Merwe, D. V. D., Obiedkov, S., & Kourie, D. (2004). AddIntent: A new incremental
References algorithm for constructing concept lattices. In P. Eklund (Ed.), Concept lattices
lecture notes in computer science (pp. 372–385). Berlin Heidelberg: Springer.
<http://link.springer.com/chapter/10.1007/978-3-540-24651-0_31> <http://
Barsalou, L. W. (1999). Perceptual symbol systems. The Behavioral and Brain Sciences,
www.springerlink.com/index/6r03tfahg6y9wt1r.pdf>.
22, 577–609. http://dx.doi.org/10.1017/S0140525X99252144. Discussion 610–
Neisser, U. (1976). Cognition and reality: Principles and implications of cognitive
660.
psychology. W.H. Freeman and Company.
Barto, A. G., Bradtke, S. J., & Singh, S. P. (1995). Learning to act using real-time
Norris, E. M. (1978). An algorithm for computing the maximal rectangles in a binary
dynamic programming. Artificial Intelligence, 72, 81–138. http://dx.doi.org/
relation. Revue Roumaine de Mathématiques Pures et Appliquées, 23, 243–250.
10.1016/0004-3702(94)00011-O.
A.I. Panov / Biologically Inspired Cognitive Architectures 19 (2017) 21–31 31

Osipov, G. S. (1997). Applied semiotics and intelligent control. In Proceedings of the Redko, V., & Burtsev, M. (2016). Modeling of mechanism of plan formation by new
second workshop on applied semiotics, seventh international conference on artificial caledonian crows. Procedia Computer Science, 88.
intelligence and information-control systems of robots (AIICSR’97) (pp. 27–34). Richter, S., & Westphal, M. (2010). The LAMA planner: Guiding cost-based anytime
Bratislava. planning with landmarks. Journal of Artificial Intelligence Research, 39, 127–177.
Osipov, G. S. (2015). Signs-based vs. symbolic models. In G. Sidorov & S. N. Galicia- http://dx.doi.org/10.1613/jair.2972. Available from arXiv:1401.3839.
Haro (Eds.), Advances in artificial intelligence and soft computing lecture notes in Sardina, S., Silva, L. D., & Padgham, L. (2006). Hierarchical planning in BDI agent
computer science (pp. 3–11). Springer International Publishing. http://dx.doi. programming languages: A formal approach. In Proceedings of the fifth
org/10.1007/978-3-319-27060-9_1 <http://link.springer.com/10.1007/978-3- international joint conference on Autonomous agents and multiagent systems
319-27060-9_1>. (AAMAS’06) (pp. 1001–1008). http://dx.doi.org/10.1145/1160633.1160813.
Osipov, G. S., Panov, A. I., & Chudova, N. V. (2014). Behavior control as a function of Skrynnik, A., Petrov, A., & Panov, A. I. (2016). Hierarchical temporal memory
consciousness. I. World model and goal setting. Journal of Computer and implementation with explicit states extraction. In A. V. Samsonovich, V. V.
Systems Sciences International, 53, 517–529. http://dx.doi.org/10.1134/ Klimov, & G. V. Rybina (Eds.), Biologically Inspired Cognitive Architectures (BICA)
S1064230714040121 <http://link.springer.com/10.1134/S1064230714040121>. for young scientists a dvances in intelligent systems and computing (pp. 219–225).
Osipov, G. S., Panov, A. I., & Chudova, N. V. (2015). Behavior control as a function of SpringerInternational Publishing. http://dx.doi.org/10.1007/978-3-319-32554-
consciousness. II. Synthesis of a behavior plan. Journal of Computer and 5_28 <http://link.springer.com/10.1007/978-3-319-32554-5_28>.
Systems Sciences International, 54, 882–896. http://dx.doi.org/10.1134/ Stanovich, K. E. (2009). Distinguishing the reflective, algorithmic, and autonomous
S106423071505010X <http://link.springer.com/article/10.1134/S106423071505010X>. minds: Is it time for a tri-process theory? In J. Evans & K. Frankish (Eds.),
Panov, A. I., & Yakovlev, K. (2016a). Behavior and path planning for the coalition of In two minds: Dual processes and beyond (pp. 55–88). Oxford University
cognitive robots in smart relocation tasks. In J.-H. Kim, F. Karray, J. Jo, P. Sincak, Press. http://dx.doi.org/10.1093/acprof:oso/9780199230167.003.0003 <http://
& H. Myung (Eds.), Robot intelligence technology and applications 4 advances in keithstanovich.com/Site/Research_on_Reasoning_files/Stanovich_Two_MInds.
intelligent systems and computing (pp. 3–20). Springer International Publishing. pdf>.
http://dx.doi.org/10.1007/978-3-319-31293-4_1 <http://link.springer.com/10. Sternberg, R. J., Forsythe, G. B., Hedlund, J., Horvath, J., Snook, S., Williams, W. M., ...
1007/978-3-319-31293-4_1>. Grigorenko, E. L. (2000). Practical intelligence in everyday life. Cambridge
Panov, A. I., & Yakovlev, K. S. (2016b). Psychologically inspired planning method for University Press.
smart relocation task. Procedia Computer Science, 88, 115–124. http://dx.doi.org/ Sun, R., & Hélie, S. (2012). Psychologically realistic cognitive agents: Taking human
10.1016/j.procs.2016.07.414 <http://linkinghub.elsevier.com/retrieve/pii/ cognition seriously. Journal of Experimental & Theoretical Artificial Intelligence,
S1877050916316702>. 25, 65–92. http://dx.doi.org/10.1080/0952813X.2012.661236.
Pospelov, D. A., & Osipov, G. S. (1997). Knowledge in semiotic models. In Proceedings Verenikina, I., & Gould, E. (1998). Cultural-historical psychology and activity theory.
of the second workshop on applied semiotics, seventh international conference on In H. Hasan, E. Gould, & P. Hyland (Eds.), Information systems and activity theory:
artificial intelligence and information-control systems of robots (AIICSR’97) Tools in context (pp. 7–18). Wollongong University Press.
(pp. 1–12). Bratislava. Vygotsky, L. S. (1986). Thought and language. MIT Press.

Vous aimerez peut-être aussi