Sie sind auf Seite 1von 6

Chaos-Based Meaning Modeling

Marius Crisan
Department of Computer and Software Engineering, Polytechnic University of Timisoara
marius.crisan@cs.upt.ro

Abstract [8][9] may offer promising perspectives in modeling


such nonlinear processes as language processing from
A central problem in language processing is the chaos theory point of view.
meaning understanding. In a previous work we
introduced the concept that meaning is structured 2. Meaning, a holistic concept
information, and we devised a model, called
differentiated-cognition, in order to describe in a A central interest in any theory of language is to
compressed manner the meaning content of sentences. explain the process of knowing the meaning of a
The purpose of this work is to develop this result and sentence, or how words and sentences can have a
explore the possibility of modeling meaning in a semantic representation [10]. There are many
dynamics-based approach by enfolding the approaches for a theory of meaning [11][12] such as
differentiated-cognition description in chaotic meaning as reference, meaning as truth, meaning as
attractors. The results of simulations revealed the usage, meaning as perception, a naturalized account of
potential offered by chaotic attractors to encode meaning, thought and language, etc. We are
meaning of words and sentences in a series of complex particularly interested in developing a computational
enfoldments. This new approach may form the basis of approach for the process of understanding. In a
an integrated language generation-understanding classical view, the information content of what a
model. sentence means can be generated from information
about the meaning of the sentence's constituent parts
and of the ways they are related to each other. In this
1. Introduction concept, natural languages are necessarily
compositional [13]. This means that complex
Attempts to explain language and identify a model expressions or sentences are understood based on the
comprehensive enough to account for language understanding of the constituent parts and their
generation and understanding are rooted in ancient arrangement in the sentence. In a compositional theory
times. In modern approaches, scientists are interested of language, for each sentence in the object-language
in developing computational techniques for both there is a theorem that states the sentence meaning.
speech (or character/word sequences) recognition and This theorem (called T-theorem) is derivable from
synthesis [1][2] in order to enable effective human- axioms that characterize the expressions and operations
machine communication. An important class of forming that sentence and the order in which the
methods of language processing is based on expressions appear. A typical critique of such an
probabilistic and statistical models [3], another one approach, as in [14], has the central argument based on
uses neural networks [4][5] (in particular self- the redundancy of such T-theorems derivations. For
organizing maps [6][7]), but the main challenge for any someone who knows the theory of truth for that
approach is dealing with the nonlinear character of language, in order to understand the sentences of that
language phenomenon. The main progress achieved so language, all knowledge necessary is reducible to
far pertains in principal with a concise articulation of knowing the axioms. Thus, derivation is epistemically
the constraints that language obeys rather than to superfluous since it is conveyed by logical form.
provide a working model of language performance. However, the compositionality constraint has to be
However, the recent evidences of neural sciences and satisfied by any theory of meaning for the simple
the development in the domain of dynamical systems reason that the theory has to show how the meanings of
sentences are determined by properties of the simple the structured information accounting for meaning has
constituents of the sentences, coupled with the the property of being communicable. We named this
combination or order in which the constituents appear information structure undivided meaning whole
[13]. Yet, there are several difficulties here. First, what (UMW). This may be similarly conceived as
constitutes the unit of meaning? Do the individual informational structure of an algorithm that exists
words or even syllables (letters) have a separate internally at the agent’s information level. The
meaning by themselves, or meaning is present only linguistic apprehension is a cognition process that takes
when they are combined together? Second, if we admit place in two phases. First, the separate syllables or
that the complete semantic unit is the sentence then words uttered by the transmitter and/or heard by the
what kind of combination of syllables and words is receiver act as a manifold at the perceptive level.
possible, for instance in the process of pronouncing or Second, this manifold has to trigger a unitary state of
hearing when the uttered words are never together as linguistic cognition or UMW. If we want to follow the
one whole? Obviously, the order of the sentence’s compositional constraint and account for an integrated
constituents (syllables and words) should be meaning at the sentence level we have to postulate the
considered. But, what law would determine the order of existence of an underlying principle of identity.
constituents? At word level, the order of syllables is Without that underlying identity, the sentence’s
very strict. Words containing the same syllables in constituents could not be related and remain only
different order have different meaning or no meaning at separate entities. Such an underlying principle can be
all. Yet, the order seems to be less strict for the words identified in the nature of nonlinear dynamic systems
in a sentence in order to convey meaning. In principle, that manifest a deterministic chaotic behavior.
the same words can be used in a rephrased sentence to The UMW concept and its type of dynamics appear
convey the same meaning. Another problem is to consistent with chaotic attractor modeling. There is a
account for the role an individual word might have in fundamental connection between chaos and
sentence’s meaning. For instance, not all words refer to information. This view is also supported by other
a specific thing, or a given word may be used in large works (such as [16][17]) that demonstrated that a
varieties of contexts and circumstances. Also, there are chaotic system can be manipulated to encode symbolic
causes that may create difficulties such as the representation of a desirable message. Also, in other
similarity/dissimilarity of words’ form (polysemy, approaches [18][19], chaotic attractors are used to code
homonymy, homophony, etc.). From such words and sentences in a process of interaction of
considerations, we may draw the conclusion that a attractors.
formal compositional theory of meaning should be A chaotic system behavior follows complicated
revised in order to encompass the nonlinear phenomena trajectories in state space, none of which dominates
of meaning manifestation from sentence’s constituents. under normal conditions. These trajectories are not
In our approach we start from several premises that random but follow specific patterns being drawn
were discussed in a previous work [15]. Language and toward the chaotic attractor. This attractor
thought are interrelated, i.e., there is no cognition characterizes the chaotic system and is determined by
without the operation of words. Although the individual the dynamic variables and parameters of system
words or even letters may have meaning, the sentence equations. Because of the rich variety of trajectories
is considered the complete form of a meaningful which are kept as a whole by the system’s nature, a
thought. In our view, linguistic communication is based chaotic system is a natural source of structured
on the concept that meaning is ‘structured information’ information, and a suitable space for manifesting the
that must have a finite description and manifests as a underlying principle of identity, of which we spoke
whole at the level of sentences. An utterance (sentence) above. If the syllables composing a word are mapped to
is meaningful if it can generate knowledge (cognition) a chaotic attractor then its dynamic behavior could
in an ideal receiver (hearer or reader). This knowledge account for the unitary conveying of meaning at the
is the result of a reaction mechanism triggered by the word level. At the sentence level, the chaotic attractors
series or words in the sentence. In other words, it’s a corresponding to the constituent words are enfolded in
process of extracting information from that sentence. a resulting chaotic attractor that accommodates the
We have to make distinction between the transmitter’s UMW, as a unitary information structure. This contains
cognition that has originated the sentence and the all the information of the component parts but in a
receiver’s cognition upon the transmitted sentence. higher-order way of integration. In the reversed
Obviously, in an ideal communication process both process, when an agent wants to communicate, it
cognitions have to be similar. The assumption is that begins with the UMW that exists internally in its
knowledge base. The agent identifies a configuration position in DCP tree determines the nature of the
for a chaotic attractor that contains the UMW meaning cognitive structure, because that element is
description. Then the subattractors for each of the cognized by its properties denoted by the elements on
component words are identified. Finally, according to the right side of the tree. The element in the head
the corresponding dynamics for each subattractor, the position can be either a nominal or a verbal word. This
syllables are generated. Therefore, when words are is of interest in the philosophy of language where
uttered in sequence, only apparently they are meaning is analyzed from an object-like (substance-
differentiated or separate entities. In fact, they are the like) or action-like (state-like) original cause (see also
component trajectories of the same chaotic information [20]). For instance, in the above sentence, the head
source. Thus, the model preserves the underlying unity word can be “the sky” which is cognized by the rest of
of meaning, providing in the same time an account for the words, as shown above, or can be the verb “is”. In
linguistic communication through series of distinct the latter case, the action “is” or state of ‘existence’
words. The unitary structure of UMW at the transmitter that has as object “blue-ness” is cognized as ‘the sky’.
is composed of cognitive semantic units that unfold and According to the differentiated-cognition description,
produce the sequence of uttered words that generates in the enfoldment process of the attractors takes place
the receiver’s knowledge base a similar UMW by a starting with the attractor that corresponds with the
correlated enfoldment process. word in the head position. In the fourth section we
The enfoldment of attractors and the formation of present an example of a DCP description and
UMW is a cognitive process that involves the dynamics correspondent enfoldment of chaotic attractors.
of chaotic systems. In order to model this process we
need a formal description of UMW. We suggest to use 3. Chaotic dynamics of meaning
a metalinguistic concept, that we named differentiated-
cognition in [15]. According to this concept, we may In our approach we suggest a model of chaotic
know an object by its capacity to be known. This attractors enfoldment based on differentiated-cognition
follows a classical approach in the philosophy of description of meaning. This model proves to be more
language where cognition is described as knowing an comprehensive and noise-tolerant. Continuing our
object in terms of other objects (see for instance [20]). previous work [21], we exploited the suitable dynamic
If a certain object x is cognized as another object y, property of chaotic attractors to map in a continuous
then we write C(x, y). This forms a semantic unit. x is manner the feature vectors according to some input
called the differenced word and y the differencer. For patterns. The unified word feature vector is obtained by
instance, the meaning of the sentence “The sky is blue” a many-to-one mapping, starting from the component
can be described or cognized in the following terms: letters (phonemes). This bears the unique information
“sky” is cognized as being “blue,” C(s, b). It’s structure of the word meaning. Slight variations in the
interesting to note that the same meaning can be input feature vectors of the component letters
extracted from the simple cluster of words, “the blue (phonemes) are tolerated without major changes in the
sky.” Syntactically, this is not a complete sentence chaotic attractor dynamics. This is an indication of
because it doesn’t contain a definite verb, but robustness in meaning preservation in the case of noise
according to the differentiated-cognition approach, we (or differences in syllables pronunciation or letters
define a sentence as any series of words that satisfies form). Each letter is modeled by a different chaotic
the differenced-differencer relationship, and hence attractor and then embedded in a higher order equation
convey meaning. This type of description can be that describes the cognitive dynamics of that word. In
generalized and applied to more complex the present work, we extend this technique and
constructions. The model can describe the meaning of construct the embedded chaotic attractor for the whole
all declarative sentences because they have similar sentence.
semantic structure. The differentiated-cognition We consider the simpler case where each letter is
universal pattern can be conveniently represented using characterized by a 3-dimensional feature vector, Z =
the differentiated-cognition phrase (DCP) tree. This is [z1, z2, z3]. The letter feature vector is mapped as a
a semantic counterpart of the X-bar schema, with which chaotic attractor selecting the appropriate values for the
it can be suitably linked [15]. The syntactical model of coefficients z1, z2, z3 in the following quadratic iterated
X-bar schema facilitates the construction of the map:
semantic DCP tree by determining the sentence’s
element that will occupy the head position in the DCP xt+1 = z1 + z2xt + z3xt2. (1)
tree. The central element which occupies the head
The quadratic map was chosen as example, for the sake a2, a3], R = [r1, r2, r3], and H = [h1, h2, h3], are mapped
of simplicity, in order to illustrate the basic principle of by the following equations:
the proposed model. It is nonlinearly stable, and can
manifest a variety of chaotic attractors suitable for st+1 = s1 + s2st + s3st2, (6)
mapping the alphabet letters. at+1 = a1 + a2at + a3at2, (7)
Based on letter feature vectors, a unified word rt+1 = r1 + r2rt + r3rt2, (8)
feature vector can be constructed. For a generic word ht+1 = h1 + h2ht + h3ht2, (9)
w, composed by a series of n letters l1l2...ln, the word
feature vector is W = [L1, L2,… Ln]. The quadratic where st, at, rt, and ht are the dynamic variables. The
maps Li, i = 1, n, corresponding to the component resulting attractor, ws, for word Sarah is constructed as
letters form the coefficients of the following
polynomial type equation, ws(t+1) = k1+k2(stwst+atwst2+rtwst3+atwst4+htwst5), (10)

wt+1 = k1 + k2(l1twt + l2twt2 + … + lntwtn), (2) where k1 and k2 are scaling parameters.

where k1 and k2 are scale parameters. Eq. (2) describes


the chaotic behavior at the word level.
Next, following the structure of differentiated-
cognition model, the compound attractor for C(x, y) is
constructed by enfolding the chaotic attractor of the
differencer, wy, in the chaotic attractor corresponding
to the differenced word, wx, in the following form:

wx(t+1) = f(wxt) + g(wyt), (3)

where f and g are the chaotic functions of the two


words constructed from the component letters as in (2).
The enfoldment process is repeated successively for
the entire sentence. A higher order chaotic attractor is Figure 1. DCP tree for (4).
finally obtained that constitutes the meaning content or
UMW at the sentence level. In Fig. 2, the four trajectories (6) – (9) are presented
along with the resulting chaotic attractor (10). The
coefficients are: s1 = 1.2, s2 = 0.01, s3 = –1.5, a1 = –
4. Simulation results 0.9, a2 = –1.6, a3 = 0.6, r1 = 0.8, r2 = –1.1, r3 = –1, h1 =
–1.1, h2 = –0.5, h3 = 1.3, k1 = 0.45, k2 = 0.7.
In order to exemplify the proposed approach, let’s
consider the following simple sentence:

“Sarah drives a red car.” (4)

According to the differentiated-cognition model, the


meaning content of (4) results as:

C(S, C(d, C(c, r))), (5)

Where S, d, c, and r are abbreviations for Sarah,


drives, car and red, respectively. The corresponding
DCP tree is presented in Fig. 1. Sarah is cognized as
the doer or ‘driver’, specified by the verb ‘drives’,
which has the object a car, which in turn is cognized
by the property of being red.
The letters s, a, r, and h in the word Sarah, with the Figure 2. Chaotic attractors for Sarah, (10),
corresponding feature vectors S = [s1, s2, s3], A = [a1, and the corresponding letters s, a, r, and h.
parameters: c1 = 0.9, c2 = –0.8, c3 = –1.1, d1 = 0.5, d2 =
–1.4, d3 = –0.6, e1 = –1, e2 = –1, e3 = 0.7, i1 = –0.4, i2 =
–1.2, i3 = 1.2, v1 = 1.1, v2 = –0.6, v3 = –1. Next, the
attractors for the other words in the sentence, wd for
drives, wc for car, and wr for red, result as follows:

wd(t+1) = k1+k2(dtwdt+rtwdt2+itwdt3+vtwdt4+etwdt5+stwdt5),
(11)
wc(t+1) = k1+k2(ctwct+atwct2+rtwct3), (12)
wr(t+1) = k1+k2(rtwrt+etwrt2+dtwrt3). (13)

In Fig. 3, 4, and 5 the dynamics of these words along


with the component letters are depicted. Comparing the
behavior of the word attractors, we can observe how
Figure 3. Chaotic attractors for drives (11) and the different pattern for each word is given by the
the corresponding letters d, r, i, v, e, and s. combinations of the component letters in the specific
order, following the compositional constraint. The first
letters (syllables) in the word have a greater influence
in the resulting pattern than the latter ones. For
instance, it’s interesting to note the slight resemblance
of the attractors (10) (Fig. 2) and (12) (Fig. 4). Because
the first syllables ‘sar’ and ‘car’ are rather similar,
their influence upon the resulting pattern is major. The
model is noise tolerant and meaning preserving. Slight
variations in the input feature vectors of the component
letters are equally well integrated without major
changes of the word attractor pattern.
Finally, the resulting attractor for the entire sentence
is formed following the differentiated-cognition
description (5), and the process of chaotic enfoldment
described by (3). The resulting attractor is presented in
Figure 4. Chaotic attractors for car (12) and Fig. 6. wrt is enfolded in wct, which is enfolded in wdt,
the corresponding letters c, a, and r. which in turn is enfolded in wst. The combined effect of
the component words meaning creates a pattern that
encapsulates in a unitary way the meaning (UMW) for
the whole sentence.

Figure 5. Chaotic attractors for red (13) and


the corresponding letters r, e, and d.

Similarly, the chaotic attractors for the letters c, d, e, i, Figure 6. Chaotic attractor that captures the
and v, can be constructed using the following meaning of the entire sentence (4).
Of course, the variety of letter attractors is not very [8] T.P. Vogels, K. Rajan, and L.F. Abbott, “Neural
rich in our examples because we used only simple Networks Dynamics”, Annual Review of Neuroscience, July
quadratic maps. Therefore, the resulting word patterns 2005, Vol. 28, pp. 357-376.
are somehow less complicated, but they are suggestive
[9] M.I. Rabinovich, P. Varona, A.I. Selverston, and H.D.I.
enough to illustrate the principles of the proposed Abarbanel, “Dynamical principles in neuroscience”, Reviews
model. In a future work, using, for instance, ordinary of modern physics, vol. 78, no. 4, 2006, pp. 1213-1265.
differential equations, more sophisticated word
attractors can be constructed. [10] H. Helbig, Knowledge Representation and the Semantics
of Natural Language (Cognitive Technologies), Springer
5. Conclusions Verlag Berlin Heidelberg, 2006.

[11] W.G. Lycan, Philosophy of Language: A Contemporary


The paper presented the basic elements of a chaos- Introduction, New York, Routledge, 2000.
based approach to a theory of meaning, starting from
the observed similitude between language apprehension [12] M. Greenberg, and G. Harman, “Conceptual Role
phenomenon and the chaotic behavior of dynamical Semantics”, Oxford Handbook of Philosophy of Language,
systems. The proposed approach was validated by the Ernie Lepore, Barry Smith, eds., Oxford Univ. Press, 2005.
simulation results. In conclusion, we suggest that the
model can be used equally well in both language [13] E. LePore, K. Ludwig, Donald Davidson: Meaning,
understanding and generation, in building Truth, Language, and Reality, Oxford University Press,
2005.
communicating agents and modeling brain dynamics in
language processing. [14] E. Fisher, “On the very idea of a theory of meaning for a
natural language” , Synthese, vol. 111, no. 1, 1997, pp. 1-16.
4. References
[15] M. Crisan, “Meaning as Cognition,” Proceedings of the I
[1] R. Mitkov (editor), The Oxford Handbook Of International Conference on Multidisciplinary Information
Computational Linguistics, Oxford Univ. Press, 2005 Sciences and Technologies-InSciT2006, Merida, Spain, 2006,
pp. 369–373.
[2] D. Jurafsky, J.H. Martin, Speech and Language
Processing: An Introduction to Natural Language [16] E.M. Bollt and M. Dolnik, "Encoding information in
Processing, Computational Linguistics, and Speech chemical chaos by controlling symbolic dynamics", Physical
Recognition, Prentice-Hall, 2000. Review E, Vol. 55, No. 6, June 1997, pp. 6404–6413.

[3] C.D. Manning, H. Schuetze, Foundations of Statistical [17] Y-Cheng Lai, “Encoding Digital Information Using
Natural Language Processing, The MIT Press, Cambridge, Transient Chaos,” International Journal of Bifurcation and
Massachussetts, 2003. Chaos, Vol. 10, No. 4, 2000, pp. 787–795.

[4] H.L. Moisl, “Artificial Neural Networks and Natural [18] H.L. Moisl, “Linguistic Computation with State Space
Language Processing”, in Encyclopedia of Library and Trajectories”, in Emergent Neural Computational
Information Science, 2nd ed. (M.J. Bates, M. Niles Maack, Architectures Based on Neuroscience: Towards
and M. Drake ed.), Taylor and Francis Group, 2003. Neuroscience-Inspired Computing (S. Wermter, J. Austin,
and D. Willshaw, ed.), Springer, 2001, pp. 442-460.
[5] Q. Ma, “Natural Language Processing with Neural
Networks”, in Proceedings of the Language Engineering [19] T. Yang, “Dynamics of vocabulary evolution”,
Conference (LEC'02) 2002, pp. 45-56. International Journal of Computational Cognition
(http://www.YangSky.com/yangijcc.htm), Vol. 1, No. 1,
[6] T. Kohonen, P. Somervuo, "Self-organizing maps of March 2003, pp. 1–19.
symbol strings with application to speech recognition", in
Proc. of Workshop on Self-Organizing Maps (WSOM'97), [20] B.K. Matilal, Logic, Language and Reality, 2nd ed.
pp. 2-7, Espoo, Finland, 1997. Motilal Banarsidass Publ., Delhi, 1990 (reprint: 1997).

[7] T. Honkela, Self-Organizing Maps in Natural Language [21] M. Crisan, “Chaos and Natural Language Processing”,
Processing, PhD Thesis, Helsinki University of Technology, Acta Polytechnica Hungarica, Vol. 4, No.3, 2007, pp. 61–74.
Espoo, Finland, 1997. (http://www.cis.hut.fi/~tho/thesis/)

Das könnte Ihnen auch gefallen