Open access

Convergence towards a Dynamic Theory of Linguistics and Semantics

Written By

Marius Crisan

Published: 01 March 2010

DOI: 10.5772/9644

From the Edited Volume

Convergence and Hybrid Information Technologies

Edited by Marius Crisan

Chapter metrics overview

2,522 Chapter Downloads

View Full Metrics

1. Introduction

From the dawn of civilization mankind was aware of the importance of language. We live, think, have knowledge and our being in language. All knowledge of the world and ourselves is expressed and mediated through language. Therefore, it is not surprising that in ancient views language (Greek logos) encompassed everything as the soul of the universe, and had mystical and religious dimensions and origins. The term logos which literary means “word,” “reason,” or “plan” denotes a deep concept of divine controlling principle manifested by speech that can be found also in Indian, Persian, and Egyptian theological systems. The opening verses from St. John’s Gospel — “In the beginning was the Word, and the Word was with God, and the Word was God.” — became a standard quotation in supporting the ancient philosophical conception of language as the universal foundation of life and thought. The philosophical investigation upon the essence of everything led inevitable, at some stage or other, every system of philosophy to consider language and its relation with thinking, cognition, and reality. The early foundations of the systematic study of language can be traced back to Socrates, Plato, and Aristotle in the Western tradition (Modrak, 2001) and to Yāska, Pānini, and Bhartrhari of the ancient Indian Grammar School (Coward, 1980; Matilal, 1985). The merit of these schools of thought is that they realize that understanding the nature of language means understanding ourselves in relation with the universe we live in.

However, the ancient dimensions faded in time, and it is only recently that modern science has begun seriously to investigate language. A considerable research effort was and is still involved. One direction of investigation is conceptual, and has the aim to answer the fundamental philosophical quests regarding the nature, origins, and usage of language. Modern philosophy of language follows the same type of speculative inquiry into language by pure a priori reasoning established by the ancients. Its main concern is the development of a theory of meaning and the relationship between language and reality (Lycan, 2000; Malmkjaer, 2009; Morris, 2007). Other topics of interest for philosophers of language are language cognition and language acquisition, generation and speech acts. Here, of particular interest is to understand how language is related to the minds of both the speaker and the hearer. Also, a theory is sought to explain how words are translated into other words.

The other direction is mostly empirical in nature (Sampson, 2002), i.e., based on observation and experimentation, and is specific to modern scientific approach. Initially, the study of language, referred by the term philology, was concerned mainly with the historical development of languages and associated literature in the cultural context. Later, the scientific study of language became known as linguistics. Its main theoretical purpose is the construction of a general theory of the structure of language (grammar) and the study of meaning (semantics) (Aronoff & Rees-Miller, 2003; Fromkin, 2000). Linguistics tries to discover the common elements of all languages or the universals and devise a predictive scientific theory of them. The applied linguistics involves the application of the theory to practical tasks such as language teaching and learning, linguistic competence, and communication (Cook, 2003; Davies & Elder, 2004)

The early attempts in linguistics were oriented towards an idealistic conception of language. Later, the dominant view became the classical premise of structuralism that language is a formal system of discrete symbolic units and their combinations. In Saussure’s conception, the language system, called langue, is made up of discrete signifying units or signs which externally manifest in different combinations in language use (the speech of an individual), called parole (Saussure, 2006). Another essential view of structuralism is the arbitrariness of the connection between the signifier (sound or letter group) and the signified (concept). A similar view that all language is conventional and temporal can be traced back to Aristotle and Plato in the Western tradition and to Indian schools such as Nyāya or Buddhism.

After a period dominated mainly by behaviorist attitudes towards language, modern linguistics became influenced by the view that language is a non-finite but denumerable set that can be defined by an algorithm. Natural languages were considered syntactically rule-governed, and the goal of linguistics became the investigation of these rules. Both statistical and algebraic approaches have been considered with much more emphasis on the latter. In Chomsky's approach (Chomsky, 1957), natural language resembles artificial formal languages and therefore is defined by a finite system of generative rules for each language. The most significant accomplishment is the transformational grammar which uses rules to express relationships among the various elements of a sentence and to generate accordingly the grammatical sentences in a language. The purpose of linguistic analysis of a language L is to make distinction between grammatical sequences and ungrammatical sequences of L and to study the structure of grammatical sequences. In this view, the capability of recognizing the meaningfulness and grammaticality of a potential infinity of sentences can be accounted by the existence of a set of rules for assigning meanings to utterances. These rules are supposed to be innate to humans and form a universal grammar shared by all languages.

Computational linguistics is a precursor of artificial intelligence and originated in the task of using computers to translate texts from other languages (Hutchins & Somers, 1992). Later, the task extended to the study of computer systems for natural language understanding and generation (Mitkov, 2005). Specifically, human-computers interaction became the field of natural language processing (NLP) which has the goal of developing techniques for both speech recognition and synthesis (Huang et al., 2001; Jurafsky & Martin, 2000). An important class of methods for language recognition and generation is based on probabilistic models (Jurafsky & Martin, 2000; Manning & Schütze, 1999), such as n-grams model, hidden Markov and Maximum Entropy model. Given a sequence of units (words, letters, morphemes, sentences, etc.) these models try to compute a probability distribution over possible labels and choose the best label sequence. Another approach in NLP is to use neural networks, in particular self-organizing maps of symbol strings (Kohonen, 2001; Somervuo, 2003). Still, an important challenge for any NLP approach, which may hinder its success, is the capability of dealing with the dynamic character of language phenomenon. The main progress achieved so far pertains in principal with a concise articulation of language competence rather than providing a working model of language performance. This is the effect of reducing language to the formal or uttered word, which is viewed only as a symbol, carrying information that a computer can store and retrieve. The higher dimensions of language, experienced for instance when the right words are found to express the nuances of a thought, are missed.

Out of dissatisfaction with formal approaches to language, a relatively new trend has manifested in linguistics. This is cognitive linguistics which starts from the fundamental premise that language reflects patterns of thought (Croft & Cruse, 2004; Evans & Green, 2006; Geeraerts & Cuyckens, 2007). In contrast with the Chomskyan approach, where an autonomous separate module for language acquisition is present in the mind, cognitive linguistics assumes that human linguistic ability is conceptual in nature and not different from other cognitive functions. Knowledge of language, being a cognitive process, emerges from language use. Cognitive semantics, as part of cognitive linguistics, goes beyond the classic theories in semantics which try to explain meaning in terms of necessary and sufficient and truth-conditions. Meaning is conceptual, i.e., corresponds with a concept present in the mind and is based on personal understanding. An increasing influence role in the domain is played by the cognitive grammar (Langacker, 2008; Taylor, 2002). According to this view, the mental grammar takes the form of symbolic assemblies, consisting of conventional pairings of form and meaning, without the need of the abstract rules that cannot be naturally discerned in language use. The competence versus performance issue is dissolved as is also dissolved the principled separation between lexicon and grammar. Grammar is an integral part of cognition as the other cognitive abilities. A central place in this theory is occupied by conceptual archetypes. These are related with the grammatical components and lexical classes and play a cognitive role in experiential gestalts. It's interesting to remark the dynamic dimensions introduced by the archetypes such as physical objects, locations, motion of an object through space, events, participants in an event, energy transfer from one participant to another, etc. (Langacker, 2008). A distinct place in the realm of dynamic view in linguistics and semantics is occupied by the Catastrophe Theory proposed by Thom (Thom, 1983), starting from Tesnière’s fundamental ideas of actant and valency (Tesnière, 1959). There is a correspondence, in this theory, between the evolution of sentence structures, where the verb has the central role, and the morphogenesis of biological forms. Thom identified a set of archetypal morphologies that play a role in the actantial interactions originated by the verb. The morphogenesis of sentence-structures is described in terms of the number of actants involved by the verb and the evolution of interactions in time.

In recent years, an increased scientific awareness is manifested towards the importance of dealing with the dynamics of phenomena for a deeper understanding in any domain of science. Within cognitive linguistics, the dynamical perspective on linguistic phenomena has been argued by several authors to be a promising alternative to the symbolic paradigm based on logics and algebraic algorithms (Andersen, 2002; Manjali, 1995; Peregrin, 2003; Wildgen, 1986; Wildgen, 2008). Any attempt in explaining natural language remains incomplete unless there is an understanding of its dynamics. The dynamical system theory describes the behavior of complex dynamical systems by employing differential and difference equations. Using the tools offered by this theory, the dynamics modeling of linguistic phenomena can be performed more effectively. Another support to the dynamical approach comes from the new established field of cognitive neurodynamics, which provided evidence that event-related brain potentials reflect a lexical-semantic integration and syntactic process that can be interpreted in terms of dynamical system theory (Graben et al., 2008; Rabinovich et al., 2006; Vogels et al., 2005)

In the above context, a reasonable interest to explore the possibility of using dynamical systems in modeling language and meaning appears motivated. Starting from the premise that natural language phenomena can be viewed as a dynamical system the purpose of this chapter is to investigate the possibility of modeling meaning of words and sentences by superposition of chaotic attractors. In the following sections we present some details of this dynamic approach.

Advertisement

2. Meaning and dynamical systems

The essence of language is to be meaningful. The inquiring into the nature of meaning is one of the most profound philosophical quests for human mind. What is meaning? What it means to mean something? How something meaningful for a person can be known or transmitted to someone else? Several approaches have been proposed for a theory of meaning such as meaning as reference, meaning as truth, meaning as usage, thought and language, a naturalized account of meaning, etc. (Collins, 2001; Greenberg & Harman, 2005; Lycan, 2000; Stainton, 1996). It is not our purpose to analyze these theories here, but some common characteristics can be outlined. All theories of meaning encounter the same difficulty: They try to explain meaning using other meaningful concepts, and for this reason are prone to limitations of one kind or another. In general, since all human knowledge is encompassed within language, in order to explain language we need to use language. However, a way out of this difficulty that can lead also to a certain degree of objectivity is to account the object language in metalinguistic terms. For instance, in formal semantics approaches, a metadescription is obtained by assigning labels to sentence constituents (the syntactic category S is replaced by the truth value t) and conflating logical terms with lexical categories (Heim & Kratzer, 1998; Nirenburg & Raskin, 2004). Denotational semantics uses mathematical objects to describe the semantics of the system. Emergent semantics principles, in order to establish semantic agreement, are committed to adoption of meta-data representational models (for instance standards like RDF or OWL) (Bozsak et al., 2002). Therefore, meaning has to be described as being something possessing metalinguistic properties. The higher is the degree of such properties the higher the degree of generality of the respective theory of meaning. In our view, we suggest that high-order metalinguistic properties can be provided by taking into account the dynamic role played by the language constituents in the formation of words and sentences.

On the other hand, we have to consider also the following problem. In the classical view, the information content of what a sentence means can be generated from information about the meaning of the sentence's constituents and of the ways they are related to each other. In this concept, natural languages are necessarily compositional. The compositionality constraint has to be satisfied by any theory of meaning for the simple reason that the theory has to show how the meanings of sentences are determined by properties of the simple constituents of the sentences, coupled with the combination or order in which the constituents appear (LePore & Ludwig, 2005). In this context, the proposed dynamic approach tries to answer several fundamental questions (Crisan, 2008; Crisan, 2009b). One is about the formation of meaning. A word is composed of phonemes which are individually uttered by the speaker and individually perceived by the hearer. If the component phonemes of a word are distinct elements in the process of word uttering and perception, how can these distinct elements be cognized as a whole so that the meaning of the word is understood as a resulting composite phoneme-unit? What constitutes the morpheme or the unit of meaning? Do the individual words or even syllables (letters) have a separate meaning by themselves, or meaning is present only when they are combined together? The individual phonemes do not manifest separate meaning by themselves. Only combined together in a word the meaning is revealed. Similarly, the several words, which are supposed to constitute a sentence, only as combined together can convey the unitary meaning at the sentence level. However, the problem is that these phonemes/words are never together in the same time as a whole. They appear in a sequence, one after another. Words containing the same syllables in different order have different meaning or no meaning at all. Yet, the order seems to be less strict for the words in a sentence in order to convey meaning. In principle, the same words can be used in a rephrased sentence to convey the same meaning. One may argue from a structuralist position that it is the last phoneme/word actually perceived combined with the memory of the previous phonemes or words that brings about the meaning as a whole. In other words, the word is nothing more than the phonemes themselves or the whole results from the sum of its parts. But how is such a combination possible which, obviously, should take also into account the order of phonemes in a word or of words in a sentence. What would determine the order of phonemes/words in the absence of an underlying dynamic condition which engenders the unitary meaning? Another problem is to account for the role an individual word might have in sentence- meaning. For instance, not all words refer to a specific thing, or a given word may be used in large varieties of contexts and circumstances. Also, there are causes that may create difficulties such as the similarity/dissimilarity of words’ form (polysemy, homonymy, homophony, etc.). From such considerations, we may expect that only a holistic dynamic concept could encompass the nonlinear phenomena of meaning manifestation from sentence’s constituents. Verbal communication is made possible because of the presence of similar dynamic linguistic properties in both the speaker and the hearer. It is the role of the dynamic approach to account for such a unifying principle of meaning generation out of the dynamic contribution of the component elements. Therefore, in the dynamic view, we encounter the principle of the gestalt theory that the whole appears greater than the sum of its parts. This is also consistent with the basic principles of cognitive linguistics (Geeraerts & Cuyckens, 2007).

We may start with the assumption that at least one kind of internal states is interrelated with language, or in other words that there is no cognition without the operation of the word. This is not in the sense that we have a thought and then we look for a word with which to express it, or that we have an isolated word which we try to associate with a thought. Our approach assumes that the speaker’s purpose is to convey a thought structure, and therefore uses language to encode that structure, hoping that this code will be understood by the hearer. Understanding is equivalent with the formation of a similar thought in the hearer’s mind. Thus, meaning appears to be inseparably tied to such concepts as belief, judgment, desire, intention, knowledge, and understanding. Therefore, meaning understanding presupposes the capacity of the receiver to extract and retrieve the thought structure of the transmitter from particular utterances.

Observations may lead to the fact that people do not speak in individual words. Linguistic communication is based on a meaning concept as a whole at the level of indivisible sentences. Although the individual words or even letters have meaning, the sentence is the complete form of a meaningful thought. An ideal receiver has to have the “capacity” to extract meaning from a sentence. This capacity is what qualifies the linguistic competency, and can be described by the cognition of the cognitive properties a sentence has assigned by the transmitter. It is useful to consider the following semantic bearing criteria: (1) semantical competency, (2) expectancy, (3) contiguity in space and time, and (4) transmitter’s intention (Matilal, 1985). These cognitive properties are the requirements for defining a grammatical and meaning-bearing sentence. A sentence is said to have semantic competency when the objects denoted by the respective words are compatible one to another. For instance, the sentence "She sees the light." is grammatically acceptable, and has semantic competency, while the sentence "She hears the color." even if it is grammatically acceptable, lacks semantic competency. Semantic expectancy refers to the capacity of an ideal receiver to infer the meaning of an incomplete sentence (utterance). Syntactic expectancy refers to the syntactic property x which has o be assigned to a sentence s when it is not grammatical, in order to make it suitable to convey the meaning. This expectancy is measured by the predictor of the entropy of the entropic source. Contiguity is the property which imposes the absence of any unnecessary spatial (in written text) or temporal (in speech) interval between the words of a sentence. However, there is a difficulty here related with the fact that the same thought can be expressed by the transmitter in different languages and within a language in different paraphrases. On one hand, the thought states of both the transmitter and the receiver are subjective mental states, and on the other hand an objective procedure is required that can provide a ‘representation’ of those cognition states.

In defining meaning as something that must have a finite and objective significance, we postulate the concept of undivided meaning whole (UMW), which exists internally in the mind (the agent’s information level or knowledge base) (Crisan, 2006). This is structured information, and may be similarly conceived as informational structure of an algorithm. A somewhat similar assumption can be found in (Steels & Hanappe, 2006). Even if UMW is a unitary information structure, it is describable rationally in terms of cognitive semantic units. These semantic units are the generating principle of producing the sequence of uttered words. When an agent wants to communicate, it begins with the UMW existing internally in its mind. A sentence (utterance) is significant or meaningful if it can generate knowledge in an ideal receiver (reader or hearer). This knowledge is a result of a reaction mechanism triggered by the series of words in the sentence. When words are uttered producing different sounds in sequence, it appears only to have differentiation. Ultimately, the sound sequence is perceived as a unity or UMW and only then the word meaning, which is also inherently present in the receiver’s mind, is identified.

The above described capacity of the receiver to extract meaning from series of words led to another assumption, that the whole word/sentence meaning has to be inherently present in the mind of each agent according to a similar dynamic process. Thus, it can be explained how it is possible the UMW to be grasped by the hearer even before the whole sentence has been uttered. The sounds which differ from one another because of difference in pronouncement cause the cognition of the one changeless UMW without determining any change in it. Sometimes, reasoning may have to be applied to the components of the sentence so that the cognition is sufficiently clear to make possible the perception of the meaning-whole. It appears that the unitary word-meaning is an object of each agent’s own cognitive perception. When a word, such as “tree” is pronounced or read there is the unitary perception or simultaneous cognition of trunk, branches, leaves, fruits, etc. in the receiver’s mind. Communication (verbal or written) between peoples is only possible because of the existence of the UMW which is potentially perceivable by all and dynamically revealed by words’ sounds or symbols.

The concept of UMW is consistent with a more general view, suggested by Bohm, regarding the possibilities for wholeness in the quantum theory to have an objective significance (Bohm, 1990). This is in contrast with the classical view which must treat a whole as merely a convenient way of thinking about what is considered to be in reality nothing but a collection of independent parts in a mechanical kind of interaction. If wholeness and non-locality is an underlying reality then all the other natural phenomena must, one way or another, be consistent with such a model. Natural language generation and understanding is a phenomenon that might be modeled in such a way. UMW is like “active information” in Bohm’s language, and is the activity of form, rather than of substance. As Bohm puts it clearly (Bohm, 1990), “…when we read a printed page, we do not assimilate the substance of the paper, but only the forms of the letters, and it is these forms which give rise to an information content in the reader which is manifested actively in his or her subsequent activities.” But, similar so called mind-like quality of matter reveals itself strongly at the quantum level. The form of the wave function manifests itself in the movements of the particles. From here, a new possibility of modeling the mind as a dynamical system is considered. In line with Kantian thought, in (Coward, 1980) we find a similar insight, as above, regarding the linguistic apprehension. This is the interplay of two factors of different levels: (a) the empirical manifold of the separate letters or words and (b) the a priori synthesis of the manifold which imparts a unity to those elements which would otherwise have remained a mere manifold. According to this kind of observations it appears motivated to use the concept of manifold for modeling the mind as the seat of language generation and understanding. Manifolds are defined as topological spaces possessing families of local coordinate systems that are related to each other by coordinate transformations pertaining to a specific class. They may be seen also as the multidimensional analogue of a curved surface. This property seems suitable to represent both the natural language constraints and semantic content of linguistic objects. The linguistic apprehension is a cognition process that takes place in two phases. First, the separate syllables or words uttered by the transmitter and/or heard by the receiver act as a manifold at the perceptive level. Second, this manifold has to trigger a unitary state of linguistic cognition or UMW. If we want to follow the compositional constraint and account for an integrated meaning at the sentence level we have to postulate the existence of an underlying principle of identity. Without that underlying identity, the sentence’s constituents could not be related and remain only separate entities. Such an underlying principle can be identified in the nature of nonlinear dynamic systems that manifest a deterministic chaotic behavior.

Usually, a dynamical system is a smooth action of the real numbers or the integers on a manifold. The manifold is the state space or phase space of the system. Having a continuous function, F, the evolution of a variable x can then be given by the equation:

x t + 1 = F ( x t ) E1

The same system can behave either predictably or chaotically, depending on small changes in a single term of the equations that describe the system. Equation (1) can also be viewed as a difference equation (x t+1x t = F(x t) – x t) and generates iterated maps. An important property of dynamical systems is that even very simple systems, described by simple equations, can have chaotic solutions. This doesn’t mean that chaotic processes are random. They follow rules, but even the simple rules can produce amazing complexity. In this regard, another important concept is that of an attractor. An attractor is a region of state space invariant under the dynamics, towards which neighboring states in a given basin of attraction asymptotically approach in the course of dynamic evolution. The basin of attraction defines the set of points in the space of system variables such that initial conditions chosen in this set dynamically evolve to a particular attractor. It is important to note that a dynamical system may have multiple attractors that may coexist, each with its own basin of attraction. This type of behavior is suitable for modeling self-organizing processes, and is thought to be a condition for a realistic representation of natural processes.

In our approach, the dynamic continuity can be found in the domain of dynamical systems and chaos theory. The UMW concept and its type of dynamics appear consistent with chaotic attractor modeling. There is a fundamental connection between chaos and information. This view is also supported by other works that demonstrated that a chaotic system can be manipulated to encode symbolic representation of a desirable message (Bollt & Dolnik, 1997; Lai, 2000). Also, in other approaches (Moisl, 2001; Yang, 2003), chaotic attractors are used for coding words and sentences in a process of dynamic interaction.

Advertisement

3. Chaos-based word modeling

In quantum experiments, when particles interact, it is as if they were all connected by indivisible links into a single whole. The same behavior is manifested by the chaotic solutions in an attractor, as we will see in this section. In spite of the apparent random behavior of these phenomena, there is an ordered pattern given by the form of the quantum wave (or potential) in the former case, and by the equations of the dynamic system in the latter.

Let’s consider the simplest case of the quadratic iterated map described by the equation:

x t + 1 = a 1 + a 2 x t + a 3 x t 2 E2

Even if it is so simple, it is nonlinearly stable and can manifest chaotic solutions. The initial conditions may be drawn to a special type of attractor called a chaotic attractor. This may appear as a complicated geometrical object which gives the form of the dynamic behavior.

In nonlinear dynamics the problem is to predict if a given flow will pass through a given region of state space in finite time. One way to decide if the nonlinear system is stable is to actually simulate the dynamics of the equation. The primary method in the field of nonlinear dynamic systems is simply varying the coefficients of the nonlinear terms in a nonlinear equation and examining the behavior of the solutions. The initial values of the components of the model vector, m i(t), were selected at random in a process of finding a chaotic attractor. Strange attractors are bounded regions of phase space corresponding to positive Lyapunov exponents. We found more than 100 chaotic attractors. In Table 1 we presented a list of several coefficients along with the Lyapunov exponent for which the chaotic attractors were found by random search (Crisan, 2009a). The initial condition x 0 was selected in the range 0.01 – 1 and lies within the basin in many cases. The Lyapunov exponent is computed in an iterated process according to the following equation (Sprott, 2003):

L E = Σ log 2 | a 2 + 2 a 3 x n | / N E3

The sum is taken from a value of n = 1 to a value of n = N, where N is some large number. LE gives the rate of exponential divergence from perturbed initial conditions. If the value is positive (for instance, greater than 0.005) then there is sensitivity to initial conditions and a chaotic attractor can manifest. If the solution is chaotic, the successive iterates get farther apart, and the difference usually increases exponentially. The larger the LE, the greater the rate of exponential divergence, and the wider the corresponding separatrix of the chaotic region may be. If LE is negative, the solutions approach one another. If LE is 0 then the attractors are regular. They act as limit cycles, in which trajectories circle around a limiting trajectory which they asymptotically approach, but never reach.

No. a1 a2 a3 LE
1 1.2 -0.9 -0.9 0.3106
2 1.1 -1 -0.6 6.6073
3 1.1 -0.7 -0.9 0.1538
4 0.8 -1.1 -1 0.2805
5 0.7 -1.2 -0.8 0.2001
6 -0.4 -1.2 1.2 0.3144
7 -0.7 -1.1 1.2 0.3033
8 -0.8 -1.1 0.7 6.9382
9 -0.8 -0.9 1.1 0.2214
10 -1.2 -0.9 0.8 0.2793

Table 1.

The coefficients values and the Lyapunov exponents for ten attractors of (2).

It’s interesting to analyze in more details the behavior of a chaotic attractor. The idea of the self-organizing maps is to project the N-dimensional data into something that can be better understood visually. We follow a similar idea in constructing iterated maps. It is convenient to plot the values in the iterated process versus their fifth previous iterate for a more suggestive aspect. In Fig. 1(a) the iterated map for the attractor No. 10 is presented. A remarkable property of the chaotic solutions, as noted above in connection with quantum physics, is the “ballet-like” behavior as iterations progress. Each new dot on the map, representing the solution x n+1, appears in a random position but orderly following the attractor’s form.

In Fig. 1(b) the same attractor is shown only after a few iterates (2000). We can observe the sparse distribution of dots but along with the ordered path. This type of behavior is similar with the quantum phenomena, such as the distribution of photons along the interference pattern lines in the two-slit interference experiment, where the photons are emitted in series one after the other. This is also akin to the quality of the perception act (understanding word meaning). It’s an observation fact that a word meaning is at first perceived vaguely and then more and more clearly. Thus, through the process of repeated perception or iterations finally the meaning is revealed. Therefore, we may suggest that meaning can be mathematically modeled as a basin of attraction.

Another interesting property is the symmetry between a 1 and a 3 and the corresponding iterated map. Considering again the chaotic attractor a 1 = –1.2, a 2 = –0.9, a 3 = 0.8, a symmetric behavior can be obtain for the values a 1 = 1.2, a 2 = –0.9, a 3 = –0.8.

Figure 1.

a) Quadratic iterated map of (2). (b) Quadratic iterated map of (2) after 2000 iterates. Note the sparse distribution of dots along the regular pattern of the attractor.

There is a large possibility to obtain other attractors by tuning the values of the coefficients. The shape of the attractor changes smoothly with small variations of the coefficients. Even if the interval of variation is rather small, dramatic changes in the shape of the map can be obtained. In Fig. 2(a), the dynamic behavior of (2) can be observed for a 1 = –1.3, a 2 = –0.65, and a 3 = 0.8. If –1.38 ≥ a 1 ≥ –0.94 (a 2 = –0.9 and a 3 = 0.8) the value of LE is negative and fixed point patterns manifest. Trajectories approach a limit cycle for a 1 = –1.3, a 2 = –1, and a 3 = 0.9615. If a 3> 0.9615 the solutions grow unbounded.

Figure 2.

a) The dynamic behavior of (2) for a 1 = –1.3, a 2 = –0.65, and a 3 = 0.8. (b) A different behavior of (2), obtained for a 1 = –1.3, a 2 = –1.06, and a 3 = 0.8.

An important change in shape can be obtained for a 1 = –1.3, a 2 = –1.06, and a 3 = 0.8, with the value of LE = 0.3648, as shown in Fig. 2(b).

The above analysis revealed the fact that chaotic attractors offer dynamic properties that can map in a continuous manner the feature vectors according to some input patterns. In the process of language communication, the dynamics of each phoneme, as it is uttered, has a contribution to the dynamics of the entire word. The goal is to construct a unified word feature that may account for the word meaning or UMW, by encapsulating the phonemes’ dynamics into a unitary description of a chaotic attractor.

For a generic word w, composed by a series of m phonemes p 1 p 2...p m, the word feature vector is W = [P 1, P 2,… P m], where P i, i = 1, m, are the quadratic maps (2) corresponding to each phoneme. In order to encapsulate the phonemes’ dynamics into a resulting attractor at the word level, we may examine two possibilities: (a) to map the phonemes’ attractors as coefficients of a higher-order polynomial type equation, and (b) to linearly superpose the phonemes’ attractors.

In the first approach, the quadratic maps P i, i = 1, m, form the coefficients of the following polynomial type equation,

w t + 1 = k 1 + k 2 ( P 1t w t + P 2t w t 2 +     + P mt w t n ) E4

where k 1 and k 2 are scale parameters. Eq. (4) describes the chaotic behavior at the word level.

Each valid word of length m will determine a corresponding attractor with a unique dynamic behavior. Small variations in the input will be tolerated and recognized with the same meaning, but other illegal combinations will be rejected. For words with higher length, higher-order iterated maps can be used.

A second possibility is to use a linear superposition of P i, i = 1, m, of the following form:

w t + 1   = z 1 P 1t +   z 2 P 2t +   z m 1 P m 1t + P mt E5

where z 1,…, z m-1 are subunitary superposition parameters. These parameters account for the progressive accumulation of the individual phoneme dynamics into the word meaning as the phonemes are uttered in sequence.

In order to exemplify our approach, let’s consider the phonemes /a/, /e/, /d/, and /r/ as they may form the words dear and dare. The corresponding feature vectors A = [a 1, a 2 , a 3], E= [e 1, e 2 , e 3], D = [d 1, d 2 , d 3] and R = [r 1, r 2 , r 3], are mapped by the following equations:

a t + 1 = a 1 + a 2 a t + a 3 a t 2 E6
e t + 1 = e 1 + e 2 e t + e 3 e t 2 E7
d t + 1 = d 1 + d 2 d t + d 3 d t 2 E8
r t + 1 = r 1 + r 2 r t + r 3 r t 2 E9

where a t, e t, d t, and r t are the dynamic variables. The four trajectories (6) – (9) are presented in Fig. 3 for the following feature vectors: A = [-0.9, -1.6, 0.6], E= [-1, -1, 0.7], D = [0.5, -1.4, -0.6] and R = [0.8, -1.1, -1].

The resulting attractor for the word dear is constructed as

w dear ( t + 1 ) = k 1 + k 2 ( d t w t + e t w t 2 + a t w t 3 + r t w t 4 ) E10

and is represented in Fig. 4(a). Similarly, the resulting attractor for the word dare appears in Fig. 4(b), according to

w dare ( t + 1 ) = k 1 + k 2 ( d t w t + a t w t 2 + r t w t 3 + e t w t 4 ) E11

Figure 3.

Chaotic attractors for phonemes /a/, /e/, /d/, and /r/.

Figure 4.

a) Chaotic attractor for dear according to (10). (b) Chaotic attractor for dare according to (11).

There is a clear difference between the dynamics of (10) and (11), although common trajectory patterns can be indentified in both chaotic attractors. This is according to our expectations since both words are composed of the same phonemes. Concomitantly, the meanings encapsulated by the respective words dynamics are clearly different.

In the second approach involving linear superposition, according to (5), the dynamics of the word dear is modeled as

w dear ( t + 1 ) = z 1 d t + z 2 e t + z 3 a t + r t E12

where z 1< z 2< z 3< 1. The resulting chaotic behavior is shown in Fig. 5(a). It is interesting to compare the dynamics of (12) with that of (10), for the same word dear. They are different because the process of phonemes’ encapsulation is linear in (12) and nonlinear in (10). Nonetheless, the dynamic contribution of the word’s phonemes is successfully captured in both cases.

A similar linear superposition can be used for the word dare in the following form:

w dare ( t + 1 ) = z 1 d t + z 2 a t + z 3 r t + e t E13

keeping the same superposition parameters as in (12). The dynamics of (13) appears in Fig. 5(b). When comparing to (12), a clear global difference can be noticed concomitantly with the identification of common trajectories patterns.

The simulation results for both linear and nonlinear superposition of phonemes’ dynamics have proved the validity of the dynamic approach in modeling meaning and semantics. The model can also account the synthetic interplay between the separate linguistic components and the ultimate unitary manifestation of meaning. The key element in this approach is to emphasize the role of individual phonemes in the formation of the composite phoneme-unit at the word level. In this regard, the above linear superposition enfoldment of the phoneme attractors is suggestive, but the enfoldment process can be further refined as we will exemplify below.

Figure 5.

a) Chaotic attractor for dear according to (12). (b) Chaotic attractor for dare according to (13).

The word is not only a linear sum of the phoneme components but a dynamic compound. Therefore, in a series of phonemes, uttered one after another during the word’s generation, the dynamic influence of one phoneme should be manifested in the behavior of the next one in sequence. Then, the influenced of these two phonemes combined is manifested on the next one, and so on. We exemplify this process in the following set of simulations (Crisan, 2009b). We may start from the general form of a two-dimensional non-linear quadratic system:

X new ( x y )   =  A 00 +  A 0 1 y +  A 0 2 y 2 + A 1 0 x +  A 11 x y + A 12 x y 2 +  A 2 0 x 2 +  A 21 x 2 y +  A 22 x 2 y 2 Y new ( x y )   =  B 00 +  B 0 1 y +  B 0 2 y 2 +  B 1 0 x +  B 11 x y + B 12 x y 2 +  B 2 0 x 2 +  B 21 x 2 y +  B 22 x 2 y 2 E14

If the coefficients were chosen from the approximate interval [–1 +1] the system would exhibit behavior that was stable or bounded, non-degenerative, non-periodic and deterministically chaotic. This can be a rich source of chaotic attractors suitable for modeling the syllable components of words.

Let’s consider a series of three phonemes, say /o/, /r/, and /t/. They can be modeled by equations of type (14) with the values of coefficients as given in Table 2. These values are chosen so that (14) manifests a typical deterministic chaotic behavior. The initial values (x 0, y 0) are selected in the interval (0.001 – 0.5). The dynamics of phoneme /o/ can be observed in Fig. 6(a), of phoneme /r/ in Fig. 6(b), and of phoneme /t/ in Fig. 6(c). There is a rich variety of trajectories in the chaotic behavior which is suitable for the simulation process.

/o/ /r/ /t/
A00 −0.375 −0.164 0.723
A01 −0.033 0.179 0.883
A02 0.065 0.895 0.178
A10 0.519 −0.377 −0.907
A11 0.533 0.442 −0.419
A12 −0.51 0.106 −0.448
A20 0.255 −0.625 −0.044
A21 −0.822 0.914 −0.08
A22 0.376 −0.117 0.124
B00 0.011 −0.663 0.34
B01 0.032 0.525 −0.169
B02 −0.683 0.43 −0.931
B10 −0.952 −0.075 −0.145
B11 0.229 0.942 −0.876
B12 0.182 0.011 −0.941
B20 −0.046 0.56 0.152
B21 −0.624 −0.728 −0.198
B22 0.032 −0.81 −0.812

Table 2.

The Values of Coefficients in (14) for Modeling the Phonemes /o/, /r/, and /t/.

Let’s consider next the generation process of the syllable or word or. The phoneme /o/ is followed immediately in time by the phoneme /r/. This means that the attractor of phoneme /o/ becomes enfolded in the attractor of phoneme /r/ from the very beginning of the word generation. Our premise is that, in order to communicate meaning, an underlying dynamic principle of unifying phonemes in the formation of words has to exist. This is of equally importance for both word generation and recognition.

Figure 6.

a) The dynamics of the phoneme /o/. (b) The dynamics of the phoneme /r/. (c) The dynamics of the phoneme /t/.

Therefore, resuming the above example, the phoneme /o/ is generated according to the equation

o X new ( x y )   =  A 00 +  A 0 1 y +  A 0 2 y 2 + A 1 0 x +  A 11 x y + A 12 x y 2 +  A 2 0 x 2 +  A 21 x 2 y +  A 22 x 2 y 2 o Y new ( x y )   =  B 00 +  B 0 1 y +  B 0 2 y 2 +  B 1 0 x +  B 11 x y + B 12 x y 2 +  B 2 0 x 2 +  B 21 x 2 y +  B 22 x 2 y 2 E15

The phoneme /r/ is generated in a similar iterated process as (15) for the pair [rX new (x, y), rY new (x, y)]. Next, in order to account the influence of /o/, the values [rX new (x, y), rY new (x, y)] are recomputed, every iteration, according to

r X new ( x y )   = w 1 [ r X new ( x y )   + w 2 o X new ( x y ) ] r Y new ( x y )   = w 1 [ r Y new ( x y )   + w 2 o Y new ( x y ) ] E16

where w 1 and w 2 are weights. By varying the values of w 1 and w 2 suggestive results of the dynamic influence of /o/ upon /r/ are obtained. For instance, in Fig. 7(a) the dynamics of the word or can be observed for w 1 = 0.93 and w 2 = 0.23.

Next, it’s interesting to study comparatively the dynamics of the syllable ro. The phoneme-unit /ro/ is generated by a similar equation as (16) of the following form,

o X new ( x y )   = w 1 [ o X new ( x y )   + w 2 r X new ( x y ) ] o Y new ( x y )   = w 1 [ o Y new ( x y )   + w 2 r Y new ( x y ) ] E17

The correspondent dynamics can be seen in Fig. 7(b) for w 1 = 0.9 and w 2 = 0.25.

Figure 7.

a) The dynamics of the word or according to (16). (b) The dynamics of the syllable ro according to (17).

The last phoneme in a generated word is dominant because the meaning is revealed only after the last phoneme is uttered. We can observe this effect by considering the influence of the phoneme /t/ upon the previous syllable or as in the word ort. The following equation models the process,

t X new ( x y )   = w 1 [ t X new ( x y )   + w 2 r X new ( x y ) ] t Y new ( x y )   = w 1 [ t Y new ( x y )   + w 2 r Y new ( x y ) ] E18

where [rX new (x, y), rY new (x, y)] are iterated according to (16). The values of weights are w 1 = 0.9 and w 2 = 0.25. The dynamic behavior of (18) is presented in Fig. 8(a).

The combined influence of the three attractors can be clearly observed in an interesting pattern. Also, following a similar process, it’s interesting to observe the formation of the phoneme-unit to in Fig. 8(b), and the dynamics of the word tor in Fig. 8(c). Completely different dynamics are obtained in both cases although the initial phonemes are identical. The meaning is determined by the combined effect of all the phonemes’ attractors in their order of appearance, in a definite time. Although the phonemes may look like separate entities, as a result of the underlying enfoldment process of their dynamic behavior, the meaning is conveyed as a whole.

The described nonlinear model of attractors’ enfoldment is stable and preserves rather well the chaotic behavior of the components. The enfoldment process of the chaotic trajectories of one phoneme into another is clearly demonstrated. This proves to be more refined in modeling the phoneme-unit than the linear superposition as used in (5). However, both methods can provide the resultant chaotic attractor with a clearly distinct pattern for the entire word.

Figure 8.

a) The dynamics of the word ort according to (18). (b) The dynamics of the syllable to. (c) The dynamics of the word tor.

Advertisement

4. Sentence as the semantic unit of language

We start considering along with other theories of language that the sentence (utterance) is the semantic unit of language. Although the individual words may have meaning this is not complete. Only at the sentence level a complete unitary meaning is revealed. Communication is done with sentences. The individual words have only a conducive role in the formation of the sentence meaning as a unity. The dynamics of the component words of a sentence can be modeled successfully by chaotic attractors. The dynamics of the entire sentence results according to an informational structure which defines the coupling process of the words’ attractors and the contribution of each word’s dynamics in the ensemble.

The role of the dynamics of each component word can be analyzed considering its relations and position in a sentence. We consider in a general formalization that a word is any phoneme-sequence possessing the property of inflection. Normally, each word takes either a verbal, i.e., conjugational inflection, in which case it is called a verb, or a nominal, i.e., declensional inflection, in which case it is of a non-verbal category (substantives, adjectives, participles, etc.). All the other words which do not have declensional inflections, such as prepositions, may be considered to possess invariant inflection. However, only classifying words in terms of their inflection property is incomplete and does not seem to help much in explaining how the meaning as structured information is conveyed by a sentence. We have to take into account the role of the specific dynamics of each component word in a sentence in relation with other words. Therefore, we suggest the employment of dynamic criteria in defining the notion of a word. The semantic criterion determines the minimum sequence length of the phonemes which convey a meaning. Thus, words may vary in complexity, from the shortest meaning-bearing ones to the more complex compound words. Based on meaningful words, we may define, in general terms, a sentence as being a cluster of words capable to generate a cognitive meaning in a competent receiver (reader or hearer). This concept is further articulated by specifying the dynamic influence and interdependence of words. Here we emphasize the importance of the verb’s function in each sentence. Containing a verb is a necessary condition for being a sentence. The verb’s role as the organizing centre that distributes the actantial places was discussed in the previous works of Tesnière and Thom (Tesnière, 1959; Thom, 1983). In the present approach we extend this concept by emphasizing the role of the verb’s dynamics in forming the sentence-meaning (Crisan, 2009c).

At the sentence level, a similar process of attractor enfoldment as in the case of words can account the formation of UMW. In order to model such a process we need a metalinguistic description. One possibility is to apply the attractor enfoldment process in conjunction with the differentiated cognition model [Crisan, 2006; Matilal, 1985]. According to this concept, we describe cognition as knowing something as something else. In other words, we may know an object by its property to be known. If a certain object x is cognized by another object y, then we write C(x, y). Differentiated-cognition description can be generalized and applied to more complex constructions. For instance, from the sentence

A d a m   r e c i t e s   ( a )   p o e m E19

the meaning can be described in the following terms of cognized objects and properties: Adam is cognized by the activity of reciting which has a poem as object. Thus, cognition appears as a series of descriptions of one object in terms of others. If we use the notation C(x, y) as it was introduced above, the structural cognitive description of the meaning content of (19) is obtained as:

C ( a   C ( r   p ) ) E20

where a stands for Adam, r for recites, and p for poem respectively. It’s interesting to remark the possibility to apply directly the differentiated-cognition model to the X-bar model. We consider again the sentence (19), but using this time a VP representation, since for the sake of simplicity we disregard the verb’s inflection information. The head of VP tree is the node V, denoting the verbal element recites. This is also the head of the corresponding DCP tree. The specifier of VP is the NP Adam, which is the qualifier of the VP node. The V node is cognized with the NP node (complement). The corresponding DCP tree is shown in Fig. 9. The compatibility with the X-bar schema is an advantage that makes the DCP description easily to integrate in the classical syntactic parsers. The semantic description results as follows:

C ( C ( V  NP )  C ( NP a ) ) E21

where a stands for ‘agent’ which is the qualifying property of the specifier NP.

Figure 9.

The corresponding DCP tree of the VP tree of sentence (19) having the verb as the central element.

The verb is modeled by a chaotic attractor that provides a suitable dynamics for making a natural connection to the other words in the sentence. The verb in the head position attracts the other words to it. The two C’ nodes in Fig. 9 represent the slots of attraction in the manifold. One slot attracts the nominal element (Adam) as agent from one dimension and the other slot attracts the verb’s object (poem) from another dimension. This approach is in a way consistent with the idea discussed in (Andersen, 2002) of using potential fields associated with each word that influences other words, and subscribes also to the more general view of quantum potential which is regarded as information whose activity is to guide the quantum behavior of particles, as discussed above (Bohm, 1990). We may also note the similitude with the actantial interaction of Thom (Thom, 1959). An interesting proposal for a unifying theory of actants was advanced by Mel’čuk (Mel’čuk, 2004). The Mel’čuk’s approach to actants is based on dependency rather than constituency. This means that the structure in which actants appear is determined by dependency relations between terminal elements. Similar dependency structures may result by coupling the attractors according to the DCP description. This is active information that structures the entire sentence as a unity and has to be resident at the receiver’s cognitive level.

If the verb’s dynamics obeys a potential field behavior as mentioned above the best description is through differential equations. We consider the following differential equations in general form:

d x / d t   =   a 00 + a 0 1 y + a 0 2 y 2 + a 1 0 x + a 11 x y + a 2 0 x 2 d y / d t   =   b 00 + b 0 1 y + b 0 2 y 2 + b 1 0 x + b 11 x y + b 2 0 x 2 E22

This can provide a very rich variety of dynamics according to the values of the coefficients and may suitably model different verb constructs. The feature vector of the verb’s component phonemes can be mapped to the twelve coefficients of (22). A remark should be made at this stage. Naturally, in order to model the dynamics of verbs, a similar process of word construct out of component phonemes, as in the previous section, should be considered. However, for the sake of simplicity, in our example, the verb recites is modeled directly by (22) according to the following values of coefficients: a 00 = 0.2, a 01 = –1, a 02 = 0, a 10 = –0.5, a 11= 0, a 20 = 0, b 00 = 0, b 01 = 1, b 02 = –1, b 10 = 1, b 11= 0, b 20 = 0. The resulting attractor is presented in Fig. 10(a). This attractor is a flow and can act as a connecting principle in forming the sentence meaning by coupling the dynamics of the nominal element and the verb’s object.

Next, we suggest that the dynamics of the nominal element can be modeled by a sine-map. This type of map can naturally adapt to the type of dynamics implied by the substance-like nominal element. In the present approach we investigate the following form of sine-map (Okuda & Tsuda, 1994):

u n + 1 =  sin ( a u n + b v n ) v n + 1 =  sin ( c u n + d v n ) E23

where the parameters a, b, c, and d can be chosen according to the feature vector of the nominal element. In our case, for the word Adam we selected the values: a = 1, b = –1.5, c = 1, and d = 0.7. Roughly, these values may correspond to the four component phonemes of the word. The dynamics of (23) is depicted in Fig. 10(b).

Figure 10.

a) Dynamics of the verb recites according to (22). (b) Dynamics of the nominal element Adam according to (23).

According to the DCP description the verb is the central dynamic element for the entire sentence which couples both the nominal element and verb’s object. When the first word in the sentence is uttered this is attracted according to its type of dynamics to the corresponding slot in relation with the verb. Even if the verb is missing, the receiver tends to supply it in order to complete the sentence meaning. In other words, the whole ideas are verbalized. For instance, after the substantive noun Adam is uttered, the receiver expects a type of dynamics specific to a verb. This can be modeled by a coupled chaotic system which involves (22) and (23):

u n + 1 =  sin ( a u n + b v n )   + k 1 x n v n + 1 =  sin ( c u n + d v n )   + k 1 y n E24

where k 1 is a coupling constant, and x n, y n are the values at the nth iteration of (22) in the discretization process. In Fig. 11(a), the dynamic behavior of the coupled words Adam recites is shown for k 1 = 2.5. We can observe a very interesting itinerant motion starting from a cluster like a distorted sine-map (23) to a path influenced by the dynamics of (22). The cluster region is formed dynamically in time and finally emerges into the trajectory of type (22). We can also observe in the cluster several instances of copying the dynamic pattern of (22). The two dynamics naturally fit one to another in a suggestive behavior.

The verb’s object can be modeled by another type of dynamics such as a two-dimensional quadratic map of the following form:

s n + 1 = c 00 +   c 0 1 t n + c 1 0 s n + c 2 0 s n 2 t n + 1 = d 00 + d 0 1 t n + d 1 0 s n + d 0 2 t n 2 E25

where the eight coefficients can be adjusted to map the word’s feature vector. In our example, for the simulation of the verb’s object poem the values are: c 00 = 0.8, c 01 = 0.1, c 10 = 0.2, c 20 = –0.4, d 00 = 0.7, d 01 = 0.9, d 10 = 1, d 02 = –1.2. The attractor is shown in Fig. 11(b).

Figure 11.

a) Dynamics of the coupled words Adam recites according to (24). (b) Dynamics of the verb’s object poem according to (25).

In the DCP description, the verb dynamics attracts naturally an object. For, instance, when a verb is uttered first, the receiver expects naturally a verb object to follow. This process can be modeled by coupling the systems (22) and (25) as follows:

d x / d t   =   a 00 + a 0 1 y + a 0 2 y 2 + a 1 0 x + a 11 x y + a 2 0 x 2 + k 2 s n d y / d t   =   b 00 + b 0 1 y + b 0 2 y 2 + b 1 0 x + b 11 x y + b 2 0 x 2 +   k 2 t n E26

where k 2 is a coupling constant. In Fig. 12(a) the dynamics of the coupled words recites poem is shown for k 2 = 0.02. We can see that the dynamic pattern of the differential equations (22) is preserved and qualified by the chaotic dots due to (25).

Figure 12.

a) Dynamics of the coupled words recites poem according to (26). (b) Dynamic behavior of (19) by coupling (24) and (26).

Finally, according to the DCP semantic description (21) we can couple (24) with (26) and obtain the dynamics for the entire sentence (19). The dynamic behavior can be seen in Fig. 12(b) for k 1 = 0.18 and k 2 = 0.33. We may observe an itinerant motion in the manifold starting from a cluster that finally bursts into three paths that manifest a repeated sequence of patterns similar with (25). This is according to our expectations. The verb dynamics has two slots that naturally have to be filled by the other two actants in the sentence. The final trajectories manifested during the formation of word Adam are linked to the verb dynamics. This is responsible for creating an itinerant motion towards the final pattern of the word poem. Other interesting patterns are obtained for different combinations of k 1 and k 2, which emphasize either the dynamics of the nominal element or the dynamics of the verb’s object. In most cases the itinerant motion between the subject and object is visible. The results support our premise that the verb dynamics is essential in structuring the sentence meaning as a whole.

Advertisement

5. Conclusion

Our purpose was to study the possibility of using dynamical systems in modeling natural language. Of particular interest in any model of meaning, and semantics in general, is to account the interplay process of the empirical manifold of individual phonemes or separate words and the unitary characteristic of meaning as a whole. We started from the premise of UMW and the observation facts of language apprehension and noted a similitude with the chaotic behavior of dynamical systems. The attractor behavior as studied for a series of iterated map seems to be robust enough to accept feature vectors for phonemes that may compose any word of length m in the dictionary. The separate dynamics of the phonemes participate in the manifestation of a unique dynamic behavior of the entire word that may represent the UMW. Two approaches have been proposed for modeling the formation of word’s dynamics based on individual phonemes, and the simulation results proved successfully in both cases. At the sentence level, the sentence-meaning is conveyed as a whole and not as the summation of the individual words. This is the result of a cognitive process of assigning cognitive properties by the speaker in order to form a grammatical and meaning-bearing sentence. A competent hearer can extract these cognitive properties of a sentence in a similar cognitive process. This process takes place on the level of sentence-meaning and not on the level of the word-meanings. Only as one hears all of the words of a sentence and grasps the whole sentence by the mind the meaning results in a complete form. The dynamics of this cognitive process is modeled by the enfoldment of chaotic attractors in a similar way as in the case of word’s phonemes. The process of meaning communication begins with the UMW in the form of a thought that exists internally in the mind. This UMW determines the configuration for a chaotic attractor. Then, the subattractors for each of the component words are identified and according to the corresponding dynamics for each subattractor the syllables are generated. The same UMW is the generating principle of producing the sequence of words having the sentence-meaning in the speaker’s mind, and is also the result of a process of extracting meaning from that sentence in the hearer’s mind. Thus, the model preserves the underlying unity of meaning, providing in the same time an account for linguistic communication through series of distinct words. The chaotic attractors corresponding to the sentence’s constituent words are enfolded in a resulting chaotic attractor that accommodates the UMW as a unitary information structure. This structure follows the differentiated cognition model based on dependency relations between actants. The sentence-meaning appears as series of cognitions of one object in terms of others, and contains all the information of the component parts but in a higher-order way of integration and completeness. The verb’s dynamics plays the central role in the formation of the complete sentence-meaning by creating two slots of attraction for coupling with the dynamics of the subject and the object. This describes naturally the basic structure of the complete UMW of a sentence or thought. In more complex sentences, the subject and the object can be further qualified individually by other properties such as substance-like and spatial-temporal information by coupling their dynamics with the corresponding dynamics of the properties. Even if only one word is uttered, its dynamics determines the corresponding slot of attraction in relation with the verb, and the hearer undergoes a cognitive process involving added word-meanings in order to complete the sentence-meaning. In the case where the single word utterance has the dynamics of a subject, some verbal meaning is added mentally so that the completeness of meaning is achieved. If only the verb is uttered then the complete meaning is mentally perceived by adding a possible subject and even an object according to the context or the cognitive properties suggested by the speaker. In conclusion, in contrast with the classical view which considers language as a collection of independent parts in a mechanical kind of interaction, the logic of the proposed dynamic approach is based on the idea that the meaning-whole is prior to the parts. The simulations results validate this view and encourage us to deepen the research in this direction.

References

  1. 1. Andersen P. B. 2002 Dynamic semiotics. Semiotica, 139-1/4, 2002, 161 210 , 0037-1998
  2. 2. Aronoff M. Rees-Miller J. (Eds.) 2003 The handbook of linguistics, Blackwell Publishing, 1-40510-252-7Malden
  3. 3. Bohm D. 1990 A new theory of the relationship of mind and matter. Philosophical Psychology, 3 2 1990, 271 286 , 0951-5089
  4. 4. Bollt E. M. Dolnik M. 1997 Encoding information in chemical chaos by controlling symbolic dynamics, Physical Review E, 55 6 June 1997, 6404 6413 , 1539-3755
  5. 5. Bozsak E. et al. 2002 Kaon- towards a large scale semantic web, Proceedings of EC-Web 2002, LNCS, Springer, 304 313 .
  6. 6. Chomsky N. 1957 Syntactic Structures, Mouton & Co., Publishers, 3-11017-279-8 Hague
  7. 7. Collins J. 2001 Truth Conditions Without Interpretation, Sorites, 13 52 71 , 1135-1349
  8. 8. Cook G. 2003 Applied Linguistics, Oxford University Press, 978-0-19437-598-6 Oxford
  9. 9. Coward H. G. 1980 The Sphota Theory of Language: A Philosophical Analysis, Motilal Banarsidass, 8-12080-181-4
  10. 10. Crisan M. 2006a Meaning as Cognition, Proceedings of the I International Conference on Multidisciplinary Information Sciences and Technologies, InSciT2006, 369 373 , 108461131053 Spain, Oct. 2006, Open Institute of Knowledge, Badajoz, Spain
  11. 11. Crisan M. 2006b Information Machine and the Gödelian Case. Scientific Bulletin of “Politehnica” University of Timisoara, Transactions on Automatic Control and Computer Science, 51 (65), 4 2006, 45 50 , 0122-4600X
  12. 12. Crisan M. 2008 Chaos-Based Meaning Modeling, Proceedings of the 4th International Conference on Networked Computing and Advanced Information Management, (NCM 2008), 2 314 319 , 978-0-76953-322-3 Gyeongju, Korea, Sep. 2008, IEEE CS, CPS, Los Alamitos, California
  13. 13. Crisan M. 2009a Dynamic Modeling of Natural Language. Scientific Bulletin of “Politehnica” University of Timisoara, Transactions on Automatic Control and Computer Science, 54 68), 1 2009, 39 44 , 0122-4600X
  14. 14. Crisan M. 2009b Upon Dynamic Natural Language Processing, Proceedings of the 2009 International Conference on New Trends in Information and Service Science, (NISS 2009), 487 492 , 9780769536873 Beijing, China, July 2009, IEEE CS, CPS, Los Alamitos, CA
  15. 15. Crisan M. 2009c Upon the Dynamic Modeling of Sentence Meaning, Proceedings of the 5th International Conference on Networked Computing and Advanced Information Management, (NCM 2009), Seoul, Korea, August, 2009, IEEE CS (in press)
  16. 16. Croft W. Cruse D. A. 2004 Cognitive Linguistics (Cambridge Textbooks in Linguistics), Cambridge University Press, 0-52166-770-4 Cambridge/New York
  17. 17. Davies A. Elder C. (Eds.) 2004 Handbook of Applied Linguistics, Blackwell Publishing, 0-63122-899-3 Oxford/Malden
  18. 18. Evans V. Green M. 2006 Cognitive Linguistics. An Introduction, Edinburgh University Press, 0-74861-832-5, Edinburgh
  19. 19. Fromkin V. A. et al. 2000 Linguistics: An Introduction to Linguistic Theory, Blackwell, 0-63119-711-7 Oxford/Malden
  20. 20. Geeraerts D. Cuyckens H. (Eds.) 2007 The Oxford Handbook of Cognitive Linguistics, Oxford University Press, 10019514378 Oxford/New York
  21. 21. Graben P. beim. Gerth S. Vasishth S. 2008 Towards dynamical system models of language-related brain potentials. Cognitive Neurodynamics, 2 3 Sept. 2008, 229 255 , 1871-4080
  22. 22. Greenberg M. Harman G. 2005 Conceptual Role Semantics. Oxford Handbook of Philosophy of Language, Ernie Lepore, Barry Smith, eds., Oxford University Press.
  23. 23. Heim I. Kratzer A. 1998 Semantics in Generative Grammar, Blackwell, 0-63119-713-3, Oxford
  24. 24. Huang X. Acero A. Hon H. W. 2001 Spoken Language Processing: A Guide to Theory, Algorithm and System Development, Prentice-Hall, 0-13022-616-5 Upper Saddle River, NJ
  25. 25. Hutchins W. J. Somers H. L. 1992 An introduction to machine translation, Academic Press, ISBN 012362830X, London
  26. 26. Jurafsky D. Martin J. H. 2000 Speech and Language Processing: An Introduction to Natural Language Processing, Computational Linguistics, and Speech Recognition, Prentice-Hall, 0-13095-069-6 Upper Saddle River, NJ
  27. 27. Kohonen T. 2001 Self-Organizing Maps, Springer-Verlag (3rd extended ed.), 3-54067-921-9, Berlin
  28. 28. Lai Y-C. 2000 Encoding Digital Information Using Transient Chaos, International Journal of Bifurcation and Chaos, 10 4 2000, 787 795 , 0218-1274
  29. 29. Langacker R. 2008 Cognitive Grammar: A Basic Introduction, Oxford University Press, 0-19533-196-7, New York
  30. 30. Lycan W. G. 2000 Philosophy of Language: A contemporary introduction, Routledge, 0-41517-116-4, London
  31. 31. Malmkjaer K. 2009 The Linguistics Encyclopedia (3rd edition), Routledge, 100415424321, London
  32. 32. Manjali F. 1995 Dynamic Semiotics or the Case for Actantial Case. Sémiotiques, 6-7 , Décembre 1995, 85 97, 1160-9907
  33. 33. Manning C. D. Schütze H. 1999 Foundations of Statistical Natural Language Processing, MIT Press, 0-26213-360-1
  34. 34. Matilal B. K. 1985 Logic, Language and Reality, Motilal Banarsidass, 8-12080-717-0 New Delhi
  35. 35. Mel’čuk I. 2004 Actants in semantics and sintax. I: actants in semantics. Linguistics, 42-1, 2004, 1 66 , 0024-3949
  36. 36. Mitkov R. (ed.) 2005 The Oxford Handbook of Computational Linguistics, Oxford University Press, 019927634X Oxford/New York
  37. 37. Modrak D. K. W. 2001 Aristotle’s Theory of Language and Meaning, Cambridge University Press, 0-52177-266-4 Cambridge/New York
  38. 38. Moisl H. L. 2001 Linguistic Computation with State Space Trajectories, In: Emergent Neural Computational Architectures Based on Neuroscience: Towards Neuroscience-Inspired Computing (S. Wermter, J. Austin & D. Willshaw, ed.), 442 460 , Springer, 3-540-42363-X London
  39. 39. Morris M. 2007 An introduction to the philosophy of language, Cambridge University Press, 978-0-521-84215-0, Cambridge/New York
  40. 40. Nirenburg S. Raskin V. 2004 Ontological Semantics. The MIT Press, 0-262-14086-1
  41. 41. Okuda H. Tsuda I. 1994 A coupled chaotic system with different time scales: Possible implications of observations by dynamical systems. International Journal of Bifurcation and Chaos, 4 4 1994, 1011 1022, 0218-1274
  42. 42. Peregrin J. (ed.) 2003 Meaning: The Dynamic Turn. Current Research in the Semantics/Pragmatics Interface, Elsevier, 0-08044-187-4, London
  43. 43. LePore E. Ludwig K. 2005 Donald Davidson: Meaning, Truth, Language, and Reality, Oxford University Press, 100199251347 Oxford/New York.
  44. 44. Rabinovich M. et al. 2006 Dynamical principles in neuroscience. Reviews of modern physics, 78 4 2006, 1213 1265 , 0034-6861
  45. 45. Sampson G. 2002 Empirical linguistics, Continuum International, 0-82644-883-6 London/New York
  46. 46. Saussure F. de. 2006 Writings in General Linguistics, Oxford University Press, ISBN 019926144X, Oxford
  47. 47. Somervuo P. 2003 Speech Dimensionality Analysis on Hypercubical Self-Organizing Maps. Neural Processing Letters, 17-2 , April, 2003, 125 136 , 1370-4621
  48. 48. Sprott J. C. 2003 Chaos and Time-Series Analysis, Oxford University Press, 0-19850-840-9 Oxford/New York
  49. 49. Stainton R. J. 1996 Philosophical perspectives on language. Peterborough, Ont., Broadview Press.
  50. 50. Steels L. Hanappe P. 2006 Interoperability through Emergent Semantics. A Semiotic Dynamics Approach. In: Journal on Data Semantics VI, 143 167 , Springer, 978-3-54036-712-3 Berlin/Heidelberg
  51. 51. Taylor J. R. 2002 Cognitive Grammar, Oxford University Press, 0-19870-033-4 Oxford/New York
  52. 52. Tesnière L. 1959 Éléments de syntaxe structurale, Klincksieck, 2-25201-861-5 Paris
  53. 53. Thom R. 1983 Mathematical models of morphogenesis, Ellis Horwood, Halsted Press, 100470274999 Chichester, New York
  54. 54. Vogels T. P. Rajan K. Abbott L. F. 2005 Neural Networks Dynamics. Annual Review of Neuroscience, 28 July 2005, 357 376 , 0014-7006X
  55. 55. Wildgen W. 1986 Procesual Semantics of the Verb. Journal of Semantics, 5 4 1986, 321 344 , 0167-5133
  56. 56. Wildgen W. 2008 The “dynamic turn” in cognitive linguistics. Studies in Variation, Contacts and Change in English, Vol. 3-Approaches to Language and Cognition, 2008, 1797-4453
  57. 57. Yang T. 2003 Dynamics of vocabulary evolution. International Journal of Computational Cognition, 1 1 March 2003, 1 19 , 1542-8060

Written By

Marius Crisan

Published: 01 March 2010