A Chaos Auto-Associative Model with Chebyshev Activation Function

*Masahiro Nakagawa*

## **Abstract**

In this work, we shall put forward a novel chaos memory retrieval model with a Chebyshev-type activation function as an artificial chaos neuron. According to certain numerical analyses of the present association model with autocorrelation connection matrix between neurons, the dependence of memory retrieval properties on the initial Hamming distance between the input pattern and a target pattern to be retrieved among the embedded patterns will be presented to examine the retrieval abilities, i.e. the memory capacity of the associative memory.

**Keywords:** Chebyshev Chaos neuron, associative memory, computer simulation, numerical analysis methods, memory retrieval

### **1. Introduction**

Over the past quarter century, it has been extensively reported that there may exist inherently chaotic dynamics in the human electroencephalogram (EEG) in the variety of biological experiments [1–8]. In addition from the viewpoints of the artificial neuron models including chaos neurons, many researchers have investigated the association memory models in terms of the various types of activation functions of artificial neurons [9–33] so as to improve their memory retrieval capabilities as reviewed in brief in the following section.

First, as an artificial neuron model, Tsuda reported a dynamic retrieval model as well as dynamic linking of associative memories and pointed out a crucial role of chaos in those dynamics [9–11]. In addition, Davis and Nara *et al*. put forward a new memory search model with a chaos control [12, 13]. Subsequently, several applications of the chaos neural networks with the sigmoidal, i.e. monotonous, activation function have been reported by Aihara *et al*. [14] and Nakamura and Nakagawa [15]. In practice, however, as was confirmed by Kasahara and Nakagawa [16], the chaotic dynamics in association process with such a sigmoidal, i.e. monotonous, activation function encounters an inevitable troublesome such that the complete association of the embedded patterns becomes considerably difficult if the loading rate, i.e. *α* ¼ *L=N*, in which *L* and *N* are the number of embedded patterns and the neurons, respectively, increases beyond a certain critical values *αc*, i.e. *α<sup>c</sup>* � 0.05, for the autocorrelation learning model.

In contrast to the aforementioned monotonous chaos neuron model [14–16], the neurodynamics with a nonmonotonous activation function was investigated by Morita [19], in which he reported that the nonmonotonous mapping in a neurodynamics may possess a certain advantage of the critical loading rate, *αc*�0.27, superior than *αc*�0.15 for the conventional association model as the Associatron with such a signum activation function [17]. Later, Shiino and Fukai examined the memory capacity for a somewhat simplified step-like nonmonotonous activation function with the continuous time model, instead of the aforementioned discrete time ones, in an appropriate statistical manner, and concluded that the complete association could be achieved up to the critical loading rate such as �0.42 [20]. Subsequently, several models of nonmonotonous activation functions have been proposed including a sinusoidal activation function to control chaos dynamics [21–29]. In practice, the present author reported a chaos neuron model with a periodic activation function in order to construct a chaos association model with a discrete time as well as the orthogonal learning scheme [21–24]. Therein, the memory capacity was found to be increased up to *α<sup>c</sup>* �0.4, even in the search mode without any key information for the input vector, beyond the previously proposed monotonous dynamic models with the discrete time [14–16]. As the further possible applications, such a chaotic dynamics was involved in the synergetic neural network [25, 26]. The earlier-noted chaos neuron models with such sinusoidal periodic activation functions, which are referred as the sinusoidal chaos neuron model in the present work, have been extensively applied to the chaos association models to mainly evaluate their memory capacities [21–24, 27–34].

From the earlier-noted aspects, in this work, we shall propose a novel chaos autoassociation model with the Chebyshev-type activation function recently proposed [35] which may have an advantage to promote the memory capacity of the associative memory beyond the conventional models, proposed up to date, e.g. the sinusoidal activation function and the signum one [9–24, 27–34].

In §2, a theoretical framework of the present association model will be described making use of the Chebyshev-type chaos activation function [35]. Therein, a chaos control related to the chaos simulated annealing as previously mentioned [21–35] will be introduced to transfer the system from a strongly chaotic initial state to a moderate one which may lead the associative model to a memory retrieval point. Furthermore, some computer simulation results for the memory retrieval characteristics will be given in §3, in which an apparent advantage of the present chaos associative model will be elucidated in comparison with the previously proposed sinusoidal activation function model [21–24, 27–33] as well as the Associatron [17]. Finally, §4 is devoted to a few concluding remarks on the presently proposed chaos associative model as well as the future works to be investigated further.

### **2. Theory**

Denoting the *i-th* component of the *r-th* embedded, or memorized, vector as *e* ð Þ*r i* , the autocorrelation connection matrix can be defined as follows in accordance with the conventional autocorrelation learning model [9–19]:

$$w\_{\vec{\eta}} = \frac{1}{N} \sum\_{r=1}^{L} e\_i^{(r)} e\_j^{(r)},\tag{1}$$

where *N* and *L* are the numbers of the neurons and the embedded vectors, respectively. Then the loading rate, which corresponds to a ratio of the number of the embedded vectors to that of neurons, *α* is to be defined by:

$$a = \mathcal{L}/\mathcal{N}.\tag{2}$$

Therefore, one may evaluate the memory retrieval performance depending on the loading rate *α* for an initial vector which may involve a certain noise in a certain vector among the embedded ones, as will be explained later in detail. Then the updating rule of the internal state of the *i-th* neuron, *σi*ð Þ*t* , in the presently concerned discrete time model may be introduced by:

$$
\sigma\_i(t+1) = (1 - k\_t)\sigma\_i(t) + k\_t \sum\_{j=1}^N w\_{ij} s\_j(t), \tag{3}
$$

where *ks*ð Þ 0< *ks* ≤1 is considered to be a relaxation parameter of the updating process of the internal state *σi*ð Þ*t* [34], and the output of the *i-th* neuron *si*ð Þ*t* is to be related to *σi*ð Þ*t* in terms of

$$\mathfrak{s}\_i(t) = f(\sigma\_i(t));$$

here *f*ð Þ• is an activation function of the neuron [9–35], which is closely related to the dynamics of the neural networks [34] as has been reported in the conventional models with several nonmonotonous activation functions [9–13, 18–20] as well as the chaos associative models [21–24, 27–34] with periodic activation function.

In the present work, we shall focus our interests on the recently proposed Chebyshev-type activation function as follows [35]:

$$\mathfrak{s}\_{i}(t) = \tanh^{-1}\{\sin\left(\nu \sin^{-1}(\tanh \sigma\_{i}(t))\right)\},\tag{5}$$

where *ν* may be regarded as a chaos control parameter to be related to the strength of chaos as will be mentioned later [35]. If we ignore the operations of tanh �<sup>1</sup> ð Þ• and tanh ð Þ• in Eq. (5), the present activation function is readily reduced to the original Chebyshev function of the second kind [36] as was previously argued in detail [35]. That is, the activation function as Eq. (5) is regarded as a modified version of the original Chebyshev function [36] so as to apply it to the practical applications where the range of the internal states *σi*ð Þ*t* may not be restricted in general as j j *σi*ð Þ*t* ≤1, which is to be forced in the original Chebyshev function [35, 36].

To review the chaotic property of the present chaotic neurons [35], let us present the fundamental properties of a single neuron in the following section. Several profiles of the activation function of Eq. (5) for several *ν* are depicted in **Figure 1**. As a specific case, assuming that *σ<sup>i</sup>* ! �∞ in Eq. (5), one finds

$$\begin{split} s\_i(t) &= \lim\_{\sigma\_i(t) \to \pm \infty} \tanh^{-1} \left\{ \sin \left( \nu \sin^{-1} (\tanh \sigma\_i(t)) \right) \right\} \\ &= \tanh^{-1} \left\{ \sin \left( \pm \nu \frac{\pi}{2} \right) \right\} = \pm \tanh^{-1} \left\{ \sin \left( \nu \frac{\pi}{2} \right) \right\}, \end{split} \tag{6}$$

#### **Figure 1.**

*The presently introduced Chebyshev-type activation function [35]. Herein, the control parameter ν is set to ν* ¼ 2, 3, 4, 5 *for each profile of the activation function defined by Eq. (5). The solid, dashed, dashed-dotted, and dotted lines are for ν* ¼ 2*, ν* ¼ 3*, ν* ¼ 4*, and ν* ¼ 5*, respectively.*

as is confirmed in **Figure 1**. Herein, the subscripts *i* of *σi*ð Þ*t* and *si*ð Þ*t* are ignored for the moment for a single neuron.

Then, in order to examine a chaotic dynamics of the present chaos neuron model, replacing *si*ð Þ*t* with *σi*ð Þ *t* þ 1 in Eq. (5), we have the following updating rule for the internal state *σi*ð Þ*t* of the *i-th* neuron [35]:

$$\sigma\_i(t+1) = \tanh^{-1}\left\{\sin\left(\nu \sin^{-1}(\tanh \sigma\_i(t))\right)\right\}.\tag{7}$$

As an example, the chaotic behavior of the internal state *σi*ð Þ*t* in terms of Eq. (7) for a single chaos neuron, in which the subscript *i* of *σi*ð Þ*t* is ignored again, is presented in **Figure 2**, where *ν* was set to *ν* ¼ 2. Moreover, the corresponding bifurcation

**Figure 2.** *An example of chaotic behavior of the internal state σ*ð Þ*t according to Eq. (7) where ν* ¼ 2 *or τ* ¼ 2*=ν* ¼ 1*.*

*A Chaos Auto-Associative Model with Chebyshev Activation Function DOI: http://dx.doi.org/10.5772/intechopen.106147*

**Figure 3.** *The bifurcation diagram for the Chebyshev-type mapping given by Eq. (7) [35].*

diagram is depicted in **Figure 3** over *τ* ¼ *τ*<sup>0</sup> � 2 0ð Þ <*τ*<sup>0</sup> < <1 , or *ν* ¼ 1 � þ∞, in the abscissa, where *ν* is related to *τ* in terms of [35]:

$$
\nu = \mathbf{2/r}.\tag{8}
$$

Therein, one may confirm the variation of the chaos strength, or the complexity of dynamics, depending on *τ*. As a special case with *ν* ¼ 1 or *τ* ¼ 2*=ν* ¼ 2, one may readily confirm the following relation:

$$
\sigma\_i(t+1) = \tanh^{-1}\left\{\sin\left(\sin^{-1}(\tanh\sigma\_i(t))\right)\right\} = \sigma\_i(t),
\tag{9}
$$

which corresponds to an identity mapping such that *si*ðÞ¼ *t σi*ð Þ*t* .

Through the updating rule of Eq.(7), the corresponding Lyapunov exponent *λ* has to be derived as [35]:

$$\lambda = \lim\_{T \to \infty} \frac{1}{T} \sum\_{n=0}^{T-1} \log \left| \frac{d\sigma(t+1)}{d\sigma(t)} \right|$$

$$= \lim\_{T \to \infty} \frac{1}{T} \sum\_{n=0}^{T-1} \log \left| \frac{\nu \cos z(t)}{\cos \nu z(t)} \right| \tag{10}$$

$$= \log \nu + \lim\_{T \to \infty} \frac{1}{T} \sum\_{n=0}^{T-1} \log \left| \frac{\cos z(t)}{\cos \nu z(t)} \right|,$$

where

$$z(t) = \sin^{-1}\{\tanh\sigma(t)\}.\tag{11}$$

To derive Eq. (10), we have noticed the following derivations:

$$\frac{d\sigma(t+1)}{d\sigma(t)} = \frac{d\mathbf{g}\_{\nu}(\eta(t))}{d\eta(t)} \frac{d\eta(t)}{d\sigma(t)}$$

$$= \frac{d\mathbf{g}\_{\nu}(\eta(t))}{d\boldsymbol{f}\_{\nu}(\eta(t))} \frac{d\boldsymbol{f}\_{\nu}(\eta(t))}{d\eta(t)} \frac{d\eta(t)}{d\sigma(t)}\tag{12}$$

$$= \nu \frac{\cos\left(\boldsymbol{z}(t)\right)}{\cos\left(\nu\boldsymbol{z}(t)\right)},$$

where *gν*, *f <sup>ν</sup>*, *η*ð Þ*t* , and *z t*ð Þ are defined as:

$$\lg\_{\nu}(\eta(t)) = \tanh^{-1}[f\_{\nu}(\eta(t))],\tag{13}$$

$$f\_{\nu}(\eta(t)) = \sin\left\{\nu \sin^{-1}(\eta(t))\right\} = \sin\left\{\nu z(t)\right\},\tag{14}$$

$$\eta(t) = \sin z(t) = \tanh \sigma(t),\tag{15}$$

respectively [35].

The dependence of the Lyapunov exponents *λ* derived by Eq. (10) on *τ* ¼ 2*=ν* are illustrated with dot symbols in **Figure 4**, where *λ* is found to be decreased as *τ* increases toward *ν* ¼ 2*=τ* ¼ 1 as Eq. (9). In addition, the following approximated expression [34] of the Lyapunov exponents is also depicted as a solid curve in **Figure 4**:

$$
\lambda \simeq \log \nu = \log \left(\frac{2}{\pi}\right). \tag{16}
$$

Therein, one may confirm that the aforementioned expression holds approximately in comparison with the numerical results with Eqs. (7) and (10). Hence,

#### **Figure 4.**

*The Lyapunov exponents for the Chebyshev-type mapping given by Eq. (7) for* 0<*τ* ¼ 2*=ν*<2*. Herein, the Lyapunov exponents in the ordinate are normalized in terms of log2 for convenience. The dots and the solid line are for the numerical results derived by Eqs. (7) and (10) and the approximated expression,* log 2ð Þ *=τ =* log 2 *as in Eq. (16), respectively.*

according to Eqs. (10) and (16), the following relation is considered to be approximately satisfied [35]:

$$\lim\_{T \to \infty} \frac{1}{T} \sum\_{t=0}^{T-1} \log \left| \frac{\cos z(t)}{\cos \nu z(t)} \right| \simeq \mathbf{0}. \tag{17}$$

Especially for a specific case of *ν* ¼ 2, or *τ* ¼ 1, an analytical expression of the Lyapunov exponent *λ* has been derived as [35]

$$
\lambda = \log 2,\tag{18}
$$

which is found to be exactly same with that of the logistic mapping because of the topological conjugacy of the Chebyshev-type mapping as Eq. (5) with the conventional logistic one [35]. In the aforementioned work [35], the corresponding invariant measure was also investigated further in comparison with the sinusoidal and the monotonous activation functions [37].

From these results, the strength of chaos is found to be properly controlled, through the parameter *τ* ¼ 2*=ν* as in the previous chaos associative memories [21–24, 27–34] as well as the learning models with the back-propagation algorithm [38–40].

In the similar manner to the previous models [21–24, 27–34, 37, 38, 41], the control dynamics of the parameter *τ*ð Þ*t* may be simply introduced as follows:

$$
\pi(t+1) = \pi(t) + \kappa(\tau\_{\circ \circ} - \pi(t)),
\tag{19}
$$

where 0< *κ* < 1 is regarded as a parameter to control the number of the steps, or the speed, of the transient retrieval process from an initial chaotic state with a relatively small *τ*ð Þ¼ 0 *τ*0ð Þ < <1 to a memory retrieval point with *τ*ðÞ� *t τ*∞. Since *ν* ¼ 2*=τ*, *ν* varies from 2*=τ*0ð Þ > > 1 to 2*=τ*<sup>∞</sup> as *τ*ð Þ*t* changes from *τ*ð Þ¼ 0 *τ*0ð Þ < <1 to *τ*ð Þ¼ ∞ *τ*<sup>∞</sup> according to Eq. (19). Solving the difference equation as Eq. (19) with an initial value of *τ*ð Þ¼ 0 *τ*0ð Þ < <1 , one may readily derive the following analytical expression for *τ*ð Þ*t* :

$$\begin{split} \tau(t) &= \tau\_{\infty} - (\tau\_{\infty} - \tau(\mathbf{0})) (\mathbf{1} - \kappa)^{t} \\ &= \tau\_{\infty} - (\tau\_{\infty} - \tau\_{0}) \exp\left(-t/T\_{c}\right), \end{split} \tag{20}$$

where *Tc* is regarded as a time constant for the retrieval process from *τ*ð Þ¼ 0 *τ*0ð Þ 0<*τ*<sup>0</sup> < <1 to *τ*ð Þ¼ ∞ *τ*<sup>∞</sup> and given by:

$$T\_c = -\frac{1}{\log\left(1 - \kappa\right)} = \frac{1}{\log\left(\frac{1}{1 - \kappa}\right)}.\tag{21}$$

Updating *τ*ð Þ*t* with a certain initial value as *τ*ð Þ¼ 0 *τ*0ð Þ 0< *τ*<sup>0</sup> < < 1 and solving Eq. (3) with Eqs. (5), (8), and (19) simultaneously under an initial vector set *si*ð Þ 0 ð Þ 1≤*i* ≤ *N* which has a Hamming distance with respect to a target vector *e* ð Þ*s <sup>i</sup>* ð Þ 1≤ *i*≤ *N*, 1≤*s*≤*L* to be retrieved among the embedded vectors, we may construct a chaos auto-associative model with the Chebyshev-type activation function [35].

### **3. Computer simulation results**

In this section, we shall show some numerical results derived from the aforementioned chaos auto-association model defined by Eqs. (1), (3), (5), (8), and (19).

For numerical simulations, the embedded vectors *e* ð Þ*r <sup>i</sup>* ð Þ 1≤*i* ≤ *N*, 1≤*r*≤*L* will be defined as random bipolar vectors as follows:

$$\mathbf{z}\_i^{(r)} = \text{sign}\left(\mathbf{z}\_i^{(r)}\right),\tag{22}$$

where *sign*ð Þ• is the signum function such that

$$\text{sign}(\mathbf{x}) = \begin{cases} +\mathbf{1}(\mathbf{x} > \mathbf{0}) \\ \mathbf{0}(\mathbf{x} = \mathbf{0}) \end{cases},\tag{23}$$
 
$$-\mathbf{1}(\mathbf{x} < \mathbf{0})$$

and *z* ð Þ*r <sup>i</sup>* ð Þ 1≤*i* ≤ *N*, 1≤ *r*≤ *L* are the pseudo-random numbers between �1 and + 1. In addition, *e* ð Þ*r <sup>i</sup>* ð Þ 1≤*r*≤ *N* are assumed to be linearly independent each other to avoid the redundancy among the embedded vectors such that

$$\det\left(\sum\_{i=1}^{N} e\_i^{(r)} e\_i^{(s)}\right) \neq \mathbf{0},\tag{24}$$

where detð Þ• means the determinant of a matrix •. Then the connection matrix, i.e. the autocorrelation matrix, between the neurons, *wij* is straightforwardly derived in terms of Eq. (1).

The initial conditions for *si*ð Þ*t* ð Þ 1≤ *i*≤ *N* may be set [21–24, 27–33] in the following manner:

$$s\_i(\mathbf{0}) = \begin{cases} -e\_i^{(s)}(\mathbf{1} \le i \le H\_d) \\ e\_i^{(s)}(H\_d + \mathbf{1} \le i \le N) \end{cases},\tag{25}$$

where *Hd* is the Hamming distance between the initial vector *si*ð Þ 0 and a concerned target vector *e* ð Þ*s <sup>i</sup>* ð Þ ∀*s* ∈f g 1, 2,•••, *L* to be retrieved which is randomly chosen among the embedded vectors *e* ð Þ*r <sup>i</sup>* ð Þ 1≤*r*≤*L* . Of course, since the presently introduced embedded patterns are defined as the random bipolar vectors as in Eq. (22), Eq. (25) is substantially equivalent to random choice of the bit-reversed components in an initial vector *si*ð Þ 0 with certain amounts of bit-reversed *Hd* components corresponding to the involved noise in *si*ð Þ 0 . For a minimum noise, or a minimum Hamming distance, included in the initial vector *si*ð Þ 0 , *Hd* is to be set to 1, which corresponds to a singlebit reverse in the initial vector*si*ð Þ 0 from *e* ð Þ*s <sup>i</sup>* ð Þ ∀*s*∈f g 1, 2,•••, *L* . For the initial values of the vector, *si*ð Þ 0 , the corresponding initial values for the internal states *σi*ð Þ 0 may be set to as:

$$\sigma\_i(\mathbf{0}) = \sum\_{j=1}^{N} w\_{ji} s\_j(\mathbf{0}),\tag{26}$$

#### *A Chaos Auto-Associative Model with Chebyshev Activation Function DOI: http://dx.doi.org/10.5772/intechopen.106147*

setting as *σi*ð Þ¼ *t* þ 1 *σi*ðÞ¼ *t σi*ð Þ 0 and *si*ðÞ¼ *t si*ð Þ 0 in Eq. (3).

For convenience, the ratio of the distorted components according to Eq. (25), or the Hamming distance *Hd*, to the number of neurons, *N*, may be introduced by:

$$r\_d = H\_d / \text{N.}\tag{27}$$

Hence, the directional cosine, i.e. the inner product, *ς*ð Þ*<sup>s</sup>* between the input vector *si*ð Þ 0 and a target vector *e* ð Þ*s <sup>i</sup>* ð Þ ∀*s*∈f g 1, 2,•••, *L* can be evaluated in terms of *rd* as follows:

$$\boldsymbol{\varsigma}^{(s)} = \frac{1}{N} \sum\_{i=1}^{N} \boldsymbol{e}\_i^{(s)} \boldsymbol{s}\_i(\mathbf{0}) = \mathbf{1} - \frac{2H\_d}{N} = \mathbf{1} - 2\boldsymbol{r}\_d. \tag{28}$$

It is well known that the succeeded memory retrieval becomes apparently troublesome if *<sup>ς</sup>*ð Þ*<sup>s</sup>* is decreased from *<sup>ς</sup>*ð Þ*<sup>s</sup>* <sup>¼</sup> 1 to *<sup>ς</sup>*ð Þ*<sup>s</sup>* � 0 as an initial condition, since the involved information in order to retrieve a target vector *e* ð Þ*s <sup>i</sup>* ð Þ ∀*s*∈f g 1, 2,•••, *L* tends to be depressed as *rd* increases from *rd* ¼ 0 to *rd* � 0*:*5 [9–24, 27–34, 41].

Then, in similar manner to the previous works [21–24, 27–34, 41] for the associative memories, the overlaps *<sup>o</sup>*ð Þ*<sup>r</sup>* ð Þ*<sup>t</sup>* with respect to the embedded vector *<sup>e</sup>* ð Þ*r <sup>i</sup>* ð Þ 1≤*r*≤*L* for the output vector *si*ð Þ*t* may be defined as:

$$\boldsymbol{o}^{(r)}(t) = \frac{1}{N} \sum\_{i=1}^{N} \boldsymbol{e}\_i^{(r)} \boldsymbol{s} \text{sign}\{\boldsymbol{s}\_i(t)\}. \tag{29}$$

If one of the overlaps *<sup>o</sup>*ð Þ*<sup>r</sup>* ð Þ*<sup>t</sup>* ð Þ <sup>1</sup>≤*r*≤*<sup>L</sup>* , e.g.*o*ð Þ*<sup>s</sup>* ð Þ*<sup>t</sup>* , leads to <sup>þ</sup>1 or �1, then one may determine that a complete memory retrieval corresponding to *e* ð Þ*s <sup>i</sup>* is achieved with *si*ðÞ¼ *t e* ð Þ*s <sup>i</sup>* . It has to be noted here that there exists an uncertainty for the sign of *<sup>o</sup>*ð Þ*<sup>s</sup>* ð Þ*<sup>t</sup>* because of the invariant property of the dynamic equation, Eq. (3), and the activation function Eq. (5), under *si*ðÞ!� *t si*ð Þ*t* as well as *σi*ðÞ!� *t σi*ð Þ*t* . Thus, the success rate for the different random embedded sets *e* ð Þ*r <sup>i</sup>* <sup>¼</sup> *sign z*ð Þ*<sup>r</sup> i* � �ð Þ <sup>1</sup><sup>≤</sup> *<sup>r</sup>*<sup>≤</sup> *<sup>L</sup>* may be evaluated by examining the succeeded memory retrieval according to the following criteria:

$$\left| \max\_{1 \le r \le L} |o^{(r)}(t)| - 1 \right| = \left| \max\_{1 \le r \le L} \left| \frac{1}{N} \sum\_{i=1}^{N} e\_i^{(r)} \text{sign} \{ s\_i(t) \} \right| - 1 \right| < \epsilon (<<1), \tag{30}$$

and

$$|\tau(t+1) - \tau(t)| < \varepsilon(<<1),\tag{31}$$

as *τ*ð Þ*t* approaches to *τ*<sup>∞</sup> starting from *τ*ð Þ¼ 0 *τ*0ð Þ < <1 in accordance with Eq. (19) [33, 34]. Provided that both conditions such as Eqs. (30) and (31) are satisfied simultaneously, the current memory retrieval may be decided as a succeeded one. Alternatively, if only the latter condition as Eq. (31) is satisfied, the retrieval trial is judged as a failed one. In any cases, the next retrieval process with Eq. (3) will be restarted with the other random set of embedded vectors as Eq. (22) to constitute the connection matrix *wij* as Eq. (1) resetting *τ*ð Þ*t* to the initial value of *τ*ð Þ¼ 0 *τ*0, *si*ð Þ 0 ,

and *σi*ð Þ 0 to Eqs. (25) and (26) with fixed *α* and *rd*. This retrieval process will be successively repeated up to the total number of trials, i.e. *Ts*.

Thus through the present numerical simulations for certain values of *α* ¼ *L=N* and *rd* ¼ *Hd=N*, one may evaluate the number of the succeeded trials *Ns*ð Þ *α*,*rd* among *Ts* trials with the aforementioned different random bipolar vector sets *e* ð Þ*r <sup>i</sup>* ð Þ 1≤*i*≤ *N*, 1≤*r*≤*L* and the connection matrices *wij* as defined in terms of Eqs. (22) and (1). Here, the number of trials *Ts* in the retrieval simulations corresponds to the number of the retrieval processes from *τ*ð Þ¼ 0 *τ*0ð Þ 0<*τ*<sup>0</sup> < <1 to *τ*ð Þ¼ ∞ *τ*<sup>∞</sup> according to Eq. (19) or Eq. (20) with the criterion Eq. (31). Then, counting up the total number of the succeeded trials as *Ns*ð Þ *α*,*rd* among *Ts* trials, the succeeded association rate *Sr*ð Þ *α*,*rd* in accordance with the aforementioned criteria with Eqs. (30) and (31) for a pair of fixed parameters, *α* ¼ *L=N* and *rd* ¼ *Hd=N*, will be introduced as follows:

$$S\_r(a, r\_d) = \frac{N\_s(a, r\_d)}{T\_s}.\tag{32}$$

Thus, the ability of the memory retrievals for the associative memory model may be evaluated in terms of the score of the succeeded memory retrievals rate to satisfy the conditions and Eqs. (30) and (31) for a certain fixed value of *rd* ¼ *Hd=N*, i.e. *Sr*ð Þ *α*,*rd* vs. *α* ¼ *L=N* characteristics as will be presented later. Hence, the memory capacity *Mc*ð Þ *rd* may be defined as the total area of *Sr*ð Þ *α*,*rd* (the ordinate) vs. *α* ¼ *L=N* (the abscissa) curves as follows:

$$\mathcal{M}\_c(r\_d) = \int\_0^1 da \mathcal{S}\_r(a, r\_d). \tag{33}$$

In the following practical simulations, the aforementioned parameters, *N*, *ks*, *κ*, *Ts*, *τ*ð Þ¼ ∞ *τ*∞, *τ*ð Þ¼ 0 *τ*0, and *ε* (see Eqs. (30) and (31).) are set to 1000, 0.2, 0.5, 100, 1, 10�10, and 10�10, respectively, if not mentioned.

In **Figure 5a**–**e**, the retrieval characteristics, i.e. the dependence of the succeeded association rate defined by Eq. (32), *Sr*ð Þ *α*,*rd* , on the loading rate *α* ¼ *L=N* are depicted for *rd* <sup>¼</sup> <sup>1</sup>*=<sup>N</sup>* <sup>¼</sup> <sup>0</sup>*:*001(*ς*ð Þ*<sup>s</sup>* <sup>¼</sup> <sup>0</sup>*:*998), *rd* <sup>¼</sup> <sup>0</sup>*:*1(*ς*ð Þ*<sup>s</sup>* <sup>¼</sup> <sup>0</sup>*:*8), *rd* <sup>¼</sup> <sup>0</sup>*:*2 (*ς*ð Þ*<sup>s</sup>* <sup>¼</sup> <sup>0</sup>*:*6), and *rd* <sup>¼</sup> <sup>0</sup>*:*3(*ς*ð Þ*<sup>s</sup>* <sup>¼</sup> <sup>0</sup>*:*4), respectively. From these results, the memory retrieval abilities of the presently proposed Chebyshev-type associative memory model are found to be apparently superior than the conventional association models [9–24, 27–34, 41]. By increasing *rd* <sup>¼</sup> *Hd=N*, i.e. decreasing *<sup>ς</sup>*ð Þ*<sup>s</sup>* as Eq. (28), the memory capacities *Mc*ð Þ *rd* defined by Eq. (33) are found to be decreased in similar to the conventional associative models [9–24, 27–34, 41]. In practice, the memory capacities *Mc*ð Þ *rd* were evaluated as *Mc*ð Þ¼ *rd* ¼ 0*:*001 0.4858, *Mc*ð Þ¼ *rd* ¼ 0*:*1 0.4292, *Mc*ð Þ¼ *rd* ¼ 0*:*2 0.3680, *Mc*ð Þ¼ *rd* ¼ 0*:*3 0.2472, and *Mc*ð Þ¼ *rd* ¼ 0*:*4 0.1168. From these findings, one may confirm the advantage of the present chaos associative model with the Chebyshev-type activation function as Eq. (5) beyond the conventional association models [21–24, 27– 34, 41]. This point will be mentioned later in detail as shown in **Figure 6**.

In **Figure 7**, let us present some examples of time dependences of the internal states *<sup>σ</sup>i*ð Þ*<sup>t</sup>* in Eq. (3), the outputs *si*ð Þ*<sup>t</sup>* derived from Eq. (5), and the overlaps *<sup>o</sup>*ð Þ*<sup>r</sup>* ð Þ*<sup>t</sup>* defined by Eq. (29) under an updating process of *τ*ð Þ*t* according to Eq. (19) with an initial value *τ*ð Þ¼ 0 *τ*<sup>0</sup> as shown in **Figure 7d** for a certain loading rate *α* ¼ *L=N* ¼ 0*:*3 and *rd* ¼ *Hd=N* ¼ 0*:*1, 0*:*2, 0*:*3. From these results, one may confirm the success of the memory retrievals for (a-1)–(a-3) with *rd* ¼ 0*:*1 as well as (b-1)-(b-3) with *rd* ¼ 0*:*2,

*A Chaos Auto-Associative Model with Chebyshev Activation Function DOI: http://dx.doi.org/10.5772/intechopen.106147*

#### **Figure 5.**

*Memory retrieval characteristics for the Chebyshev-type activation function as in Eq. (5). Here, the abscissa and the ordinate are for the loading rate α* ¼ *L=N and the succeeded association rate Sr*ð Þ *α*,*rd defined by Eq. (32). (a) rd* ¼ 0*:*001*, (b) rd* ¼ 0*:*1*, (c) rd* ¼ 0*:*2*, and (d) rd* ¼ 0*:*3*.*

whereas the memory retrieval is no longer to be attained for (c-1)–(c-3) with *rd* ¼ 0*:*3. This tendency is found to coincide with the previously shown results for the success rates for *rd* ¼ 0*:*1, 0*:*2, 0*:*3 as shown in **Figure 5b**–**d**, respectively. In addition even for *α* ¼ 0*:*3 and *rd* ¼ 0*:*2, one may find the succeeded memory retrievals in which *<sup>o</sup>*ð Þ*<sup>s</sup>* ðÞ¼ *<sup>t</sup> Maxr <sup>o</sup>*ð Þ*<sup>r</sup>* ð Þ*<sup>t</sup>* approaches to +1 or � 1 to satisfy Eq. (30). Of course, it is well appreciated that, for such a case with *α* ¼ 0*:*3 and *rd* ¼ 0*:*2, the succeeded memory retrievals considerably become troublesome in the conventional association models [9–14, 18–24, 27–34]. Comparing the time dependences of *σi*ð Þ*t* or *si*ð Þ*t* between the succeeded cases(see **Figure 7a** and **b**.) and the failed one (see **Figure 7c**.), it may be found that, in such a failed retrieval case as in **Figure 7**(c-3), there exists no longer separated regime as *σi*ð Þ*t* <0, *σi*ð Þ*t* >0, or *si*ð Þ*t* < 0, *si*ð Þ*t* > 0 as *τ*ðÞ!*t τ*<sup>∞</sup> as can been seen in **Figure 7**(c-1), (c-2). In fact this tendency in **Figure 7**(c-1), (c-2) is obviously different from **Figure 7**(a-1), (a-2) and **Figure 7**(b-1), (b-2). As can be confirmed in **Figure 7**(c-3), *Maxr <sup>o</sup>*ð Þ*<sup>r</sup>* ð Þ*<sup>t</sup>* does no longer approach to 1 under such behaviors of *<sup>σ</sup>i*ð Þ*<sup>t</sup>* and *si*ð Þ*t* , in which one may confirm no longer the separated regime as *σi*ð Þ*t* <0, *σi*ð Þ*t* >0, or *si*ð Þ*t* <0, *si*ð Þ*t* >0 as *τ*ðÞ!*t τ*∞. Now we are at the position to compare the present results with the conventional models with the sinusoidal and the signum

#### **Figure 6.**

*Comparison of the memory capacities defined by Eq. (33), i.e. the total area Sr*ð Þ *α*,*rd vs. α* ¼ *L=N curves given in Figures 5, 8 and 9, depending on rd* ¼ *Hd=N between the presently examined three models, in which the solid, the dashed-dotted, and the dotted lines are for the Chebyshev-type activation function as Eq. (5), the sinusoidal activation function as Eq. (34) and the Associatron with signum activation function as Eq. (35), respectively.*

activation functions. In practice, we shall compare the present numerical results of the Chebyshev-type activation function with the previously reported sinusoidal activation function model [21–24, 27–34] and the Associatron with the signum activation function proposed by Nakano [17], which are given by:

$$s\_i(t) = \sin\left(\frac{\pi}{2\pi}\sigma\_i(t)\right),\tag{34}$$

and

$$s\_i(t) = \text{sign}(\sigma\_i(t)),\tag{35}$$

respectively; here *τ* ranges over *τ*<sup>0</sup> ≤*τ* ≤*τ*<sup>∞</sup> in Eq. (34), which may be controlled in terms of Eq. (19), and *sign*ð Þ• is again the signum function defined by Eq. (23). The fundamental chaotic property of a sinusoidal chaos neuron defined by Eq. (34) has been reported in the previous paper [35], in which, in similar manner to the present Chebyshev-type activation function, the strength of chaos, i.e. the Lyapunov exponents are found to be approximated as [34, 35, 37]:

$$
\lambda \sim \log\left(\frac{\pi}{4\pi}\right). \tag{36}
$$

Therefore, the chaos dynamics is expected to be properly controlled by the parameter *τ* [35] which is to be updated through Eq. (19) or Eq. (20) in similar to the aforementioned Chebyshev-type association model.

The resultant retrieval characteristics for Eqs. (34) (sinusoidal activation function) and (35) (signum activation function) are given in **Figures 8** and **9**, respectively. First, in **Figure 8** for the sinusoidal activation function as Eq. (34), according to these *α* vs. *Sr*ð Þ *α*,*rd* curves, the memory capacities as Eq. (33) were evaluated as

*A Chaos Auto-Associative Model with Chebyshev Activation Function DOI: http://dx.doi.org/10.5772/intechopen.106147*

*Mc*ð Þ¼ *rd* ¼ 0*:*001 0.3234, *Mc*ð Þ¼ *rd* ¼ 0*:*1 0.2078, *Mc*ð Þ¼ *rd* ¼ 0*:*2 0.1254, *Mc*ð Þ¼ *rd* ¼ 0*:*3 0.1028, and *Mc*ð Þ¼ *rd* ¼ 0*:*4 0.0560. Here, the control parameter *τ* ¼ *τ*ð Þ*t* in Eq. (34) was updated in terms of Eq. (19) as in the same manner to the previously mentioned Chebyshev-type model in **Figure 5**.

Then, according to *α* vs. *Sr*ð Þ *α*,*rd* curves in **Figure 9**, the memory capacities *Mc*ð Þ *rd* as Eq. (33) for the Associatron with Eq. (35) were evaluated as *Mc*ð Þ¼ *rd* ¼ 0*:*001 0.2282, *Mc*ð Þ¼ *rd* ¼ 0*:*1 0.1454, *Mc*ð Þ¼ *rd* ¼ 0*:*2 0.1402, and *Mc*ð Þ¼ *rd* ¼ 0*:*3 0.1278, *Mc*ð Þ¼ *rd* ¼ 0*:*4 0.070. These values of the memory capacities *Mc*ð Þ *rd* are found to be lower than the aforementioned Chebyshev-type associative model with Eq. (5) over

#### **Figure 7.**

*Some examples of time dependences of the internal states σi*ð Þ*t in Eq. (3), the outputs si*ð Þ*t derived from Eq. (5), and the overlaps o*ð Þ*<sup>r</sup>* ð Þ*<sup>t</sup> defined by Eq. (29) under an updating process of <sup>τ</sup>*ð Þ*<sup>t</sup> according to Eq. (19) with an initial value τ*ð Þ¼ 0 *τ*<sup>0</sup> *for α* ¼ *L=N* ¼ 0*:*3 *and rd* ¼ *Hd=N* ¼ 0*:*1, 0*:*2, 0*:*3*. (a-1) the time dependence of the internal states σi*ð Þ*t in Eq. (3) for rd* ¼ 0*:*1*. (a-2) the time dependence of the outputs si*ð Þ*t according to Eq. (5) for rd* ¼ 0*:*1*. (a-3) the time dependence of overlaps o*ð Þ*<sup>r</sup>* ð Þ*<sup>t</sup> defined by Eq. (29) for rd* <sup>¼</sup> <sup>0</sup>*:*1*. (b-1) the time dependence of the internal states σi*ð Þ*t in Eq. (3) rd* ¼ 0*:*2*. (b-2) the time dependence of the outputs si*ð Þ*t according to Eq. (5) for rd* <sup>¼</sup> <sup>0</sup>*:*2*. (b-3) the time dependence of overlaps o*ð Þ*<sup>r</sup>* ð Þ*<sup>t</sup> defined by Eq. (29) rd* <sup>¼</sup> <sup>0</sup>*:*2*. (c-1) the time dependence of the internal states σi*ð Þ*t in Eq. (3) for rd* ¼ 0*:*3*. (c-2) the time dependence of the outputs si*ð Þ*t according to Eq. (5) for rd* <sup>¼</sup> <sup>0</sup>*:*3*. (c-3) the time dependence of overlaps o*ð Þ*<sup>r</sup>* ð Þ*<sup>t</sup> defined by Eq. (29) for rd* <sup>¼</sup> <sup>0</sup>*:*3*. (d) the time dependence of τ*ð Þ*t , which follows Eq. (19), whose behavior is common to all figures, (a-1)–(c-3).*

0*:*001<*rd* ¼ *Hd=N* <0*:*4 and lower than the sinusoidal associative one with Eq. (34), especially for a relatively low *rd* such that *rd* ¼ *Hd=N* <0*:*1.

In **Figure 6**, let us present the dependence of the memory capacities *Mc*ð Þ *rd* on *rd* ¼ *Hd=N* for the presently concerned three models, i.e*.* the Chebyshev-type activation function as Eq. (5) [35], the sinusoidal activation function as Eq. (34) [21–23, 26–34], and the Associatron as Eq. (35) [17] for comparison. From these results, one may find the advantage of the presently proposed Chebyshev-type activation function as Eq. (5) beyond such previous models as Eqs. (34) [21–23, 26–34] and (35) [17] as well as the other association models with the autocorrelation connection matrix *wij* and the discrete time updating scheme as Eq. (3) [9–15, 18, 20]. Especially, for relatively small values of *rd* such that *rd* ¼ *Hd=N* <0*:*1, it is confirmed that the chaos retrieval models with the Chebyshev-type and the sinusoidal activation functions possess a remarkable advantage beyond the non-chaos association model such as the Assocciatron [17]. Alternatively,

*A Chaos Auto-Associative Model with Chebyshev Activation Function DOI: http://dx.doi.org/10.5772/intechopen.106147*

#### **Figure 8.**

*Memory retrieval characteristics for the sinusoidal activation function as in Eq. (34) instead of Eq. (5). Here the abscissa and the ordinate are for the loading rate α* ¼ *L=N and the success rate Sr*ð Þ *α*,*rd defined by Eq. (32). (a) rd* ¼ 0*:*001*, (b) rd* ¼ 0*:*1*, (c) rd* ¼ 0*:*2*, and (d) rd* ¼ 0*:*3*.*

for relatively large *rd* such that *rd* ¼ *Hd=N* > 0*:*2, the memory capacities of the sinusoidal chaos associative memory are found to be depressed and become slightly lower than those of the Associatron [17]. As a whole, however, the Chebyshev-type chaos association model shows an apparent advantage beyond these models, i.e. the sinusoidal chaos model [21–23, 26–34] and the Associatron [17] over *rd* ¼ 0*:*001 � 0*:*4. Therefore, one may conclude that the chaos retrieval memories such as the Chebyshev-type and sinusoidal models have an great advantage, especially for a relatively small *rd* rather than for a relatively large *rd*. This finding may indicate that the information involved in the input vector *si*ð Þ 0 related to *rd* as defined by Eq. (25) is to be closely related to the retrieval ability even for the chaos memory retrieval processes in similar to the conventional non-chaos association model such as the Assosiatron [17, 34]. It seems to be instructive to note here that the memory capacity *Mc*ð*rd* ! 1*=N* � 0Þ � 0*:*486 for the present Chebyshev-type association model with autocorrelation connection matrix *wij* is considerably close to �0.5 which is the critical value of the memory capacity for the orthogonal learning models with the signum-type activation function as Eq. (35), investigated by means of a statistical approach in the analogy with the spin glass system as previously reported [42, 43].

#### **Figure 9.**

*Memory retrieval characteristics for the signum function as in Eq. (35) instead of Eq. (5). Here the abscissa and the ordinate are for the loading rate α* ¼ *L=N and the association rate defined by Eq. (32). (a) rd* ¼ 0*:*001*, (b) rd* ¼ 0*:*1*, (c) rd* ¼ 0*:*2*, and (d) rd* ¼ 0*:*3*.*

### **4. Concluding remarks**

In this chapter, we have proposed a chaos associative memory model with the Chebyshev-type activation function [35] instead of the conventional ones including such as the sinusoidal one [21–24, 27–34] as well as the signum one [17]. From the present computer simulation results, we have confirmed the apparent advantage of the present chaos associative model beyond the aforementioned conventional models from the viewpoint of the memory capacities related to the robustness for the included noise, i.e. the distorted components in an input vector for the memory retrieval. In practice for the extreme case such as *rd* ¼ 1*=N*ð Þ < <1 , the memory capacities of the Chebyshev-type activation function as Eq. (5), the sinusoidal activation function as Eq. (34), and the signum function as Eq. (35) are found to be 0.486, 0.341, and 0.228, respectively. Moreover, it has been found that the presently focused Chebyshev-type activation function has a remarkable advantage beyond the sinusoidal association models [21–24, 27–34] as well as the non-chaos association model as the Assocciatron [17] as can be confirmed by *rd* vs. *Mc*ð Þ *rd* characteristics in **Figure 6**. Moreover, for relatively large values of *rd* such that *rd* >0*:*2, the memory capacities of *A Chaos Auto-Associative Model with Chebyshev Activation Function DOI: http://dx.doi.org/10.5772/intechopen.106147*

the sinusoidal chaos associative memory were found to be slightly lower than those of the Associatron [17] as can be seen in **Figure 6**, whereas that of the Chebyshev-type model has shown a relatively large memory capacities beyond both the sinusoidal chaos model [21–24, 27–34] and the Associatron [17]. Especially, it has to be borne in mind here again that the memory capacity of the Chebyshev-type model for *Hd* ¼ 1, or *rd* ¼ 1*=N*, is promoted up to *Mc*ð Þ� *rd* 0.486 which is remarkably close to that of the Associatron with the orthogonal learning model [42, 43] in which *Mc*ð Þ� *rd* 0.5.

Several investigations on an optimization scheme of the included model parameters such as *ks*, *κ*, *τ*0, and *τ*<sup>∞</sup> are now in progress and then will be reported elsewhere in the near future.

Moreover, it seems to be also interesting to investigate the orthogonal learning model [42, 43] as well as the continuous time model [20]. It seems to be worthwhile to introduce the present Chebyshev-type activation function to the chaotic leaning problems as previously reported [38–40] as well as the design of the classifier applied to the sensibility measurement scheme [44] in connection with an application to the communication technology, e.g. a silent speech interface as seen in the human-brain interfaces. In addition, a chaos learning model with the present Chebyshev-type activation function is to be investigated in comparison with the other chaos learning models with the sinusoidal activation functions [38–40].

## **Acknowledgements**

The present work was supported in part by START program by Japan Science and Technology Agency in 2014–2016. The present author would like to greatly appreciate for the Grants-in-Aid for Scientific Research of MEXT, No.21300081, No.23650109, and No.24300084. The present author also would like to sincerely express his gratitude to his academic staffs and also to his family for their continuously heartwarming supports.

### **Author details**

Masahiro Nakagawa Faculty of Engineering, Department of Electrical Engineering, Nagaoka University of Technology, Japan

\*Address all correspondence to: masanaka@vos.nagaokaut.ac.jp

© 2022 The Author(s). Licensee IntechOpen. This chapter is distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/3.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
