1. Introduction

There have been many studies on learning of fuzzy systems [1–8]. Their aim is to construct learning methods based on SDM. Some novel methods on them have been developed which (1) generate fuzzy rules one by one starting from any number of rules, or reduce fuzzy rules one by one starting from a sufficiently large number of rules [2]; (2) use genetic algorithm (GA) and particle swarm optimization (PSO) to determine fuzzy systems [3]; (3) use fuzzy inference systems composed of a small number of input rule modules, such as single input rule modules (SIRMs) and double input rule modules (DIRMs) methods [9, 10]; and (4) use a

© 2016 The Author(s). Licensee InTech. This chapter is distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/3.0), which permits unrestricted use, distribution, and eproduction in any medium, provided the original work is properly cited. © 2018 The Author(s). Licensee IntechOpen. This chapter is distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/3.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

self-organization or a vector quantization technique to determine the initial assignment of parameters [11–15, 19]. Specifically, it is known that learning methods using vector quantization (VQ) and steepest descent method (SDM) are superior in the number of rules (parameters) to other methods [16, 19]. So, why is it effective to combine VQ with SDM in fuzzy modeling? First, let us explain how to combine SDM with methods other than VQ. (1) Although the learning time is short, the generation method is known to have low test accuracy, while the reduction method has high test accuracy but takes long learning time [2]. (2) The method using GA and PSO shows high accuracy when the input dimension and the number of rules are small, but it is known that there is a problem of scalability [3]. (3) SIRM and DIRM methods are excellent in scalability, but the accuracy of learning is not always sufficient [9]. As described above, many methods are not necessarily effective models because of the difficulty of learning accompanying the increase of the input dimension and the number of rules and the low accuracy. On the other hand, the method combining VQ with SDM is possible to efficiently conduct learning of SDM by arranging suitably the initial parameters of fuzzy rules using VQ [1, 16]. However, since VQ is unsupervised learning, it is easy to reflect the input part of learning data, but how to capture output information in learning is difficult. With their studies, the first learning method is the one using VQ only in determining the initial parameters of the antecedent part of fuzzy rules using input part of learning data [1, 11–14]. The second method is the one determining the same parameter using input/output parts of learning data [15, 19]. Further, the third method is one iterating learning process of VQ and SDM for the second method. Kishida and Pedrycz proposed the method based on the third one [13, 15]. These methods are the ones determining only the antecedent parameters by VQ. Therefore, we introduced generalized inverse matrix (GIM) to determine the initial assignment of weight parameters for the consequent part of fuzzy rules as the fourth method and showed the effectiveness in the previous paper [16, 17]. In this paper, improved methods for learning process of SDM in learning methods using VQ, GIM, and SDM are introduced and show that the method is superior in the number of rules to other methods in numerical simulations.

A membership value μ<sup>i</sup> of the antecedent part for input x is expressed as

Then, the output y<sup>∗</sup> of fuzzy inference method is obtained as

If Gaussian membership function is used, then Mij is expressed as

Mij xj

)|p∈ZP } and D<sup>∗</sup> = {(xp

output y<sup>r</sup> and the inference output y<sup>∗</sup>

, yr

on SDM using the following relation:

, … , xp

square error (MSE) as

Let D = {(xp

where cij and bij denote the center and the width values of Mij, respectively.

.

<sup>E</sup> <sup>¼</sup> <sup>1</sup> P X P

where yp∗ and yr mean inference and desired output for the pth input x<sup>p</sup>

∂E ∂wi

<sup>¼</sup> <sup>μ</sup> P i n <sup>I</sup>¼<sup>1</sup> <sup>μ</sup><sup>I</sup>

<sup>¼</sup> <sup>μ</sup> P i n <sup>I</sup>¼<sup>1</sup> <sup>μ</sup><sup>I</sup>

∂E ∂cij

∂E ∂cij

where t is iteration time and K<sup>α</sup> is a learning constant [1].

<sup>¼</sup> <sup>μ</sup> P i n <sup>I</sup>¼<sup>1</sup> <sup>μ</sup><sup>I</sup>

<sup>μ</sup><sup>i</sup> <sup>¼</sup> <sup>Y</sup><sup>m</sup> j¼1

<sup>y</sup><sup>∗</sup> <sup>¼</sup>

� � <sup>¼</sup> exp � <sup>1</sup>

P<sup>n</sup>

Mij xj

<sup>i</sup>¼<sup>1</sup> <sup>μ</sup><sup>i</sup> � wi P<sup>n</sup> <sup>i</sup>¼<sup>1</sup> <sup>μ</sup><sup>i</sup>

2

The objective function E is determined to evaluate the inference error between the desirable

, …, x<sup>p</sup>

p¼1

In order to minimize the objective function E, each parameter of c, b, and w is updated based

<sup>∙</sup> <sup>y</sup><sup>∗</sup> � yr ð Þ<sup>∙</sup> wi � <sup>y</sup><sup>∗</sup> ð Þ<sup>∙</sup>

<sup>∙</sup> <sup>y</sup><sup>∗</sup> � yr ð Þ<sup>∙</sup> wi � <sup>y</sup><sup>∗</sup> ð Þ<sup>∙</sup> xj � cij � �<sup>2</sup>

of input part of D, respectively. The objective of learning is to minimize the following mean

y∗ <sup>p</sup> � yr p � �<sup>2</sup>

xj � cij

� � (2)

http://dx.doi.org/10.5772/intechopen.79925

Learning Algorithms for Fuzzy Inference Systems Using Vector Quantization

bij � �<sup>2</sup> ! (4)

)|p∈Zp} be the set of learning data and the set

.

<sup>∙</sup> <sup>y</sup><sup>∗</sup> � yr ð Þ (6)

xj � cij b2 ij

> b3 ij

(3)

131

(5)

(7)

(8)
