Upload
edgar-maya-perez
View
245
Download
0
Embed Size (px)
Citation preview
7/28/2019 Fuzzy vs Neural
1/10
356 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 11, NO. 2, MARCH 2000
The Equivalence Between Fuzzy Logic Systems andFeedforward Neural Networks
Hong-Xing Li and C. L. Philip Chen
AbstractThis paper demonstrates that fuzzy logic systemsand feedforward neural networks are equivalent in essence.First, we introduce the concept of interpolation representationsof fuzzy logic systems and several important conclusions. Wethen define mathematical model for rectangular wave neuralnetworks and nonlinear neural networks. With this definition,we prove that nonlinear neural networks can be represented byrectangular wave neural networks. Based on this result, we provethe equivalence between fuzzy logic systems and feedforwardneural networks. This result provides us a very useful guidelinewhen we perform theoretical research and applications on fuzzylogic systems, neural networks, or neuro-fuzzy systems.
Index TermsEquivalence on fuzzy logic systems, fuzzy logic
systems, feedforward neural networks, interpolation representa-tion, rectangular wave neural networks.
I. INTRODUCTION
MANY researches focused on combining neural networksand fuzzy logic systems, such as neuro-fuzzy systems,or fuzzy neural networks [1][3], [10][16]. They have shown
valuable results undoubtedly. However, this paper takes a dif-
ferent approach to demonstrate the equivalent relationship be-
tween fuzzy logic systems and neural networks. Based on this
result, we wish to provide more significant theoretical result
on combining both systems. To provide this, we first introduce
representations of fuzzy logic systems and then we elaboratethe idea of interpolation representation of fuzzy logic systems
and provide new results under some weaker conditions. We first
prove rectangular wave activation function neural networks can
represent nonlinear neural networks. Based on this result, we
further prove that fuzzy logic systems and neural networks are
equivalent essentially under some restriction.
First of all, we introduce interpolation representation of fuzzy
logic systems. We prove that the antecedents of inference of a
fuzzy logic system are the base functions of interpolation and
the consequents of inference only relate to their peak values but
not to the shape of the membership functions.
Second, we define mathematical model of rectangular wave
neural networks, where their activation functions are rectan-
Manuscript received February 8, 1999; revised July 29, 1999 and November22, 1999. This work was supported by the National Natural Science Foundationof China, U.S. AFOSR Grant F49620-94-0277, and NSF-EIA-9601670.
H.-X. Li is with the Department of Mathematics, Beijing Normal Uni-versity, Beijing 100875, China (e-mail: [email protected]). He is currentlyon leave with Wright State University, Dayton, OH 45435 USA (e-mail:[email protected]).
C. L. P. Chen is with the Department of Computer Science and En-gineering, Wright State University, Dayton, OH 45435 USA (e-mail:[email protected]).
Publisher Item Identifier S 1045-9227(00)01743-4.
gular waveforms. Then we prove that a nonlinear neural net-
work can be represented by a rectangular wave neural network.
By means of this result, we prove the equivalence between
fuzzy logic systems and feedforward neural networks. This con-
clusion provides an important theoretical tool or basis for fuzzy
logic systems or neural networks, especially when combining
both of them.
Similar work can be found from [12] and [16]. Jang [12]
treated the system using radial basis function and Sugeno
model. The equivalence is found by finding the mapping
between Sugeno model and radial basis function without some
rigorous proof. Buckley et al. [16] focused on the equivalencefor fuzzy expert system. In this paper, we study the interpola-
tion and equivalence under weak condition, i.e., Kronecker's
property.
II. INTERPOLATION REPRESENTATIONS OF FUZZY LOGIC
SYSTEMS
This section reviews and elaborates the interpolation mech-
anism or representations of fuzzy logic systems. The interpo-
lation mechanism or representation of fuzzy logic systems is
discussed in detail in [1]. Here, we review, briefly, some main
issues under weaker restrictions as follows.
A. Some Necessary Concepts and Notations
Let us quickly recall fuzzy logic systems, taking a two in-
puts and one output system as an example. Let and be
the universes of input variables and , respectively, and be
the universe of output variable . Denote
and , where
and , and and
is the family of all fuzzy sets on and , respec-
tively. We regard and as linguistic variables, so that a
group of fuzzy inference rules is formed as follows:
If is and is then is (1)
where , and and
are called base variables.
According to the Mamdanian algorithm, the inference rela-
tion of the th inference rule is a fuzzy relation from
to , where
10459227/00$10.00 2000 IEEE
http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-7/28/2019 Fuzzy vs Neural
2/10
7/28/2019 Fuzzy vs Neural
3/10
http://-/?-7/28/2019 Fuzzy vs Neural
4/10
LI AND CHEN: THE EQUIVALENCE BETWEEN FUZZY LOGIC SYSTEMS AND FEEDFORWARD NEURAL NETWORKS 359
approximately a binary piecewise interpolation function taking
for its base functions
(16)
Furthermore, when the system is normal and
is an equidistant partition, we have
(17)
Proof: In (2), after is replaced by
(18)
For given inputs and , we get
(19)
Similar to the proof of Lemma 2, we have
(20)
where
and . Let
and we get (16).
When the system is normal and is an
equidistant partition, it is easy to see that . So
(17) is also true.
E. Interpolation Representations of Weighted -Centroid
Algorithm
The weighted -centroid algorithm is proposed in [6],
which means the inference rules are joined by weighed or.Taking (18) for an example, we should have
(21)
where , and usually .
Lemma 5: Under the conditions in Lemma 2, there exists a
group of base functions such that
the weighted -centroid algorithm with two inputs and one
output is approximately a binary piecewise interpolation func-
tion taking for its base functions
(22)
where and
(23)
The proof is omitted for it is similar to the proof of lemma 4.
F. Interpolation Representation of the Simple
Inference Algorithm
The simple inference algorithm is proposed in [7] and [8]
which is thought by some people to be a kind of quick and
simple algorithm with respect to fuzzy inference. In fact, it
has been most widely used in neuro-fuzzy systems involving
learning mechanisms.
In the algorithm, the fuzzy sets representing inference con-sequents are replaced by numbers. For example, inference rules
with two inputs and one output are as follows:
If is and is then is (24)
Then for a given input , the response value is calcu-
lated in the following steps:
Step 1)
or
Step 2)
(25)
Lemma 6: The simple inference algorithm with two inputs
and one output is exactly (not approximately)a binary piecewise
interpolation. When
(26)
where
(27)
When and the system is normal
(28)
The proof is omitted for it is not difficult.
G. Interpolation Representation of Function
Inference Algorithm
The function inference algorithm is proposed in [9] which is
the generalization of the simple inference algorithm. We con-
http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-http://-/?-7/28/2019 Fuzzy vs Neural
5/10
7/28/2019 Fuzzy vs Neural
6/10
LI AND CHEN: THE EQUIVALENCE BETWEEN FUZZY LOGIC SYSTEMS AND FEEDFORWARD NEURAL NETWORKS 361
which represents a fuzzy logic system. Naturally, the interpola-
tion representation of the system can be given by the following
expression:
This provides us a general guideline to construct models or al-gorithms for fuzzy logic systems.
III. RECTANGULAR WAVE NEURAL NETWORKS AND
NONLINEAR NEURAL NETWORKS
As two of basic tools for proving the equivalence of fuzzy
logic systems and feedforward neural networks, we define and
discuss rectangular wave neural networks and nonlinear neural
networks in this section. First, let us recall the basic structure of
a neuron. At present, the models of neurons most in use are the
units with multiinput and single output, for example, MP model
(i.e., McCullochPitts model, see Fig. 1).
It is well known that the relation between the inputs and out-
puts of the neuron is expressed as follows:
(43)
If we take , then ,
which means that the threshold value has been absorbed into
. We reuse the notation instead of and (43) is changed as
the following:
(44)
where in is regarded as a -ary function
Let . Clearly, is just the mathematical representa-
tion of the artificial neuron. Pay attention to the simple fact that
is the composition of and , where plays a role in the
synthesizing of -dimensional Euclidean space and plays a
role in activating signals. In fact, is just a kind of multifacto-
rial function (see [10]); so usually we call a space synthesizer
or a synthetic function of space. And is called a signal acti-vator or an activation function.
Now we should understand that a neuron can be interpreted
as a function of several variables (including unary functions)
and from mathematical viewpoint a neural network is regarded
as a certain combinatorial form of some functions of several
variables.
Definition 1: A neuron is called a nonlinear neuron if its ac-
tivation function is a nonlinear function. A neural network is
called a nonlinear neural network, if there at least exists one
nonlinear neuron.
The following definition shows an important concept in this
paper.
Fig. 1. MP model of artificial neurons.
Definition 2: A neuron is called a rectangular wave neuron if
itsactivation function is just a characteristic function of a certain
real number interval (including infinite intervals; see Fig. 2). A
neural network is called a rectangular wave neural network if
every neuronof theneural network is a rectangular wave neuron.
Usually for the sake of convenience, we do not distinguish
between a neuron and its activation function. So in Fig. 1, weuse the notation of the activation function of a neuron to stand
for the neuron.
It is worthwhile to mention that rectangular wave (pulsesignal) neural networks can be easily implemented by hard-
ware.
Lemma 8: A nonlinear neural network can be, approxi-
mately, represented by a rectangular wave neural network.
Proof: For a given nonlinear neural network, without loss
of generality, we take the neural network as in Fig. 3, where
and are the activation functions of these
neuron with nonlinear activation function. Also without loss of
generality, we can suppose that these activation functions are
all nonlinear and continuous functions.
According to the relation between the input and output of the
neural network, we have
(45)
Let have the universe (domain) ( is the
set of real numbers). For any , we create a partition of ,
denoted by , where , and
form a group of zero-order base functions, as
shown in Fig. 4
otherwise(46)
such that where .
This means that . So
(47)
For any and for any , in a similar
way, we can obtain zero-order interpolation functions
(48)
http://-/?-http://-/?-7/28/2019 Fuzzy vs Neural
7/10
362 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 11, NO. 2, MARCH 2000
Fig. 2. The activation function of a rectangular wave neuron.
such that , i.e., , where the
universe of is and the base function
is as in Fig. 5. Let
(49)
which is just the relation between the input and output of
the simple neural network shown in Fig. 6. We have shown
that . We now prove
that can approximate
within an arbitrary accuracy. In fact, we consider the following
expression:
, since is a continuous function, such
that, , when .
Let . Because of being
arbitrary, we can make . Thus
Fig. 3. A nonlinear neural network.
Fig. 4. Rectangle wave activation functions as the basic functions.
Fig. 5. Zero-order activation function g ( v ) .
So . Also because of being
arbitrary, we can take . Hence
. This means
7/28/2019 Fuzzy vs Neural
8/10
LI AND CHEN: THE EQUIVALENCE BETWEEN FUZZY LOGIC SYSTEMS AND FEEDFORWARD NEURAL NETWORKS 363
Fig. 6. A rectangular wave neural network, where I ( x ) = 1 1 1 = I ( x ) = I ( x ) = I ( x ) = i d ( x ) .
, i.e., can approximate
within an arbitrary accuracy.
IV. THE EQUIVALENCE BETWEEN THE TWO SYSTEMS
Theorem: A fuzzy logic system is approximately equivalent
to a feedforward neural network.
Proof: Necessity: Arbitrarily given a fuzzy logic system,
for the convenience of the proof, we consider the case with
two-input and one-output. Without loss of generality, based on
the conclusions in Section II, the system can be regarded as an
interpolation function
Now we create a feedforward neural network shown in Fig. 7,where and are the neurons (regarded as functions of sev-
eral variables), holding
, i.e., considered as hyperbolic functions, and is
given as ; that is
Thus, the output of the network is as follows:
So . Clearly, the neural network is a non-
linear neural network. According to Lemma 8, it can be changedinto a rectangular wave neural network. This finishes the proof
of the necessity.
Sufficiency: Arbitrarily given a feedforward neural net-
works, without loss of generality, we consider the case with two
inputs and one output and three layers as an example shown in
Fig. 8.
We consider the output of the network
(50)
From (50) we should get a fuzzy logic system such that
is approximately the output of the fuzzy logic system. Using
7/28/2019 Fuzzy vs Neural
9/10
364 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 11, NO. 2, MARCH 2000
Fig. 7. The three-layer feedforward neural network with two inputs and one output.
Fig. 8. An arbitrarily given neural network.
(50), because can be regarded an unary function (such that
there exist at most finite discontinuous points) and for any
, there exists a group of base functions
such that
where
otherwise
Thus
(51)
7/28/2019 Fuzzy vs Neural
10/10
LI AND CHEN: THE EQUIVALENCE BETWEEN FUZZY LOGIC SYSTEMS AND FEEDFORWARD NEURAL NETWORKS 365
where is the characteristic function of the
set , (clearly, , for any ),
and
(52)If we form a group of fuzzy inference rules
If is then is (53)
for a fuzzy logic system, then (51) is just the interpolation rep-
resentation of the system according to the statement of Section
II-H. This finishes the proof of the sufficiency.
V. CONCLUSION REMARKS
We briefly summarize our main conclusions in this paper as
follows.
1) Fuzzy logic systems are equivalent to feedforward neural
networks, which means that, an arbitrarily given fuzzy
logic system can be approximately represented by a feed-forward neural network. Conversely, an arbitrarily given
feedforward neural network can be approximately repre-
sented by a fuzzy logic system. It is worth noting that here
approximately used by us means the approximation can
reach an arbitrarily given accuracy.
2) We define mathematical model of rectangular wave
neural networks and nonlinear neural networks. Then we
prove that nonlinear neural networks can be represented
by rectangular wave neural networks, which can be
implemented easily by hardware.
3) From Lemma 1 to Lemma 7, we discover a very impor-
tant and interesting conclusion: under the condition of
Kronecker's property (see Section II-A), the membership
functions defined on output variable universes do not take
effect in fuzzy logic systems but only their peak values do.
So, in many real applications, as long as the Kronecker's
property is satisfied, we do not need to care for the shapes
of these membership functions but only need to consider
whether their peak values are suitable.
REFERENCES
[1] H.-X. Li, The mathematical essence of fuzzy controls and fine fuzzycontrollers, in Advances in Machine Intelligence and Soft-Computing,P. P. Wang, Ed. Durham, NC: Bookwright, 1997, vol. IV, pp. 5574.
[2] W. A. Farag, V. H. Quintana, and G. Lambert-Torres, A genetic-based
neuro-fuzzy approach for modeling and control of dynamical systems,IEEE Trans. Neural Networks, vol. 9, pp. 756767, 1998.
[3] C. T. Lin and C. S. G. Lee, Neural Fuzzy Systems: A Neural-FuzzySynergism to IntelligentSystems. Englewood Cliffs, NJ: Prentice-Hall,1996.
[4] M. Mizumoto, The improvement of fuzzy control algorithm, part 4:( + ; ) -centroid algorithm, in Proc. Fuzzy Syst. Theory, vol. 9, 1990, p.9. (in Japanese).
[5] P. Z. Wang and H.-X. Li, Fuzzy Information Processing and Fuzzy Com-puters. New York: Science, 1997.
[6] T. Terano, K. Asai, and M. Sugeno, Fuzzy Systems Theory and Its Ap-plications. New York: Academic, 1992.
[7] M. Sugeno, Fuzzy Control (in Japanese). Tokyo, Japan: Japan Ind.Press, 1988.
[8] T. Takagi and M. Sugeno, Fuzzy identification of systems and its ap-plications to modeling and control, IEEE Trans. Syst., Man, Cybern.,vol. SMC-15, pp. 1116, 1985.
[9] H.-X. Li, Multifactorial functions in fuzzy sets theory, Fuzzy SetsSyst., vol. 35, pp. 6984, 1990.
[10] B. Kosko,Neural Networks and Fuzzy Systems. EnglewoodCliffs, NJ:Prentice-Hall, 1992.
[11] M. Brown and C. Harris, Neurofuzzy Adaptive Modeling and Con-trol. Englewood Cliffs, NJ: Prentice-Hall, 1994.[12] J.-S. R. Jang and C.-T. Sun, Functional equivalence between radial
basis function networks and fuzzy inference systems, IEEE Trans.Neural Networks, vol. 4, pp. 156159, 1993.
[13] D. Nauck, F. Klawonn, and R. Kruse, Neuro-Fuzzy Systems. NewYork: Wiley, 1997.
[14] L. H. Tsoukalas and R. E. Uhrig, Fuzzy and Neural Approaches in En-gineering. New York: Wiley, 1997.
[15] J.-S. R. Jang, C.-T. Sun, and E. Mizutani, Neuro-Fuzzy and Soft Com-puting. Englewood Cliffs, NJ: Prentice-Hall, 1997.
[16] J. J. Buckley, Y. Hayashi, and E. Czogala, On the equivalence of neuralnets and fuzzy expert systems, Fuzzy Sets Syst., vol. 53, pp. 129134,1993.
Hong-Xing Li received the B.S. degree from NankaiUniversity, China, in 1978 and the Ph.D. degree inmathematics from Beijing Normal University, China,in 1993.
He has been a Professor atBeijing Normal Univer-sitysince 1994.He wasa visiting Professorat Depart-ment of Computer Science and Engineering, WrightState University, Dayton, OH, during 1998 to 1999.His research interests are fuzzy control theory, adap-tive fuzzy control, neural networks, and mathemat-ical theory of knowledge representation.
Dr. Li is on the editorial board ofJournal of Fuzzy Systems and Mathematics,and Journal of Systems Engineering.
C. L. Philip Chen (S88M88SM94) received
the B.S.E.E. degree from NTIT, Taiwan, in 1979, theM.S. degree from the University of Michigan, AnnArbor, in 1985, and the Ph.D. degree from PurdueUniversity, West Lafayette, IN, in Dec. 1988.
In 1988 to 1989, he was a Visiting AssistantProfessor at the School of Engineering and Tech-nology, Purdue University, Indianapolis, IN. SinceSeptember 1989, he has been at the ComputerScience and Engineering Department, Wright StateUniversity, Dayton, OH, where he is currently an
Associate Professor. He is also a Visiting Research Scientist, at the MaterialsDirectorate, Wright Laboratory, Wright-Patterson Air Force Base. He has beena Senior Research Fellow sponsored by National Research Council, NationalAcademy of Sciences. He also worked as a Research Faculty Fellow at NASAGlenn Research Center during summer of 1998 and 1999. He was on sabbaticalleave to Purdue University and Case Western Reserve University for Fall1996 to Fall 1997. His current research interests and projects include neural
networks, fuzzy-neural systems, neuro-fuzzy systems, intelligent systems,robotics, and CAD/CAM.Dr. Chen was a Conference Cochairman of the International Conference
on Artificial Neural Networks in Engineering in 1995 and 1996, a TutorialChairman for the International Conference on Neural Networks in 1994, theConference Cochairman of the Adaptive Distributed Parallel Computing in1996, a Program Committee of OAI Neural Networks Symposium in 1995, theIEEE International Conference on Robotics and Automation in 1996, and theIntational Conference on Intelligent Robotics and Systems (IROS) in 1998 and1999. He is a member of Eta Kappa Nu. He is the founding faculty advisor ofthe IEEE Computer Society Student Chapter at Wright State University and arecipient of the 1997 College Research Excellent Faculty Award.