0% found this document useful (0 votes)
40 views

Fuzzy Neural Logic Network and Its Learning Algorithms

1) The document introduces the fuzzy neural logic network model, which generalizes three-valued neural logic networks to allow nodes to have fuzzy activations represented by ordered pairs of numbers between 0 and 1. 2) The network consists of input, hidden, and output nodes connected by weighted directed connections. Node activations propagate through the network according to a specified propagation rule. 3) The network can be trained using delta rule learning or one of two mathematical construction algorithms to learn from a set of training examples to perform classification or recommendation tasks.

Uploaded by

Rafael Belarmino
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
40 views

Fuzzy Neural Logic Network and Its Learning Algorithms

1) The document introduces the fuzzy neural logic network model, which generalizes three-valued neural logic networks to allow nodes to have fuzzy activations represented by ordered pairs of numbers between 0 and 1. 2) The network consists of input, hidden, and output nodes connected by weighted directed connections. Node activations propagate through the network according to a specified propagation rule. 3) The network can be trained using delta rule learning or one of two mathematical construction algorithms to learn from a set of training examples to perform classification or recommendation tasks.

Uploaded by

Rafael Belarmino
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 10

Fuzzy Neural Logic Network and its Learning Algorithms

Sing-Chai Chan Fui-Hoon Nah


Department of Information Systems & Computer Science
National University of Singapore
10 Kent Ridge Crescent
Singapore 05 11

with three columns or layers. They are the input layer, the
Abstract
hidden layer and the output layer. Each layer is made up of a set
This paper introduces the basic features of fuzzy neural of nodes. The nodes in the input layer are defined as input
logic network. Each fuzzy neural logic network model is trained nodes, the nodes in the hidden layer as hidden nodes and the
from a set of knowledge in the form of examples using one of the nodes in the output layer as output nodes. There is a directed arc
three learning algorithms introduced. These three learning from every node in the input layer to every node in the hidden
algorithms are the delta rule controlled learning algorithm and two layer, and from every node in the hidden layer to every node in
mathematical construction algorithms, namely, the local learning the output layer. Figure 1 shows a fuzzy neural logic network
method and the global learning method. Once the fuzzy neural model with m input nodes I, p hidden nodes H and n output
logic network model is constructed, it is ready to accept any nodes 0.
unknown input from the user. With a low percentage of
mismatched features, output solution can be obtained.

Introduction

Fuzzy neural logic network is both a generalization and an


extension of three-valued neural logic network [1][2][3]. In a
three-valued neural logic network, the activations of the nodes are
confined to ordered pairs (l,O), (0,l) and (0,O) where (1,O)
represents TRUE, (0,l) represents FALSE and (0,O) represents
UNKNOWN (in Kleene's logic) [12]. As human logic
reasoning usually involves a certain degree of bias, fuzziness or
uncertainty, the three-valued logic concept would not be Figure I :A Fuzzy Neural Logic Network
representative of the real world. As such, its representation is
extended to an ordered pair of real, non-negative numbers (a,b)
where a and b represent the amount of evidence for and against a The input nodes are used to accept inputs in the form of
particular condition respectively. This fuzzy-valued logic concept patterns or feature extractions from an application problem. The
[4][5][6][7][11J is applied in fuzzy neural logic network. hidden nodes receive contributions from the input nodes through
the connection weights between the input and hidden layers.
Each fuzzy neural logic network model can be trained by
Once the contributions at a hidden node reach a preset threshold
incremental approximation using the delta rule or by mathematical
value, the hidden node is in an excitatory state and produces an
construction [lo].
output solution at the output layer. On the other hand, if the
contributions at a hidden node fall below the threshold value, the
Network Structure
hidden node is in an inhibitory state and hence, does not yield
A fuzzy neural logic network model is a directed graph any solution.

476
.OO 0 1991 IEEE
0073-1129/91/0000/0476$01
Fuzzy Activation having an activation of (1,O). Otherwise, the hidden node is in
an inhibitory state, i.e., with an activation of (0,l) or (0,O).
Each input, hidden and output node represents a
condition. The activation of a node is denoted by an ordered pair Each output node 0 represents a possible
of real, non-negative numbers (a,b). These numbers are recommendation action. The activation value (a,b) of the output
normalized in such a way that the following constraints are node represents the eagerness with which an action is
satisfied: recommended to be taken (indicated by the quantity "a") or not to
OIall be taken (indicated by the quantity "b"). Since (1-a-b) represents
OIbSl the uncertainty, a large (1-a-b) means that the system does not
OIa+b<l know whether to recommend for or against the action.
The interpretation of these numbers is as follows
There is a connecting arrow between every node in the
M[51[61[71[111: input layer to every node in the hidden layer and from every node
in the hidden layer to every node in the output layer. Each
1. The quantity 'la'' denotes the amount of evidence
for the condition represented by the node, connecting arrow is assigned an ordered pair of real numbers
(a$) as its weight. Unlike the activations, these weights are not
2. The quantity "b" denotes the amount of evidence
against the condition represented by the node, normalized and negative values are allowed.
3. The quantity "c" expresses the lack of evidence
Propagation Rule
regarding the condition.
The network is activated by assigning value (a,b) to each
Each input node I to a condition' If an input of the input ,,odes where a and b represent the amount of
attribute can take Only two possible it is represented by evidence for and against the condition represented by the input
one input node. On the other hand, if an input attribute is
node respectively. These activation values then propagate along
classified into more than two categories, say n categories, then it
the network using the rule of propagation defined as follows:
is represented by n input nodes where each input node represents
the amount of evidence for and against a category. Suppose there are m incoming arrows connecting m
nodes Pi whose activations are (ai,bi) to a node R where
The activations of these input nodes are assigned by the i=1,2,...,m. Let the connecting weight between each Pi node
user. For example, if 100 customers were asked about the price
and R be (ai,pi) where a i and pi are the weights associated with
of a product, 60 customers feel that the price is high, 20 the evidence for (indicated by the quantity "ai") and against
customers think that the price is low and 20 customers do not
(indicated by the quantity "bi") the condition represented by node
give any comments, then the activation of the input node
Pi respectively. The following figure illustrates this portion of
representing the condition 'price of product is high' is assigned a
the network.
value of (0.6,0.2).

Each hidden node H represents a logical operation.


Attached to each hidden node is a threshold value 8 and an
activation value (a,b). This activation value is computed using a
propagation rule (discussed in the next section). The net
excitatory input at a hidden node indicates the level of agreement
between the given evidence (i.e., input information) and the
attribute values associated with the hidden node. On the other
hand, the net inhibitory input indicates the level of disagreement
between the given evidence (i.e., input information) and the
attribute values associated with the hidden node. When the net
Figure 2 :Propagation of Strengths
excitatory input at a hidden node is greater than or equal to its net
inhibitory input by 8, the threshold operation in the propagation The activation of node R is calculated by the following
rule causes the hidden node to be in an excitatory state, i.e., steps:

411
For each value of i, take the inner products of drawn from every node in the first column to every node in the
(ai,bi) and (ai.pi) to form the ordered pair (t,f) = second column, and from every node in the second column to
(aiai&iPi). every node in the last column.

Make components of the ordered pair positive. Once the network structure is constructed, it is trained
using one of the following three techniques:
If both t and f a r e positive, set ( r i f i ) = (t,f).
1. Iterative learning using delta rule
If both t and f a r e negative, set (ti&) = (-f,-t).
2. Local learning by construction
If only t is negative, set (ti&) = (0,f-t).
3. Global learning by construction
If only f is negative, set ( t i f i ) = (t-f,O).
In the iterative learning method, weights are allowed to
The net excitatory input at node R is given by
m vary while the system is fed with examples. In the construction
T = Cti. methods (local and global learning), weights are assigned based
i=l on the given set of training examples. However, weights that are
assigned using the construction methods are so precise that any
The net inhibitory input at node R is given by
m deviation in inputs cannot be tolerated. Hence, a refinement
F= 5 i . algorithm is introduced to solve this problem.
i=l
It is assumed that the evidence for and against a condition
Perform a strong threshold calculation. The given
are weighted equally in these three learning algorithms. In other
threshold for node R is denoted by e r . The
words, for all the weights (a,p)that connect nodes in the first
strong threshold calculation is: layer to nodes in the second layer, a=P.
If T-F2 er. set pr = 1 and qr = 0.
If T-FI-Br, set Pr = 0 and qr = 1. Iterative Learning using Delta Rule
If IT-FI<Or, set Pr = 0 and qr = 0.
Like any other conventional neural network models, a
The activation of node R is given by (Ppqr). fuzzy neural logic network model can be trained using an iterative
learning algorithm. Being enhanced perceptrons, it uses a
The threshold e r is set to 1 in our approach.
controlled learning algorithm modified from the delta rule [ 111.
The same propagation process is repeated until no further During the process of delta rule controlled learning, the
change of value is needed. The neural network has then reached network is repeatedly presented with a set of initial stimulus
a "stable" state. At this point, the values of the output nodes are patterns or input vectors, with a corresponding set of outputs.
the output values of the given input values. After the learning process, each hidden node becomes sensitive to
one initial stimulus pattern or input vector, and the rest of the
Learning
hidden nodes are representatives for other stimulus patterns or

A fuzzy neural logic network model is constructed input vectors'


directly from the binary representation of a training set. The first The following are the steps for the delta rule controlled
step in the creation of a fuzzy neural logic network model is the
learning algorithm:
construction of its network structure.
Step 1: Construction of Network Structure. Construct
A fuzzy neural logic network model is a directed graph
the fuzzy neural logic network structure using the training
with three columns. The first column has m nodes, the second
examples.
column has k nodes and the third column has n nodes, where m
refers to the number of terms in the input vector, k refers to the Step 2: Initial Weights Assignment to Edges Joining
number of training examples and n refers to the number of terms Nodes in Columns 1 & 2. Let R = [(l,O), (O,l), (0,O)jand
in the output vector. The nodes in the first column are defined as S = [(l,O), (0,l)j. Let I1,12, ...,Ik be k (distinct) members of
input nodes, the nodes in the second column as hidden nodes and Rm such that Ii (i=l,2,...,k) refers to the input vector of the ith
the nodes in the third column as output nodes. A directed arc is training example and m refers to the number of terms in each

478
input vector. Similarly, let 0 1 , 0 2 ,...,Ok be k (not necessary CASE 1:
distinct) members of Sn such that Oj (i=1,2, ...,k) corresponds to If i=r,
the output vector of the ith training example and n refers to the a = a + q [1-(c-d)] (a-b)
number of terms in each output vector. p=P + q [1-(c-d)] (a-b)

Take one node say the jth node of column 1 and take any CASE 2:
node say the ith node of column 2. Attach the edge joining theJZh If i#t,
node of column 1 to the ith node of column 2 by the ordered pair a = a + q min(0,O-(c-d)) (a-b)
(a$),where a and are real numbers. p = P + q min(0,O-(c-d)) (a-b)
where
Let (a,b) be the jth term of the input vector Ij. = a small positive constant that determines the
q
learning rate (O.Ol<q<O.l).
CASE 1:
If (a,b) = (O,O), assign For each fixed t , modify the weights for each pair of
(a,P)= (0,O). nodes j=1,2 ,...,m; i=1,2 ,...,k.

CASE 2: Step 5: Repetition of Step 4 until Convergence of


If (a,b) f (O,O), assign Weights. One cycle of training is defined as the repetition of
(a$)= small, random and reflexive (a=P)initial step 4 for t=l,t=2, ...,t=k. This entire cycle is iterated until all
weights. new weights are equal to old weights, i.e., until all weights
converge.
Step 3: Weights Assignment to Edges Joining Nodes
in Columns 2 & 3. Take one node say the ith node of column
A fuzzy neural logic network model that is trained using
2 and one node say the jth node of column 3. Attach the edge
the delta rule controlled learning algorithm can tolerate error to a
joining these two nodes by the ordered pair (a'$'),where a',
small degree, depending on the value of the gain term q. The
p' are real numbers obtained as follows:
larger the gain term, the greater the degree of error tolerance.
Let (a',b) be the jth term of the output vector Oi. However, too large a gain term may result in oscillations such
that the weights never converge.
CASE 1:
If (a',b') = (l,O), assign Nevertheless, there are several drawbacks in this iterative
learning method. The final weights obtained are very sensitive to
(a'$')
= (1,O).
the initial weights. Hence, the initial weights must be kept as
CASE 2: small as possible. Moreover, this learning method cannot
If (a',b) = (O,l), assign identify the relative importance of the attributes. Instead, it
(a'$') = (-1,O). assumes that all input attributes are equally important. Lastly, the
time taken and the amount of computations required to train the
Do this to every pair of nodes between column 2 and
network is very large compared to the construction methods
column 3.
discussed in the next two sections.
Step 4: Training the First Layer of Connection
Local Learning by Construction
Weights. Take a fixed vector It (t=1,2, ...,k) and assign its
values to input nodes in column 1. Then compute the values of The local learning by construction method is a self-
the nodes in column 2. learning mechanism for fuzzy neural logic network. It is

Let (a$) be the weight associated with the edge joining performed by means of "learning without teacher". By weighing
the input attributes of each training example equally, the learning
the jthnode of column 1 to the ith node of column 2. Let (a,b) be
the jthvalue of Ii. Let (c,d) be the propagated value of the ith of each initial stimulus pattern is independent of the rest of the
stimulus patterns. In other words, each hidden node identifies a
node of column 2.
stimulus pattern by weighing each of the features to be of equal
Modify the values (a$)according to the following cases: importance.

419
During the process of learning, the network is presented Step 3: Weights Assignment to Edges Joining Nodes
with a set of initial input and output patterns. After the self- in Columns 2 & 3. Take one node say the ith node of column
learning process, each hidden node becomes very sensitive to one 2 and one node say the jth node of column 3. Attach the edge
initial stimulus pattern. Each hidden node has the ability to joining these two nodes by the ordered pair (a'$'),where a',
recognize a pattern and has become selectively responsive only to p' are real numbers obtained as follows:
that stimulus pattern. In fact, each initial stimulus pattern has
become elicit to one output through a representative hidden node. Let (a',b') be thejth term of the output vector Oi.

The following are the steps for this local learning by CASE 1:
construction algorithm: If (a',b') = (l,O), assign
= (1,O).
(a'$')
Step 1: Construction of Network Structure. Construct
the fuzzy neural logic network structure using the training CASE 2:
examples. If (a',b') = (O,l), assign
(a'J3')= (-1,O).
Step 2: Weights Assignment to Edges Joining Nodes
in Columns 1 & 2. Let R = ((l,O), (O,l), (0,O)) and S = Do this to every pair of nodes between column 2 and
((l,O), (0,l)). Let 11,12,...,Ik be k (distinct) members of Rm column 3.
such that Ii (i=1,2, ...,k) refers to the input vector of the ith
Step 4: Modification of Weights to Achieve Desired
training example and m refers to the number of terms in each
Error Tolerance. Unlike the Hopfield network, a fuzzy neural
input vector. Similarly, let 0 1 , 0 2 ,...,Ok be k (not necessary
logic network model constructed using the above construction
distinct) members of Sn such that Oi (i=1,2,...,k) corresponds to
algorithm can recognize all the initial stimulus patterns. The
the output vector of the ith training example and n refers to the
mapping from the input vector set (11,12,...,1k) to the output
number of terms in each output vector.
vector set ( 0 1 , 0 2 ,...,O k ) can be confirmed by direct
Take one node say the jthnode of column 1 and take any verification. However, the weights obtained are so precise that
node say the ith node of column 2. Attach the edge joining thejth the network cannot tolerate any minor deviation in its inputs. This
node of column 1 to the ith node of column 2 by the ordered pair problem is solved by applying the following refinement
(a$), where a and P are real numbers. algorithm:

Let (a,b) be thejth term of the input vector Ii.


Modify the weights (a$)associated with the edges
CASE 1: joining nodes in layer 1 to nodes in layer 2 such that:
If (a,b) = (l,O), assign
(a$)= (
1 1
c c1
7
a=- a
1 - &
where c is the number of terms in Ii that is either (1,O) or p=-- P
(0,l). Hence c l m . 1-E
where
CASE 2:
E = desireddegree of error tolerance required for the
If (a,b) = (O,l), assign
1 1 network ( 0 2 x 1 ) .
(a,P)= ( - c, c)
-

where c has the same meaning as in case 1. Step 5: Relearning using Delta Rule. After the weights
are modified, the network model may no longer mapped the input
CASE 3: vector set (I1,12,..., Ik) to the output vector set ( 0 1 , 0 2,...,O k ) .
If (a,b) = (O,O), assign Hence, the delta rule controlled learning algorithm (i.e., steps 4
(a$)= (0,O). and 5 of the iterative learning using delta rule method given
earlier) is applied to relearn the pattern mappings.
This operation will be performed on every pair of nodes
between column 1 and column 2. Although the local learning by construction cum

480
refinement method does not identify the relative importance of the k
attributes, it is an improvement over the iterative learning method Let N be the size of U {Oi)
because the time taken to train the network has been significantly i= 1
reduced. Furthermore, the amount of computations required for For each term (say p h term) of the input vector
both the construction and refinement algorithms are very much Initialize S(1,o) and S(o,1) to be empty sets
less than that required for iterative learning. For each training example (say qth training example)
Let (a,b) be the pth term of the input vector Iq
An advantage of using this method is the flexibility in If (ab) = (l,O), then
setting the error tolerance requirement. Not only does the S(1,O) = S(1,O) U (Oq)
network matches the given pair of pattern sets {11,12,...,1k) and If (a,b) = (O,l), then
{ 0 1 , 0 2 ,...,0 k ) correctly, it also tolerates certain degree of error S(0,l) = S(0,l) U (Oq)
depending on the input vector set {11,12,..&) as well as the Assign T(1,o) (p) = N + 1 - size of S(1,o)
desired error tolerance requirement. Assign T(o,l) @) = N + 1 - size of S(o,1)

Step 3: Association between Features' Importance and


Global Learning by Construction
Connection Weights. This algorithm associates the
In real situations, not all attributes are of equal importance of each feature with its corresponding connection
importance. A more powerful learning algorithm that can be weight:
trained within a short time and yet can identify the relative
For each term (say pth term) of the input vector of each training
importance of the factors is required. As such, a global learning
example (say 4th training example)
by construction cum refinement method is introduced.
Let (a,b) be the pth term of the input vector Iq
This new learning method is an improvement over the If (a,b) = (l,O), then
local learning by construction cum refinement method because it wpq = T( 1,O) (PI
weighs input attributes according to their importance. The only If (a,b) = (O,l), then
difference between the two methods is in the construction wpq = T(0,l) @)
algorithm used. Like the local learning by construction cum If (a,b) = (O,O), then
refinement method, this method consists of two steps -- wpq= 0
construction followed by refinement.
Step 4: Weights Assignment to Edges Joining Nodes
The following are the steps for this global learning by in Columns 1 & 2. Take one node say the jth node of column
construction cum refinement method: 2 and take any node say the jth node of column 1 . Attach the
edge joining the jth node of column 1 to the ith node of column 2
Step 1: Construction of Network Structure. Construct by the ordered pair (a$), where a and p are real numbers.
the fuzzy neural logic network structure using the training
examples. Let (a,b) be the jfh term of the input vector Ii.

CASE 1:
Step 2: Determination of Features' Importance. Let R =
If (a,b) = (l,O), assign
{(l,O), (O,l), (0,O)) and S = {(l,O), (0,l)). Let 11,12,...,Ik be
k (distinct) members of Rm such that Ii (i=1,2,...,k) refers to the
input vector of the ith training example and m refers to the
p=1 p=l
number of terms in each input vector. Similarly, let
0 1 , 0 2 ,...,0 k be k (not necessary distinct) members of Sn such
CASE 2:
that Oi (i=1,2,...,k) corresponds to the output vector of the ith If (a,b) = (O,l), assign
training example and n refers to the number of terms in each
(%PI =( -wii ' 1- m
output vector. The following algorithm determines the
importance of each feature:

48 1
CASE 3: refinement process. In addition, the refinement algorithm
If (a,b) = (O,O), assign provides the users the flexibility to set the error tolerance
(a$) = (0,O). requirement. Hence, this method is definitely more superior than
the delta rule controlled learning method.
This operation will be performed on every pair of nodes
between column 1 and column 2. The third method, the global learning by construction
method, is an improvement over the local learning by
Step 5: Weights Assignment to Edges Joining Nodes construction method. This method retains all the strong points
in Columns 2 & 3. Take one node say the ith node of column and overcomes the weakness of the second method. By
2 and one node say the jth node of column 3. Attach the edge weighing input attributes according to their importance, more
joining these two nodes by the ordered pair (a'$'),where a',
realistic models are simulated.
p' are real numbers obtained as follows:
To summarize, the construction cum refinement methods
Let (a',b) be the jth term of the output vector Oi.
are more efficient than the iterative method. The local learning by
construction method is appropriate when input attributes are to be
CASE 1:
weighted equally. The global learning by construction method is
If (a',b') = (l,O), assign
used when the relative importance of the attributes are to be taken
(a'$')= (1,O).
into consideration.
CASE 2:
Knowledge Acquisition
If (a',b) = (O,l), assign
(a'$')= (-1,O). A fuzzy neural logic network model is set up using a set
of knowledge in the form of examples. This knowledge can be
Do this to every pair of nodes between column 2 and
acquired from many sources, including textbooks, documents,
column 3.
human experts, our own experience and generalized observations
Step 6: Error Tolerance Refinement. Although a fuzzy Of the environment.
neural logic network model constructed using the above
A training example indicates the conditions (or attribute
construction algorithm can recognize all the initial stimulus
values) to satisfy a conclusion. Before a set of training examples
patterns, the weights obtained are such that even minor error are used to set up the model, they are converted into the binary
cannot be tolerated. Hence, the refinement algorithm (i.e., steps
form with (l,o), (o,l) and (o,o) A training example that
Of the local learning by consmction method given
is in its binary form is made up of two components, namely, the
is applied to enable the network to certain degree Of error. input vector and the output vector. The input vector corresponds
Comparison of the Learning Algorithms to the conditions or attribute values to be satisfied and the output
vector corresponds to the conclusion. The input vector is made
Three learning algorithms are proposed to train the fuzzy UP of (1,0), (0,l) and (0,O) terms and the output vector is made
neural logic network. up of (1,O) and (0,l) terms. The number of terms in.the input
vector is determined as follows:
The first method, the delta rule iterative learning method,
weighs all input attributes equally. The major drawback of this For each input attribute
method is the great amount of time required to train the network. if it can take only two possible values, it is
represented by only one ordered pair term.
The second method is named 'local learning by
if it can take more than two possible values, say j
construction' because each training example is learned
values, it is represented by j ordered pair terms.
independently of the others. This is possible because the input
attributes of each training example are given equal weights. Suppose there are k input attributes in the training set.
Since the network is constructed by assignment of weights, it can Let r1,r2,...,rk be the number of ordered pair terms that
be set up in a relatively short time. Unlike the iterative learning are required to represent the lSt, 2"d ,..., kth input
method, only a few computations are required, if at all, in the attributes respectively.

482
Hence, the number of terms in the input vector, m = B. Conversion of outDut attributes' values or conclusions to
rl+r2+ ...+rk. binarv outuut vector

The number of terms in the output vector is determined as If there are only two possible conclusions
follows: then the first conclusion is represented by (1,O) and the
second conclusion by (0,l).
If there are only two possible conclusions, they are If there are more than two possible conclusions, say
represented by only one term in the output vector. n conclusions,
If there are more than two possible conclusions, say n then the qth value is represented by n ordered pair
conclusions, they are represented by n terms in terms such that the qth term is (1,O) and the other
the output vector. terms are (O,l)s, i.e.,
the first conclusion is represented by n terms of
The conversion procedure of a training example from its
natural language form to its binary form is outlined as follows: (1 ,O)(O, 1) ( O , 1)...,
the second conclusion is represented by n terms of
A. (0,1)(I ,O)(O, I)...,

The ith input attribute is represented by the (rl+...+ri.l


the nth conclusion is represented by n terms of
+l)th to the (r,+ ...+ri)th terms of the input vector, where ro = 0
and ri refers to the number of ordered pair terms that are required
(0>1)(0,1)...(I m.
to represent the ih input attribute. Note: The ranking of the values and conclusions are arbitrary.

For each input attribute


The following example illustrates the conversion process.
if it can take only two possible values
then the first value is represented by (1,O) and the In a product data base [8] of a media strategy selection
second value by (0,l) system, products are classified into four categories by product
if it can take more than two possible values, say j type. These four categories are:
values, o Convenience
then the q* value is represented by j ordered pair temis o Impulsive
such that the qth term is (1,O) and the other tenns o Shopping
are (O,O)s, i.e., o Speciality
the first value is represented by j terms of
..,
( 1 ,~)(O,O)(O,O).
Product type can be defined by four key influencing
the second value is represented by j terms of variables:
(0,0)(1,O)(O>O)..., o Targetmarket
o Product involvement
o Product decision making
the jth value is represented by j terms of
o Brand loyalty
(O,O)(O,O) ...( 1,O).
In other words, four input attributes, namely, 'target
Hence, given the ith input attribute's value of a training
example, the (rl+...+ri.l+l)th to the (rl+ ...+ri)th terms of the market', 'product involvement', 'product decision making' and
'brand loyalty', are involved to categorize 'type of product' as
input vector are assigned the binary representation of the value.
'convenience', 'impulsive', 'shopping' or 'speciality'. If 'target
However, if the ith input attribute is irrelevant with respect to the
market' is 'selective', 'product involvement' is 'high', 'product
training example, i.e., the ith input attribute's value is not
indicated in the training example, the (rl+...+ri.l+l)th to the decision making' is 'complex' and 'brand loyalty' is 'high', 'type
(rl+ ...+ri)th terms of the input vector are assigned ordered pairs of product' is concluded as 'speciality'. However, if 'product
of (0,O)s. involvement' is 'high', 'product decision making' is 'complex'
and 'brand loyalty' is 'low', 'type of product' is categorized as

483
'shopping' without further information on 'target market'. In Hence, the binary representation of the training examples
other words, 'target market' is irrelevant with respect to the in table 1 is as follows:
training example. On the other hand, if 'target market' is
'general', 'product involvement' is 'low', 'product decision
making' is 'limited' and 'brand loyalty' is 'low', then 'type of
product' is 'impulsive'. Lastly, 'product decision making' is
'convenience' if 'target market' is 'general', 'product
involvement' is 'low', 'product decision making' is 'habitual' A fuzzy neural logic network model for the above set of
and 'brand loyalty' is 'high. training examples are constructed using the global learning by
construction method with a desired error tolerance of 10%. This
Table 1 illustrates these training examples in tabular form.
fuzzy neural logic network model is illustrated in figure 3.
Table I :A Sample Set of Training Examples Target mark&
(Seleaive,Ceneral)
Produd t y p
Product
Target Product decision Brand Product
market + involvement +makine + lovaltv --> twe \\ \034

1. Selective+ High + Complex+ High --> Speciality


2. High + Complex+ Low --> Shopping
3. General + Low + Limited + Low --> Impulsive
4. General + Low + Habitual+ High --> Convenience

These set of training examples are converted to their


binary form before they are used to train the fuzzy neural logic
network model.

1. Since the input attribute 'target market' can take


only two possible values, we represent the value
'selective' by (1,O) and the value 'general' by
(0,l).
2. Similarly, the input attributes 'product
Figure 3 :Fuzzy Neural Logic Network Model for Training
involvement' and 'brand loyalty' can take only
Examples in Table I
two possible values 'high' and 'low'. Hence, the
value 'high' is represented by (1,O) and the value Note: In figure 3, the connection weights between the input and
'low' by (0,l). hidden layers are represented using single weights instead
3. The input attribute 'product decision making' can of ordered pairs because the weights are reflexive. In
take three possible values 'complex', 'limited and other words, a single weight a is to be interpreted as an
'habitual'. As such, three terms are required in orderedpair (a,a).
the input vector to represent the attribute. The
value 'complex' is represented by (1 ,0)(0,0)(0,0), Inferencing
the value 'limited' by (O,O)( 1,0)(0,0) and the
Once a fuzzy neural logic network model is constructed, it
value 'habitual' by (O,O)(O,O)(l ,O).
is ready to accept any unknown input. The input to the model is
4. There are four possible conclusions for 'type of
assigned by the user through a question and answer session. The
product', namely, 'speciality', 'shopping',
activation of the input nodes are propagated to the output nodes
'impulsive' and 'convenience'. 'Speciality'
via the hidden nodes. Output solution can be obtained at the
product is represented by (l,O)(O,l)(O,l)(O,l),
output layer with a low percentage of mismatched features.
'shopping' product by (0,1)( 1,0)(0,1)(0, l ) ,
'impulsive' product by (O,l)(O,l)( 1,O)(O,l) and Suppose there is evidence (as shown in table 2) on the
'convenience' product by (O,l)(O,l)(O,l)( 1,O). following characteristics of a product: target market, product

484
involvement, product decision making and brand loyalty. Given References
the evidence, the product type is to be determined.
1. Chan S.C., Loe K.F., Teh H.H. (1987), "Modelling Intelligence
Table 2 : Evidence on the Characteristicsof a Product
Using Neural Logic Networks", Department of Information Systems
CHARACTERISTICS EVIDENCE
Target market 0% selective100% general 0% uncertainty and Computer Science, National University of Singapore.
product involvement 3% high 95% low 2%unceaainty 2. Chan S.C., Hsu L.S., Brody S., Teh H.H. (1988), "On Neural
Product decision making 10% complex 85% limited 0% habitual 5% uncertainty
0% uncertainty
Logic Networks", Neural Network Journal, 1 (Supplement 1): 428.
Brand loyalty 0% high 100% low
3. Chan S.C., Hsu L.S., Brody S., Teh H.H. (1989). "Neural Three-
Hence, the input nodes take the following values: valued-logic Networks", Proceedings of the International Joint
Conference on Neural Networks, 2: 594.
(0,1)(0.03,0.95)(0.1,0.85)(0.85,0.1)(0,0.95)(0,1).
4. Hsu L.S., Teh H.H., Chan S.C., Loe K.F. (1989). "Fuzzy Decision
The output solution that is produced at the output layer is Making based on Neural Logic Networks", Proceedings of the Inter
(O,l)(O,l)( l,O)(O,l). Since this is the binary representation for Faculty Seminar on Neuronet Computing, Technical Report
'impulsive', the conclusion is 'impulsive product type'. TRA6/89, Department of Information Systems and Computer
Science, National University of Singapore, June.
Conclusion 5. Hsu L.S., Teh H.H., Chan S.C., Loe K.F. (1989), "NELONET
based Decision Making", Department of Information Systems and
In this paper, a new class of neural networks called
Computer Science, National University of Singapore.
'Fuzzy Neural Logic Network' has been proposed. It uses
6. Hsu L.S., Teh H.H., Chan S.C., Loe K.F. (1990). "Imprecise
fuzzy-valued logic to handle fuzziness, bias and uncertainty.
Reasoning using Neural Networks", Proceedings of the Twenty-
As fuzzy neural logic network is massively parallel, Third Annual Hawaii International Conference on System Sciences,
decisions can be made at high speed. The network can recognize 4: 363-368.
all the initial stimulus patterns even if these patterns are very 7. Hsu L.S.;Teh H.H., Chan S.C., Loe K.F. (1990). "Fuzzy Logic in
similar. Furthermore, the network is able to draw conclusions Connectionists' Expert Systems", International Joint Conference on
despite minor variations in its inputs. Neural Networks, 2: 599-602.
8. Mockler R.J. (1989), "Knowledge-based Systems for Management
Only the basic features of fuzzy neural logic network are Decisions", Prentice Hall.
introduced in this paper. In fact, research on fuzzy neural logic 9. Nah F.H. (1990), "Fuzzy Neural Logic Network with Application
network has progressed more than what are presented. to Decision Support Systems", Master thesis, Department of
Information Systems and Computer Science, National University of
Multi-level fuzzy neural logic network has been designed
Singapore, in preparation.
[9] to solve complex problems. With this design, multi-level
10. Teh H.H., Yu W.C.P. (1988). "A Controlled Learning Environment
decision making can be supported by allowing the output of a
of Enhanced Perceptron", IEEE Proceedings, Future Trend in
lower level model to become the input of a higher level model.
Distributed Computing Systems, Hong Kong.
This is consistent with human's structured way of thinking.
11. Teh H.H., Chan S.C., Hsu L.S., Loe K.F. (1989). "Probabilistic
Lastly, a decision support system based on multi-level Neural-Logic Networks", Proceedings of the International Joint
fuzzy neural logic network has been implemented on the SUN Conference on Neural Netwofks, 2: 600.
workstation using the 'C' programming language. 12. Turner R. (1984). "Logics for Artificial Intelligence", Ellis
Honvood, Chichester.

485

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy