CA2077626A1 - Translation of a neural network into a rule-based expert system - Google Patents

Translation of a neural network into a rule-based expert system

Info

Publication number
CA2077626A1
CA2077626A1 CA002077626A CA2077626A CA2077626A1 CA 2077626 A1 CA2077626 A1 CA 2077626A1 CA 002077626 A CA002077626 A CA 002077626A CA 2077626 A CA2077626 A CA 2077626A CA 2077626 A1 CA2077626 A1 CA 2077626A1
Authority
CA
Canada
Prior art keywords
layer
hidden
processing elements
rules
neural network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
CA002077626A
Other languages
French (fr)
Inventor
Li Min Fu
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Florida
Original Assignee
Li Min Fu
University Of Florida
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Li Min Fu, University Of Florida filed Critical Li Min Fu
Publication of CA2077626A1 publication Critical patent/CA2077626A1/en
Abandoned legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/042Knowledge-based neural networks; Logical representations of neural networks

Abstract

ABSTRACT OF THE DISCLOSURE
A rule-based expert system is generated from a neural network. The neural network is trained in such a way as to avoid redundancy and to select input weights to the various processing elements in such a way as to nullify the input weights which have smaller absolute values. The neural network is translated into a set of rules by a heuristic search technique.
Additionally, the translation distinguishes between positive and negative attributes for efficiency and can adequately explore rule size exponential with a given parameter. Both explicit and implicit knowledge of adapted neural networks are decoded and represented as if - then rules.

Description

TRANSLATION OF A NEURAL NETWORK
INTO A RULE~BASE EXPERT SYSTEM
CopYriqht Notic~
A portion of the disclosure of this patent document contains material which is subject to copyright protection. The copyright owner has no objection to the facsimile reproduction by anyone of the patent document or the patent disclosure, as it appears in the Patent and Trademark Office patent file or records, but otherwise reserves all copyright rights whatsoever.
Backqround of the Invention This invention relates to expert systems. More specifically, this invention relates to translation of a neural network into a rule-based expert system.
Further, thls invention relates to a neural network which is trained as to better allow its translation.
The use of computers to store knowledge has been common for quite some time. In addition to simply storing data or information, computers have been used for drawing conclusions. For example, a computer may be used to predict the weather.
There has been a continuing need to develop computers or data processing systems which can interpret data or otherwise draw conclusions from data.
The data processing system might draw the conclusions based upon numerous sample cases which have previously been supplied to it. For example, one might input current weather conditions and the data processing system or computer might predict the weather based upon the known weather patterns which occurred in the past.
Instead of simply providing previous occurrences, one may want to supply rules to a computer or data processing system. The rules would tell the computer that when the facts corresponding to the premise of a .
.

~ ~ '1 r? ~

rule occur, then the conclusion of the rule is true (or is most likely true).
Data processing systems have been developed which include neural networks. Such neural networks include layers of processing elements sometimes called neurons because they are designed to roughly simulate the neurons in a human brain. Various electrical circuits may be used to rea]ize a processing element having an output which is a complex function of a series of inputs. In addition to the hardware neural networks, neural networks have been provided by software components in a digital computer system. The software components in such a software neural network simulate the hardware components of a hardware neural network.
As used herein, a neural network will include both hardware and software neural networks and exclude neurons and connections in the brain of a human or animal. In other words, a neural network as used herein is artificial (non-human, non-animal).
Data processing systems have been used to establish, possibly through use of a machine learning program, rule-based expert systems. Such expert systems include a rule memory, often called a rule base, and an inference engine. Such systems determine which rules are applicable and then fire or execute the applicable rules in order to draw conclusions. Such systems have become increasingly popular because of many of their features. Generally, an expert system holding a given amount of knowledge would run significantly faster than a neural network holding the same amount of knowledge. Further, an expert system is structured in such a way that a human expert can follow the process by which the computer expert system reached its conclusion. The human expert can see which rules ,, ~ ~ `
`

have been executed by the expert system and readily evaluate whether the expert system i5 functioning correctly. In contrast, it is very dif~icult for a human expert to follow the complex process used by a neural network, usually composed of one or more neural chips, because of the structure. A rule-based expert system is also advantageous relative to a neural network in that it is much easier usually to add a new rule to a rule based expert system than it is to add further information to a neural network.
Although the features of a rule-based expert system as discussed above help explain the increasing popularity of such systems, such systems also have several disadvantages as compared to neural networks.
Such expert systems search directly over the primary attribute space and this limits their search width.
The primary attribute space refers to characteristics which are in the premises of the rules. If there are 100 different attributes or conditions, it may be difficult or impossible to search all possible combinations of the 100 attributes. In contrast, a neural network allows for an exponential search width because of its hidden layered structure. Thus, a neural network provides for an adequate search over practical problem domains. A fllrther advantage of neural networks over expert systems is that neural networks may handle multivariate analysis, whereas rule-based expert systems are usually more limited in this aspect due to the relatively limited search width.
A rule-based expert system can be established by a machine learning program in which one supplies specific cases such that these expert systems lack statistical generalization. In contrast, a neural network can handle statistical information relatively well and f imply conclusions which are not addressed by a particular case supplied to i~. A further advant~g~ of neural networks is that they handle noisy data significantly better than an expert system as established by a symbolic learning program.
OBJECTS AND SUMMARY OF THE INVENTION
Accordingly, it is a primary object of the present invention to provide a rule-based expert system which avoids or minimizes many of the disadvantages previously associated with such systems.
A further object of the present invention is to provide a rule-based expert system which is established in such a way as to have many of the advantages of neural networks including handling multivariat~
analysis, the ability to be constructed using data which is noisy, and the ability to incorporate implications drawn from the raw data which is provided, which implications reach conclusions in cases which were not part of the raw data.
Yet another object of the present invention is to provide for translation of knowledge from a neural network to a rule-based expert system having equivalent performance to the neural network.
A still further object of the present invention is to provide a neural network which is specifically constructed as to allow transfer of its knowledge into a rule-based expert system.
The above and other objects of the present invention which will be more apparent as the description proceeds are realized by a data processing system. The system includes a translating means for translating knowledge in a neural network having an input layer of input units and different layers of processing elements including an output layer of output , ~' .~,-. ' .

~ J~3~,J

processing elements and at least orle hidden layer of hidden processing elements. The translating means translates knowledge in the ou~er layer and ea~h hidden layer of the neural network into a layer se~ of rules.
A rewriting means rewrites rule~ from the layer set of rules by reformulating rules from one layer in terms of rules of another layer to eliminate undefined hidden concepts and thereby generate a rewritten set oP rules.
The rewritten set of rules have an equivalent performance to the neural network. A memory recei~es and stores the rewritten set of rules. The data processing system ma~ further include the neural network and the translation means and rewriting means are operably connected to the neural network. The neural network may be a hardware network with the input units, output processing elements, and hidden processing elements all being hardware components. The processing system may include a CPU and a memory operably connected to the CPU and having the translating means and rewriting means disposed in the memory. As an alternative to the hardware network, the neural network may be disposed in the memory and the input units, output processing elements, and hidden processing elements would all be software components.
In another embodiment, the translation means and rewriting means may be part of one or more of the output processing elements and hidden processing elements. In that case, the memory ma~ be within at least one of the output processing elements and the hidden processing elements.
The translating means translates knowledge from the neural network while distinguishing between positive attributes, which link to a concept corresponding to an output of one of the processing elements by a positive ~ J/)~

weight and negative a~tributes, which link to an output concept of one of the processing elements by a negative weight. The translating means translates by presenting each processing element with positive input combinations corresponding to positive attributes up to k for that processing element to develop a set PA of all combinations of at most k positive attributes/ each of which confirms one of the concepts if all negative attributes are absent. The translating means then determines which of the input combinations in PA
confirm one of the concepts when combined with the absence of some or no negative attributes to develop a set PAN. The translating means then presenks each processing element with negative input combinations up to k corresponding to negative attributes for that processing element to develop a set NA of all combinations of at most k negative attributes each of which disconfirms one of the concepts if all positive attributes are absent. The translating means then determines which of the input combinations in NA
disconfirm one of the concepts when ~ombined with the absence of some or no positive attributes to develop a set NAP.
The system may further include means for performing an initial backpropagation to train a neural network, means for finding hidden processing elements within a common layer having similar patterns of information and for replacing the similar pattern hidden processing elements with a single hidden processing element having information from one or more of the similar pattern hidden processing elements, and means for performing a further backpropagation on the neural network.
The data processing system may further include an inference engine in the memory. The inference engine ':

. ;

6i1 and stored rules in the memory are together a part of a rule-based expert system.
The method of produciny a rule-based expert system according to the present invention uses a neural network and initially translates knowledge in the output layer and each hidden layer of the neural network into a layer set of rules. The rules from the layer set of rules are rewritten by reformulating rules from one layer in terms of rules of another layer to eliminate undefined hidden concepts and thereby generate a rewritten set of rules having an equivalent performance to the neural network. As used herein, equivalent performance means that the application of the rules would produce the same conclusions as produced by the neural network in at least 95~ of samples supplied to the neural netwoxk and to the rules, which samples would collectively be a random 10%
of all situations in the problem domain. Equivalent performance would also extend to any situations where the application of the rules would produce the same conclusions as produced by the neural network in at least 95% of samples supplied to the neural network and to the rules, where the samples are random and a lower percentage (i.e., less than 10%) of the problem domain which one of skill in this field would consider as providiny a reasonable basis to evaluate the performance of the rules and the neural network.
Even more preferably, the rewritten set of rules will yield the same conclusions as the neural network in at least 95% of all possible combinations of attributes (i.e., input conditions) within a problem domain. The rewritten rule set is stoxed in a memory. Prior to the translation step, the neural network is trained by performing the initial hackpropagation as discussed 3~

above, finding and replacing hidden processing elements as discussed above, and performing a ~urther backpropagation as discussed above.
The translation step includes the substeps which produce the sets, PA, PAN, NA, and NAP in the manner described above. The presenting of each processing element with positive input combinations corrPsponding to positive attributes proceeds heuristically in that, upon finding a member of set PA having m possib]e attributes P1 to Pm~ m being any positive integer less than k, one or more further combinations corresponding to Pl to Pm combined with other positive attributes are considered members of PA without such further combinations being supplied to the processing element.
The presentiny of each processing element with negative input combinations corresponding to negative attributes proceeds heuristically in that, upon finding a member of set NA having negative attributes N1 to Nm~ one or more further combinations corresponding to N1 to Nm combined with other negative attributes are considered members of NA without any further combinations being supplied to the processing element. The rules are translated by searching for rules in each layer of processing elements starting at a hidden layer closest to the input layer and proceeding to the outer layer by way of any intervening hidden layers. The translation further includes the steps of detecting implausible combinations of outputs from processing elements in one layer and using the knowledge of implausible combinations to limit the search in the next layer to be searched.
The method of the present invention may alternately be described as including the step of performing an initial backpropagation to train a neural network ,~ .

~; ~3 7 S~ ~ ~iJ
_ 9 _ having processing elements with input ~eights, firlding hidden processing elements within a common layer having similar patterns of information, replacing the similar pattern hidden processing elements with a single hi~den processing element having information from one or more of the similar pattern hidden processing elements such that each hidden processing element in a layer eventually has a distinctive pattern of information, and performing a further backpropagation on the neural network. The method may further include, after the initial backpropagation and before the further backpropagation, the step of nullifying relatively small absolute value input weights to each processing element in a layer such that only a certain number of the larger absolute value input weights remain with non-~ero values. The nullifying step is preferably performed before the finding step. The replacing step replaces similar pattern hidden processing elements with a single hidden processing element having input weights which are the vector average of the input weights of the similar pattern hidden processing units which are replaced. The finding step locates similar pattern hidden processing elements by comparing vectors corresponding to input weights of the hidden processing elements. The method may further include the translation of knowledge in the neural network into a layer set of rules, the rewriting of the rules and storage of a rewritten rule set in a memory.
The present invention may alternately be considered as an apparatus having a neural network with an input layer of input units and different layers of processing elements including an output layer having output processing elements and at least one hidden layer having hidden processin~ elements, each processing , .

.

~;J~ J~J ) element having input w~ights. Means for training the neural network by performing an initial backpropagation on the neural network, finding hidden processing elements within a common layer having similar patterns of information, replacing the similar pattern hidden processing elements with a single hidden processing element having information from one or more of the similar pattern hidden processing elements such that each hidden processing element in a layer eventually has a distinctive pattern of information, and performing a further backpropagation on the neural network are included. ThP neural network may be a hardware network with the input units, output processing elements, and hidden processing elements all being hardware components.
BRIEF DESCRIPTION OF THE DRAWINGS
The above and other features of the present invention will be more readily understood when the following detailed description is considered in conjunction with the accompanying drawings wherein like characters represent like parts throughout the several views and in which:
FIG. 1 shows a simplified overview of the method of the present invention;
FIG. 2 shows a first embodiment system of the present invention;
FIG. 3 shows a neural network according to the present invention;
FIG. 4 is a simplified flowchart showing how a neural network i5 trained by the present invention;
FIGS. 5A and 5B show parts of a neural network before and after replacement of redundant hidden processing elements;

.

~7~

FIG. 6 is a simplified flowchart showing how a neural network is translated into rules by the present invention;
FIG. 7 is a tree diagram illustrating how the present invention searches for rules within a neural network;
FIG. 8 is another tree diagram further illustrating how the present invention searches for rules;
FIG. 9 shows the structure of a processing element which may be used in a sophisticated version of the neural computer according to the present invention;
FIG. 10 shows how a neural computer according to the present invention may supply rules to a digital computer rule-based expert system;
FIG. 11 shows the hardware configuration for an alternate embodiment of the present invention which uses a software neural network; and FIG. 12 is a chart showing the classification accuracy of a neural network and a rule-based expert system according to the present invention.
DETAILED DESCRIPTION
Turning first to FIG. 1, an overview of the method of the present invention will be presented. Briefly, the present invention allows one to combine many of the advantages of neural networks with rule based expert systems. In order to realize those benefits, data is used to train a neural network as shown at block 10.
Although the present invention will, in its broadest respects, work with neural networks which have been trained using standard training procedures including backpropagation, the more specific aspects of the present invention use a special form of training for the neural network. This special form of training will be discussed in further detail below.

,., ~ ' :
.~ .
2 ~ r3~

Following the completion of training of the neural network, the neural network will be translated at block 12 into rules. Significantly, the rules will be sufficiently complete to be the performance equivalent of the neural network. The specifics o~ the translation will be discussed in more detail below. At this stage, it is sufficient to indicate that rules are provided from block 12 to block 14. At block 14 the rules are stored in the memory of a rule-based expert system. The expert system generated by this technique will provide many of the advantages of rule-based expert systems, while avoiding or minimizing many of the disadvantages. The expert system produced according to the present invention will have several of the advantages normally associated with a neural network, while avoiding or minimizing disadvantages usually associated with a neural network.
As shown in FIG. 2, a first embodiment 16 of the present invention uses a neural co-processor 18 combined with a computer having CPU 20, disc/tapes 22, CRT terminal 24, input output devices 26, and memory 28.
Within the neural co-processor 18 is a trainer 30 which is used for training the neural network or co-processor. The structure of a neural network of co-processor 18 is indicated schematically at 32 within block 18. The structure will be discussed in more detail below.
As shown, the memory 28 includes a translator 34 and an expert system 36 disposed therein. The expert system 36 includes a rule-base 38 and an inference engine 40.
The overall operation o~ the system 16 of FIG. 2 proceeds as outlined in FIG. 1. Initially, the neural . .
.
.
, , ~ $j 7 '~ J j network 32 of co processor 18 has not yet been trained.
The trainer 30 trains the neural networ~ 32 by using data supplied to it (data not shown). After the neural network 32 is trained, the translator 3~ translates the neural network 32 into rules which are supplied to a rule base 38, which rule base was initially empty.
Thus, the expert system 36 has been produced using the neural network 32.
Turning now to FIG. 3, the neural network 32 is shown in detail as including five input elements 42 extending horizontally in an input layer. The input elements or nodes simply receive input data and convert it into the proper signal level or format such that it may be received by one or more of the processing elements 44, which processing elements are arranged in a horizontal row of four processing elements in a hidden layer and three processing elements in an output layer, also extending horizontally. For ease of illustration, not all of the processing elements or input elements have been labeled with the numerals.
Those processing elements in the hidden layer may be referred to as hidden processing elements, whereas those in the output layer may be referred to as output processing elements. Although the structure has been shown as only including a single hidden layer, it will be apparent that more than one hidden layer may be used.
The neural network 32 of FIG. 3 operates in a generally conventional manner except for certain important features which will be discussed below.
Thus, the operation will be understood by those of skill in the art. However, it may be useful to briefly review the principles of operation of such a neural ~ , ;

~ 14 -network 32, which networks often use neural chips and are sometimes referred to as connectionist systems.
Each of the processing elements ~ receives inpuks from each of the elements (either input elements or processing elements) in the upstream (lower in FIG. 3) layer. In the arrangement of FIG. 3, each element in one layer is connected to each element in the next (i.e., closer to output) layer and there are no connections directly from one layer to a non-adjacent layer. However, neural networks with other structures might also be used with the present invention. The output of any of the processing elements is given by OUT = g (~ wixi-e) where Wi and Xi are input weights and inputs respectively, e is the threshold, and the function g is the squashing function (sometimes called a transfer function or activation function) and is often chosen to be a sigmoid. A sigmoid function is:
g(net)= 1 ~ t 1 + e ~ ne where ~ determines the steepness of the function and net is simply the input to the function. For ease of discussion, the following description shall use the term sigmoid-sum to denote the sigmoid of the sum minus e, that is, sigmoid-sum = g (sum - e) By use of the multiple layers of processing elements, the neural network 32 can, after proper training, make decisions or draw conclusions based upon a pattern of data supplied at the input layer.
The inputs supplied to each processing element may be considered as attributes. These attributes correspond to Xi and affect the concept corresponding to the output of a particular processing element. A

.:- -."-:~ ' ' :' , , , concept at the output of a processing element may be a hidden concept if it is at the output of a processing element in the hidden layer or the concept may be a final or target concept if the concept corresponds to an output of a processing element in the output layer.
An example may be useful in order to explain the terminology. If the neural network 32 has been trained to determine if a mushroom is poisonous or edible, the input elements 42 of the input layer would be supplied with information about the characteristics, such as shape, coloring, odor, etc. of the mushroom in question. The input elements 42 encode these attributes and supply them to the processing elements 44 within the hidden layer. For example, if the shape of the cap of the mushroom was conical, an input would be supplied to the processing elements 44 in the hidden layer, which input corresponds to the attribute of a conical cap shape.
In the mushroom example, the final or target concept would be whether the mushroom is edible or poisonous. That conclusion is a concept corresponding to the output of the output layer. In order to provide that final or target concept, the processing elements 44 of the output layer use attributes corresponding to the outputs or concepts supplied by the processing elements 44 of the hidden layer. For example, one of the processing elements 44 within the hidden layer might output a concept indicating that the mushroom was in a particular family of mushrooms. That hidden concept would be supplied as an attribute to the processing elements 44 in the output layer and might be used in combination with other signals supplied by the hidden layer in order to reach the final or target concept indicating whether the mushroom is poisonous or , . . , ~ .
' : . ~

. . .
.
,, edible. Note that the concept out of one layer of processing elements serves as an attribllte at the next downstream layer of processing elements. The example of a hidden concept correspondiny to the fact that the mushroom is in a particular ~amily of mushrooms is a defined hidden concept, meaning that an expert on mushrooms might readily appreciate and understand the concept of whether a mushroom was in a particular family or not. However, hidden concepts are often undefined, meaning that they are complex combinations of characteristics. For example, a hidden concept might correspond to the mushroom stalk shape being tapering, the cap shape being conical, the spore print color being purple, and the odor being almond. This arbitrary example might be considered as an undefined hidden concept because an expert on mushrooms would not ~e able to say that this comhination of characteristics identifies a particular family or grouping of mushrooms or otherwise identifies a particular single property commonly referred to by experts in this field.
The individual processing elements 44 within the neural network 32 may be constructed using known structures, which structures are occasionally referred to as neurons because of the rough similarity between the operation of a neural or connectionist network and the neurons in a human brain. Of course, the processing elements 44 could be made of resistors, capacitors, and/or other components commonly used in neural chips, which chips are in turn used to make neural networks.
The outpu~ of a processing element may be considered to range between 0 and 1. If the output of a processing element is close to 1, this means that the concept defined by that processing element is true, . .
.. . ..

. ~

~ ~ ~ r~
~ 17 --whereas if the output of a processing element 44 approaches o, this indicates that the concept corresponding to that processing element is not true or does not exist. Turning back to the example of the hidden processing element having a hidden concept at its output which indicates whether a mushroom is in a particular family of mushrooms, a level of approximately 1 at the output would indicate that the mushroom is indeed within that family of mushrooms. An output of about 0 would indicate that the mushroom in question is not within the particular family of mushrooms~ The closeness required for the output to be considered 1 and the closeness required for the output to be considered 0 will be discussed below.
Two kinds of attributes to a particular processing element may be considered. For a given concept corresponding to the output of a particular processing element, an attribute tending to confirm that concept (making the activation or output approach 1) is considered a positive attribute. A negative attribute for the given concept refers to attributes which disconfirm (making the activation or output approach 0) the concept. An attribute is a positive attribute for a concept if and only if it links to the concept through a connection with a positive weight. An attribute is a negative attribute for a concept if and only if it links to the concept through a connection with a negative weight. Note that an attribute which is positive relative to one concept may be a negative attribute relative to another concept. The given definitions of positive and negative attributes are based on the fact that the sigmoid function is a monotonically increasing function and based on the restriction of activation to the range between 0 and 1.

.

r~
-- 18 ~
It should also be mentioned that in a multi-layer network (two or more layers of processing elements) positive attributes and negative attributes are only defined relative to each layer. This is krue because an attribute may support a hidden concept (i.e., the mushroom is in a particular family) which is a positive attribute for the final concept (the mushroom is edible) and at the same time support another hidden concept which is a negative attribute for the final concept. It may turn out that in this case, the type of contribution of an attribute to a final concept is conditioned on the values of other attributes.
In order to train the neural network 32 of FIG. 3, training cases or data are supplied to the neural network. As the training process proceeds, the weights Wi are adjusted in magnitude in polarity.
The training process used with the present invention is shown in FIG. 4 and begins with block 46 wherein backpropagation i5 performed. This backpropagation involves a search along the gradient surface and error propagation from the output layer to hidden layers. However, as backpropagation is known in this field, the details of the backpropagation need not be discussed in detail. However, the backpropagation would be of the type described in Rumelhart, Hinton, and Williams 1986 article entitled "Learning Internal Representation by Error Propagation" which appeared in Parallel Distributed Processin~: Explorations in the Microstructures of Coqnition, Volume 1, MIT Press, Cambridge, Massachusetts. That article is hereby incorporated by reference.
Following the backpropagation of block 46, the training process of FIG. 4 leads to block 48. In that block, discriminatory weight training is performed.

"1 This discriminatory weight tr~ining involves nulli~ying (setting to zero) relatively small absolute value input weights to each processing element in a layer such that only a certain number of the larger absolute value input weights remain with non-zero values. For example, if a processing element had input weights Wi of .1, .2, .1, .5, .6, .5, and .7, discriminatory weight training in which the four input weights having the largest absolute value are selected and the remaining input weights are set to zero would involve changing the first three listed input weights to zero.
Note that the polarity of the input weights does not affect this decision as the discriminatory weight training nulls the smallest weights in terms of absolute value. Assuming that the procedure selected only the top four weights in terms of their absolute value, the input weights listed above would be changed to be 0, 0, 0, .5, .6., .5, and .7. Note that in actual practice, one might want to use the top seven weights in terms of absolute value or one might use any other number. The discriminatory weight training avoids having very small effects on a concept at the output of a processing element and will ease the translation of the knowledge in the neural network 32 into the rules format for a rule-based expert system.
More generally, the discriminatory weight training simply puts the neural network 32 in a better form for later translation than would otherwise be the case.
Following block 4~, control is transferred to block 50 whexein clustering of hidden processing elements is performed in order to remove redundancy within the hidden layers. This clustering may best be explained by use of an example. Assume that a hidden processing element or unit H has input weights of w1h, w2h, and , ..
' ~

~7~i2 ~j w3h and that another processing element G within that same hidden layer as H has input weights of w1g, w2g, and w3g. The input weights of the two different hidden units may be considered to constitute vectors. If these input weights have similar patterns of information, there will be some redundancy between the information stored in element G and the information stored in processing element H. The present invention compares the cosine of the angle between ~he two vectors defined by the three input weights for each of the two processing alements. If the cosine of the angle is approximately equal to one, this means that the patterns of the two input weight vectors are approximately the same. By using a threshold value to determine the similarity of the two input weight vectors as indicated by the cosine of the angle between the weight vectors, one can determine if the corresponding processing elements are encoding redundant information. Although the example given here simply refers to weight vectors having three components, it will be readily appreciated that the input weight vectors may have a larger number of components and the same test for similari~y may be applied. It should also be noted that each of the processing elements ~ithin a particular hidden layer are tested for similarity relative to each of the other processing elements within that hidden layer. Thus, the clustering may indicate that three processing elements have redundant information in that they are essentially duplicative of each other and the clustering process may indicate that two other hidden processing elements are redundant relative to each other. In that case, the three hiddan processing elements of the first group are replaced by a new , .
.

: ' ' ' .
.

~l3~ ~ 2 ~ 3 hidden unit having an input weight vector corresponding to the average of the three elements or uni~s which are being replaced. Likewlse, the gxoup of two redundant hidden processing elements would be replaced by a single processing element having an input weight vector which is the average of the two input weight vectors of those two units being replaced. Block 50 illustrates a group 51A of two hidden processing elements or units which are clustered and a group 51B o~ two other clustered units.
Turning momentarily to FIGS. 5A and 5B, a small portion of a neural network is illustrated in order to show how this clustering and replacement process works.
Processing elements H1 and H2 of FIG. 5A receive signals from three different previous units (not shown) and have input weights as illustrated. Assuming that the threshold test indicates the redundancy between the information of elements H1 and H2, these two units would be replaced by a single unit new H3 of FIG. 5B.
Note that H1 and H2 supply signals to processing element G1 which is in the output layer or in a hidden layer closer to the output than the layer in which H1 and H2 are disposed. In order to remove the redundancy of H1 and H2 essentially duplicating each other, H1 and H2 of FIG. 5A are replaced as shown at FIG. 5B where new H3 has input weights which are the average of the input weights for ~ormer Hl and H2. Note also that the output weight vector o~ the new H3 element or unit would also be given the average output weight vector of the units H1 and H2. In other words, the input weight at unit Gl is replaced by the average o~ the two output weights (i.e., input weights to the next element) which are being replaced.

. , ~
.t :

~ ~J

Following the removal of the redundancy in the hidden units by the cluster and replacement process as described, block 50 of FIG. 4 proceeds to block 52 wherein a further step of backpropagation is perforrned.
This further step of backpropagation helps to improve the accuracy of the neural network.
Although FIG. 4 shows the simplified process for the training of the neural network according to the present invention, an Appendix A is attached as part of this application and shows the actual computer program named KTBP in LISP for the indicated training process.
Once the neural network is trained, the present invention interprets or translates its knowledge into rules. In order to do so, a mechanism is required to translate numerical knowledge into symbolic knowledge.
As the activation level at each node in the neural network may range between 0 and 1, parameters ~ and are used such that the activation level smaller than the ~ value is treated as a logical no and the activation level greater than the ~ value is treated as a yes. The ~ must be greater than the ~ value in order to avoid ambiguities. The possibilities of the ~ value range between 0 and .5, whereas the possibilities for the ~ value fall between .5 and 1 when the choices for ~ and ~ approach .5, rules formed are more general, but may be noisy. On the other hand, when the choices go to the extremes t~ almost 0, ~ almost 1) rules tend to be more specific and restrictive. It appears that rules in the former case reflect more the continuous nature of the neural network, whereas rules in the latter case capture the more discreet nature of the neural network. The choice of values ~or ~ and ~ will vary depending on the domain under study. In addition, the present invention preferably will use different ,'; , ` ~i"' ' '' ' - ' . .

~ ~ 7 l ~3 ~, 1 values for ~ and ~ for the hidden units than the values used for ~ and ~ for the output units. For example, the values used ~or ~ and ~ at the output processing elements or units may be .1 and .9 respectively, whereas the values at the hidden layers may be .4 and .6.
Turning now to FIG. 6, there is shown a simplified flowchart of the translator 34 of FIG. 2 corresponding to the translation operation of block 12 of FIG. 1. As shown at block 54, the translation process initially requires that the knowledge of the output layer and each hidden layer are translated into rules. The form of the rules is diagramed at the bottom of block 54 which also shows how the rules relate to different layers within the neural network. In order to translate the knowledge in each layer into rules, block 54 includes within it blocks 56 and 58 which are different procedures. Each o~ the procedures is p~rformed for each processing element, also referred to as a concept node.
The block 56 procedure FORM-CONFIRM-RULE searches for rules, each o~ which is able to confirm a particular concept independently. This procedure of block 56 involves simulating the firing of a rule by presenting the network with any input matching the antecedent of any such rule such that the activation of the corresponding processing element will be greater than the ~ value. The procedure FORM-CONFIRM-RULE
calls the procedures EXPLORE-POS and NEGATE-NEG which involve heuristic search derived from numerical constraints associated with positive attributes and negative attributes as will be discussed below.
The rules which are to be generated for a rule- ;
based expert system will have one or multiple .: ' . .

. ~.

~ ~ 7 '~

conditions and one conclusion. q~he size of a rule is determined by the number of conditions. In each layer of processing elements, the translation process explores the number of attributes in combina~ions up to k. The maximum possible rule size is determined by kd where d is the number of layers. For example, given k equals 5 and d equals 3, the largest possible rule size is 125. This is believed to be more than adequate in practical domains.
10 The output of EXPLORE-POS is a set PA of all combinations of at most k pos-atts Pi (positive attributes) each of which can confirm the concept i~
all neg-atts Ni (negative attributes) are absent.
Then, NEGATE-NEG is applied to each such combination in an attempt to finding all rules each of which can confirm the concept in the absence of some or no neg-atts. In brief, EXPLORE-POS searches for combinations of pos-atts ~ ~Pl, .., Pi) ~....} ....}
such that sigmoid-sum-wt(P1, .., Pi) > ~ where sigmoid-sum-weight denotes the sigmoid-sum of weights. NEGATE-NEG searches for combinations of pos-atts and negated neg-atts { ~Pl, .., Pi, (not Nl),.. (not Nj)~ ~...) ....~
such that sigmoid-sum-wt (Pl, .., Pi, Nj+l, .., Nq) >
. Note that ~Nj+l, .., Nq~ is the complement of ~Nl, .., Nj} with respect to the set of all neg-atts.
The search tree in EXPLORE-POS shown in FIG. 7 starts with an empty set as the root node. A node is expanded by adding a new tnot in the node) pos-att in all possible ways. Redundant nodes, nodes with conflicting attributes as declared (such as mutually exclusive attributes), and nodes with more than k attributes are pruned. Mutual exclusiveness can be .
. ' ' ' .

.

~ 0'~ 3 declared explicitly for input attribute values, or can be dsduced usiny theorem proving for hidden attribu~es once their necessary characteristics are learned.
Suppose a node has 1 attributes. If the sigmoid-sum of their weights plus those of k - 1 other strongest (strongest in terms of the absolute magnitude of associated weights), non-conflicting pos-atts is not greater than the ~ value, then prune the node, because the node cannot produce any good confirming rule. A
node kept in the tree will be put in the output set PA
if the sigmoid-sum of the associated weights of all attributes contained in the node is greater than the value. Although the translation process attempts to find the most general rules, EXPLORE-POS may keep all nodes which meet the above condition. So, for example, if EXPLORE-POS accepts the combination of attributes Al and A2, it also accepts the combination of attributes Al, A2, and A3. Though the latter combination is more restrictive (or specific) than the former, they may take different combinations of negated neg-atts in conjunction to form confirming rules and the subsumption relationship may no longer exist. Thus, EXPLORE-POS may keep all rather than the most general combinations of pos-atts under the predefined criterion. However, there is one case when a node needs not be further specialized; that is, when the attributes of the node in conjunction with all neg-atts (for mutually exclusive attributes, select the strongest one) yield a sigmoid-sum greater than the ~
value. Under the circumstance, adding more pos-atts is not necessary. The depth of the search tree (or graph) will not exceed k.
The search of EXPLORE-POS is heuristic in that, as it generates the combinations of positive attributes , ~

, `

~ ~ '7 ~ '3 for a particular node or processing element, it does not supply unnecessary combinations to the processing element. An example may be useful in understanding this with reference to FIG. 7. Assuming that the single positive attribute Pl is sufficient to cause a processing element to have an activation level higher than ~ (in other words the output concept of that processing element is confirmed), one knows that any other positive attributes combined with P1 will likewise confirm the output concept. Thus, the combinations {Pl P2~, {P1 P3), {P1 P4) will necessarily result in the output concept being confirmed as well.
Thus, those last three combinations may be included in a set PA without necessarily having to supply those input combinations to the processing element being tested. From the fact that P1 confirms the concept, and the knowledge that P2, P3, and P4 are positive attributes, one will realize that they in combination with P1 will confirm the concept.
The EXPLORE-POS program is further illustrated by programming design language corresponding to it as shown in Appendix B attached to the pr~sent application.
The search tree shown in FIG. 8 in NEGATE-NEG
(programming design language attached as Appendix C) starts with any combination of pos-atts ~set PA) returned by EXPLORE-POS. A node is expanded by adding a new, negated neg-att in all possible ways. Redundant nodes, nodes with conflicting attributes as declared, and nodes with more than k attributes are pruned. A
node kept in the tree will not generate successors if the sigmoid~sum of the associated weights of all non-negated attributes present plus those neg-atts absent in the node is greater than the ~ value (the neg-atts 2~7~7 absent in a rule may be present or absent in the environment, and we assume the worst scenario). Since the conjunction of all attributes in such a node is already strong enough to form a confirming rule, there is no need to add more negated neg-atts. ~ecall that the translation is intended to find the most general rules. Such a node will be put in the output.
Note that the output of NEGATE-NEG is a set tPAN3 of premises to confirm the given concept. A confirming rule is formed by putting the conjunction of all members in a premise on the left hand side and the concept on the right hand side.
Consider a simple example here. Suppose the pos-atts and neg-atts relevant to forming rules for a certain concept are: P1, P2, P3, N1, N2, and N3; their respective associated weights are: 0.05, 0.09, 0.03, -0.05, -0.01, and -0.01. Assume that the nodal threshold for the concept ~ = 0, the sigmoid slope ~ =
20, 5 < ~ < 30, and ~ = 0.9. Then, the combination of Pl and P2 ({P1 P23) will be in the output of EXPLORE-POS since their sigmoid-sum is calculated to be 0.94. And the combination of P1, P2, negated N1 ((Pl P2 (not Nl))) will be in the output of NEGATE-NEG
because the sigmoid sum of P1, P2, N2, and N3 is 0.92.
The procedure FORM-DISCONFIRM-RULE of block 58 of FIG. 6 is similar to FORM-CONFIRM-RULE. The difference is that the roles of pos-atts and neg-atts are exchanged. It searches for rules each of which produces an activation level less than the a value for the given concept as long as its premise is true. The pruning heuristics are changed accordingly. EXPLORE-NEG generates a set NA and NEGATE-POS generates a set NAP.

' , , .
.
, , :

J ~. I

The rules are kept in the most general ~orm. In terms of the cost and the ef~iciency of using rules, general rules are more deslrable than specific rules.
Some machine learning programs perform maximally specific generalization when no sufficient number of counter-examples is available for learning.
To deal with multi-layer networks, the translation learns rules on a layer-by-layer basis, then rewrites rules to obtain rules which link attributes directly to a final (target) concept. In forming rules between a hidden layer and a concept, each hidden unit in the layer is treated as an attribute. Since this is not an original attribute, it is called a hidden attribute (or a hidden concept). The hidden attributes with non-zero weights are then categoried into pos-hidden-atts and neg-hidden-atts. The KT translator algorithm treats pos-hidden-atts and neg-hidden-atts the same as pos-atts and neg-atts respectively.
The purpose of rewriting is to eliminate undefined hidden concepts. However, if some hidden units encode useful intermediate concepts (e.g., the example above of membership in a family of mushrooms), they can be kept. In other words, predefined or useful hidden concepts are not rewritten. In this way, the present invention can generate a hierarchical rule-based system as well.
The procedure REWRITE of block 60, which follows block 5~ of FIG. 6, rewrites rules of one layer in terms of rules of another layer. For example, in layerl, there is one rule:
M1 and NOT (M2) ~ C.
"X ~ Y" reads "if X, then Y". In layer2, there are three rules:

, ~.' .

A1 and NOT (B1) ~ M1.
A3 ~ M1.
B2 and NOT (A2) ~ NOT (M2).
Rewriting rules of layerl in terms of rules o~ layer2, we obtain:
Al and ~OT (B1) and B2 and NOT (~2) ~ C.
A3 and B2 and NOT (A2) ~ C.
This rewriting process is backward in the sense that the antecedent of a rule is rewritten on the basis o~
rules whose consequents deal with its antecedent.
Every time, REWRITE rewrites rules of a layer in terms of rules of the next layer closer to the input of the net. Rewriting repeats until rules which associate attributes with final concepts result.
After each rewriting, if a rule contains an attribute and its negation, then delete the rule; if a rule contains conflicting attributes, then delete the rule; if a rule contains redundant attributes, then delete the redundant ones; remove redundant rules: in addition, remove a rule whose premise part is a superset of that of any other rule (that is, remove subsumption). A rule which cannot be rewritten will be discarded.
The present translator can learn rules based on a single case by focusing on those attribute combinations applicable to that case (the case-based mode). In the hidden layer, only those hidden units which get activated with the case input are considered for rule formation. Because of a smaller search space, more efficient search is permitted. This case-based mode is taken to supplement the general mode to learn rules.
The overall translation process o~ FIG. 6 including the generation of the layer set of rules and the rewriting of the rules is shown in detail by the - : . : ,. . ,, :.

translator program KT in the LISP language in Appendix D to the present application.
With reference now to FIG. 9, there is shown a processing element 62 which may be used in the arrangement of FIG. 3 in order to provide a neural computer. In other words, this may provide an alternative to the structure of FIG. 2 wherein the translator 34 is within the memory of a digital computer. The processing element 62 of FIG. 9 would, when configured with other such processing elements according to the structure of FIG. 3, realize a distributed computing network. The processing element 62 is sophisticated in that it includes a main memory 64 having the trainer and translator therein in addition to the weights and rules. The memory 64 i5 connected to a CPU 66 and in turn connected to communication devices 68 used to communicate with other such processing elements. By using a processing element as sophisticated as that of FIG. 9, with the structure of FIG. 3, one may realize a neural computer which, following its training, may translate and provide rules without requiring a separate digital computer to perform the translation process. In such a neural computer arrangement network training and rule translation are performed in parallel. Each processing element has its stored weights translated into rules independently and in parallel. Rewriting proceeds from the input towards the output on a layer-by-layer basis, but processing elements in each layer rewrite in parallel.
Following the use of a neural computer structured like FIG. 3 and using the sophisticated processing elements 62 of FIG. 9, such a neural computer shown as 2 0 ~7~ 3 70 in FIG. 10 could be connected to a digita] computer 72 for supplying the rules thereto.
Turning now to FIG. 11, there is shown an alternate embodiment of the system of the present invention. The system 116 of FIG. 11 has the components with the same last two digits as the corresponding component, if any, from the embodiment of FIG. 2. The FIG. 11 embodiment uses a software neural network 132 which is disposed in the main memory 128. The software neural network 132 would be structured using software components corresponding to the processing elements and input nodes or input elements as discussed above. The memory 12~ would also include trainer 130, translator 13~, and expert system 136. The expert system 136 may be configured as shown in more detail for the expert system 36 of FIG. 2.
The memory 128 connects to the CPU 120, which in turn is connected to the disc/tapes 122, terminal 124, and input/output devices 126.
Note that the rules produced by any of the versions of the present invention may be printed so as to provide a written set of rules for use by a person.
The present invention (software version of FIG. 11) has been applied to create expert systems in several different domains. As a relatively simple problem, the present invention was used to provide an expert system on the characteristics of an exclusive or gate. The present invention easily found the four rules governing the exclusive or function. Although the exclusive or problem or function is a relatively small problem, it is nonlinear.
In addition to the exclusive or problem, the present invention has been applied to wind shear . . . :

: . :
.
, ' ' ~.
, "~
.

~ ~ y~

detection, genetic trait analysis, and mushroom classification.
In the domain of wind shear detection by infrared sensors, the present invention can efficiently produce a small set of rules from a neural network with nearly 100% mutual coincidence rate in prediction. ln other words, the rules, which may be used as a rule base in an expert system, correspond well with the neural network.
In genetic trait analysis, the present invention has successfully identified five genes related to diabetic mellitus trait on mouse chromosome 3, 6, 9, 11, 17 among 62 genetic loci. The present invention offers a useful mechanism to appraise the phenotypic effect of gene combinations.
In the publicly accessible mushroom domain which contains 8,124 cases, the present invention generated a rule-based system of 233 rules based upon 1,000 training instances or cases. The neural network built for this domain included 127 input units (encoding 127 nominal feature values; there are 22 features, and the number of values for each feature ranges from 2 to 12), 63 hidden units, 2 output units, and 8,1~7 connections.
The result of classification by the rule base system is given in FIG. 12. Unclassified cases refer to those cases which do not match any rule in the rule base.
This rule-based system is able to classify the 1,000 training instances with 100% accuracy and a disjoint set of 1,000 test instances with 99.6 percent accuracy (both misclassified and unclassified cases were considered incorrect). The contrast in performance between the neural network and the rule-based system is shown in FIG. 12. Somewhat surprisingly, the rule -based system classified the test set more accurately , - 3~ ~
than the neural network. This result may be explained by the capability of the present invention to sort out decision knowledge from information distributively entangled in the network; the loss of desired information versus noise seems negligible. Among the 233 rules, 67 rules are matched by at least one training instance. The other rules are statistically implied by the training instances and could be invoXed by unseen instances. In other words, cases which were not used in training may be answered by the system.
Appendix E to the present application includes a small number of the rules generated by the present invention relative to the mushroom example.
Although various specific structures and details have been presented herein, it is to be understood that these are for illustrative purposes only. Various modifications and adaptations will be apparent to those of skill in this art. Accordingly, the scope of the present invention should be determined by reference to the claims appended hereto.

; ~

, : .

Claims (24)

1. A data processing system comprising:
a translating means for translating knowledge in a neural network having an input layer of input units, and different layers of processing elements including an output layer of output processing elements and at least one hidden layer of hidden processing elements, said translating means operable to translate knowledge in the outer layer and each hidden layer of the neural network into a layer set of rules;
a rewriting means for rewriting rules from the layer set of rules by reformulating rules from one layer in terms of rules of another layer to eliminate undefined hidden concepts and thereby generate a rewritten set of rules, the rewritten set of rules having an equivalent performance to the neural network; and a memory for receiving and storing the rewritten set of rules.
2. The data processing system of Claim 1 further comprising a neural network having an input layer of input units, and different layers of processing elements including an output layer of output processing elements and at least one hidden layer of hidden processing elements and wherein said translation means and rewriting means are operably connected to said neural network.
3. The data processing system of Claim 2 wherein said neural network is a hardware network with said input units, output processing elements, and hidden processing elements all being hardware components.
4. The data processing system of Claim 3 further comprising a CPU and a memory operably connected to said CPU and wherein said translating means and rewriting means are disposed in said memory.
5. The data processing system of Claim 2 further comprising a CPU and a memory operably connected to said CPU and wherein said translating means, rewriting means, and neural network are disposed in said memory, and said input units, output processing elements, and hidden processing elements all being software components.
6. The data processing system of Claim 1 further comprising a neural network having an input layer of input units, and different layers of processing elements including an output layer of output processing elements and at least one hidden layer of hidden processing elements, and wherein said translation means and said rewriting means are part of one or more of said output processing elements and said hidden processing elements.
7. The data processing system of Claim 6 wherein said memory is within at least one of said output processing elements and said hidden processing elements.
8. The data processing system of Claim 1 further comprising a neural network having an input layer of input units, and different layers of processing elements including an output layer of output processing elements and at least one hidden layer of hidden processing elements, and wherein said translating means is operable to translate knowledge from said neural network while distinguishing between positive attributes, which link to a concept corresponding to an output of one of said processing elements by a positive weight, and negative attributes, which link to an output concept of one of said processing elements by a negative weight, and wherein said translating means translate by:
presenting each processing element with positive input combinations up to a limit k corresponding to positive attributes for that processing element to develop a set PA of all combinations of at most k positive attributes, each of which confirms one of said concepts if all negative attributes are absent;
determining which of said input combinations in PA confirm one of said concepts when combined with the absence of some or no negative attributes to develop a set PAN;
presenting each processing element with negative input combinations up to k corresponding to negative attributes for that processing element to develop a set NA of all combinations of at most k negative attributes each of which disconfirms one of said concepts if all positive attributes are absent; and determining which of said input combinations in NA disconfirm one of said concepts when combined with the absence of some or no positive attributes to develop a set NAP.
9. The data processing system of Claim 1 further comprising means for performing an initial backpropagation to train a neural network having an input layer of input units and different layers of processing elements including an output layer having output processing elements and at least one hidden layer having hidden processing elements; means for finding hidden processing elements within a common layer having similar patterns of information and replacing the similar pattern hidden processing elements with a single hidden processing element having information from one or more of the similar pattern hidden processing elements; and means for performing a further backpropagation on the neural network.
10. The data processing system of Claim 1 further comprising an inference engine in said memory, said inference engine and stored rules in the memory being part of a rule-based expert system.
11. A method of producing a rule-based expert system by use of a neural network having an input layer of input units and different layers of processing elements including an output layer and at least one hidden layer, the steps comprising:
translating knowledge in the output layer and each hidden layer of the neural network into a layer set of rules;
rewriting rules from the layer set of rules by reformulating rules from one layer in terms of rules of another layer to eliminate undefined hidden concepts and thereby generate a rewritten set of rules, the rewritten set of rules having an equivalent performance to the neural network: and storing said rewritten rule set in a memory.
12. The method of Claim 11 wherein, prior to said translation step, the neural network is trained by the following substeps in order:
performing an initial backpropagation to train a neural network having an input layer of input units and different layers of processing elements including an output layer having output processing elements and at least one hidden layer having hidden processing elements;
finding hidden processing elements within a common layer having similar patterns of information and replacing the similar pattern hidden processing elements with a single hidden processing element having information from one or more of the similar pattern hidden processing elements; and performing a further backpropagation on the neural network.
13. The method of Claim 11 wherein said translation step includes the substeps of:
presenting each proc2ssing element with positive input combinations corresponding to positive attributes up to k for that processing element to develop a set PA of all combinations of at most k positive attributes each of which confirms one of said concepts if all negative attributes are absent;

determining which of said input combinations in PA confirm one of said concepts when combined with the absence of same or no negative attributes to develop a set PAN;
presenting each processing element with negative input combinations corresponding to negative attributes up to k for that processing element to develop a set NA of all combinations of at most k negative attributes each of which disconfirms one of said concepts if all positive attributes are absent; and determining which of said input combinations in NA disconfirm one of said concepts when combined with the absence of same or no positive attributes to develop a set NAP.
14. The method of Claim 13 wherein said presenting of each processing element with positive input combinations corresponding to positive attributes proceeds heuristically in that, upon finding a member of set PA having m positive attributes Pl to Pm, m being any positive integer less than k, one or more further combinations corresponding to Pl to Pm combined with other positive attributes are considered members of PA without such further combinations being supplied to the processing element, and wherein said presenting of each processing element with negative input combinations corresponding to negative attributes proceeds heuristically in that, upon finding a member of set NA having m negative attributes Nl to Nm, m being any positive integer less than k, one or more further combinations corresponding to Nl to Nm combined with other negative attributes are considered members of NA without such further combinations being supplied to the processing element.
15. The method of Claim 11 wherein the rules are translated by searching for rules in each layer of processing elements starting at a hidden layer closest to the input layer and proceeding to the output layer by way of any interviewing hidden layers, and further comprising the steps of detecting implausible combinations of outputs from processing elements in one layer and using the knowledge of implausible combinations to limit the search in the next layer to be searched.
16. A method comprising the steps in the following order of:
performing an initial backpropagation to train a neural network having an input layer of input units and different layers of processing elements including an output layer having output processing elements and at least one hidden layer having hidden processing elements, each processing element having input weights;
finding hidden processing elements within a common layer having similar patterns of information;
replacing the similar pattern hidden processing elements with a single hidden processing element having information from one or more of the similar pattern hidden processing elements such that each hidden processing element in a layer eventually has a distinctive pattern of information: and performing a further backpropagation on the neural network.
17. The method of Claim 16 further comprising, after said initial backpropagation and before said further backpropagation, the step of nullifying relatively small absolute value input weights to each processing element in a layer such that only a certain number of the larger absolute value input weights remain with non-zero values.
18. The method of Claim 17 wherein said nullifying step is performed before said finding step.
19. The method of Claim 16 wherein said replacing step replaces similar pattern hidden processing elements with a single hidden processing element having input weights which are the vector average of the input weights of the similar pattern hidden processing units which are replaced.
20. The method of Claim 19 wherein said finding step locates similar pattern hidden processing elements by comparing vectors corresponding to input weights of the hidden processing elements.
21. The method of Claim 16 further comprising, following said further backpropagation:
translating knowledge in the output layer and each hidden layer of the neural network into a layer set of rules;
rewriting rules from the layer set of rules by reformulating rules from one layer in terms of rules of another layer to eliminate undefined hidden concepts and thereby generate a rewritten set of rules, the rewritten set of rules having an equivalent performance to the neural network; and storing said rewritten rule set in a memory.
22. An apparatus comprising:
a neural network having an input layer of input units and different layers of processing elements including an output layer having output processing elements and at least one hidden layer having hidden processing elements, each processing element having input weights; and means for training said neural network by doing the following in order: performing an initial backpropagation on the neural network;
finding hidden processing elements within a common layer having similar patterns of information;
replacing the similar pattern hidden processing elements with a single hidden processing element having information from one or more of the similar pattern hidden processing elements such that each hidden processing element in a layer eventually has a distinctive pattern of information; and performing a further backpropagation on the neural network.
23. The apparatus of Claim 22 wherein said neural network is a hardware network with said input units, output processing elements, and hidden processing elements all being hardware components.
24. The apparatus of Claim 23 further comprising:
a translating means for translating knowledge in a neural network having an input layer of input units, and different layers of processing elements including an output layer of output processing elements and at least one hidden layer of hidden processing elements, said translating means operable to translate knowledge in the outer layer and each hidden layer of the neural network into a layer set of rules;
a rewriting means for rewriting rules from the layer set of rules by reformulating rules from one layer in terms of rules of another layer to eliminate undefined hidden concepts and thereby generate a rewritten set of rules, the rewritten set of rules having an equivalent performance to the neural network;
a memory for receiving and storing the rewritten set of rules.
CA002077626A 1991-09-09 1992-09-04 Translation of a neural network into a rule-based expert system Abandoned CA2077626A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US757,040 1985-07-19
US07/757,040 US5438644A (en) 1991-09-09 1991-09-09 Translation of a neural network into a rule-based expert system

Publications (1)

Publication Number Publication Date
CA2077626A1 true CA2077626A1 (en) 1993-03-10

Family

ID=25046108

Family Applications (1)

Application Number Title Priority Date Filing Date
CA002077626A Abandoned CA2077626A1 (en) 1991-09-09 1992-09-04 Translation of a neural network into a rule-based expert system

Country Status (4)

Country Link
US (2) US5438644A (en)
AU (1) AU2505892A (en)
CA (1) CA2077626A1 (en)
WO (1) WO1993005473A1 (en)

Families Citing this family (41)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE4410709A1 (en) * 1994-03-28 1995-10-05 Bodenseewerk Geraetetech Monitoring device for monitoring the flight safety of aircraft
US5583964A (en) * 1994-05-02 1996-12-10 Motorola, Inc. Computer utilizing neural network and method of using same
US5787408A (en) * 1996-08-23 1998-07-28 The United States Of America As Represented By The Secretary Of The Navy System and method for determining node functionality in artificial neural networks
US6249252B1 (en) 1996-09-09 2001-06-19 Tracbeam Llc Wireless location using multiple location estimators
US7714778B2 (en) 1997-08-20 2010-05-11 Tracbeam Llc Wireless location gateway and applications therefor
US7903029B2 (en) 1996-09-09 2011-03-08 Tracbeam Llc Wireless location routing applications and architecture therefor
US9134398B2 (en) 1996-09-09 2015-09-15 Tracbeam Llc Wireless location using network centric location estimators
WO1998010307A1 (en) 1996-09-09 1998-03-12 Dennis Jay Dupray Location of a mobile station
US6236365B1 (en) 1996-09-09 2001-05-22 Tracbeam, Llc Location of a mobile station using a plurality of commercial wireless infrastructures
US6336109B2 (en) * 1997-04-15 2002-01-01 Cerebrus Solutions Limited Method and apparatus for inducing rules from data classifiers
US5897627A (en) * 1997-05-20 1999-04-27 Motorola, Inc. Method of determining statistically meaningful rules
US6289329B1 (en) * 1997-11-26 2001-09-11 Ishwar K. Sethi System for converting neural network to rule-based expert system using multiple-valued logic representation of neurons in feedforward network
US6054710A (en) * 1997-12-18 2000-04-25 Cypress Semiconductor Corp. Method and apparatus for obtaining two- or three-dimensional information from scanning electron microscopy
US8135413B2 (en) 1998-11-24 2012-03-13 Tracbeam Llc Platform and applications for wireless location and other complex services
WO2002000316A1 (en) 1999-09-24 2002-01-03 Goldberg Sheldon F Geographically constrained network services
US6564198B1 (en) 2000-02-16 2003-05-13 Hrl Laboratories, Llc Fuzzy expert system for interpretable rule extraction from neural networks
US10684350B2 (en) 2000-06-02 2020-06-16 Tracbeam Llc Services and applications for a communications network
US10641861B2 (en) 2000-06-02 2020-05-05 Dennis J. Dupray Services and applications for a communications network
US9875492B2 (en) 2001-05-22 2018-01-23 Dennis J. Dupray Real estate transaction system
US20030167454A1 (en) * 2001-03-30 2003-09-04 Vassil Iordanov Method of and system for providing metacognitive processing for simulating cognitive tasks
US8082096B2 (en) 2001-05-22 2011-12-20 Tracbeam Llc Wireless location routing applications and architecture therefor
US20030069869A1 (en) * 2001-10-05 2003-04-10 Jane Gronau Computer aided strategic planning systems and methods
US20040010481A1 (en) * 2001-12-07 2004-01-15 Whitehead Institute For Biomedical Research Time-dependent outcome prediction using neural networks
US20080154809A1 (en) * 2006-10-20 2008-06-26 Genalytics, Inc. Use and construction of categorical interactions using a rule gene in a predictive model
US20080154808A1 (en) * 2006-10-20 2008-06-26 Genalytics, Inc. Use and construction of time series interactions in a predictive model
JP5582707B2 (en) * 2009-02-27 2014-09-03 キヤノン株式会社 Medical decision support apparatus and control method thereof
US9538493B2 (en) 2010-08-23 2017-01-03 Finetrak, Llc Locating a mobile station and applications therefor
US10540354B2 (en) * 2011-10-17 2020-01-21 Micro Focus Llc Discovering representative composite CI patterns in an it system
US10427192B2 (en) 2015-05-15 2019-10-01 Ronald G. Presswood, Jr. Method to recycle plastics, electronics, munitions or propellants using a metal reactant alloy composition
US10268965B2 (en) 2015-10-27 2019-04-23 Yardi Systems, Inc. Dictionary enhancement technique for business name categorization
US11216718B2 (en) 2015-10-27 2022-01-04 Yardi Systems, Inc. Energy management system
US10274983B2 (en) 2015-10-27 2019-04-30 Yardi Systems, Inc. Extended business name categorization apparatus and method
US10275708B2 (en) 2015-10-27 2019-04-30 Yardi Systems, Inc. Criteria enhancement technique for business name categorization
US10275841B2 (en) 2015-10-27 2019-04-30 Yardi Systems, Inc. Apparatus and method for efficient business name categorization
US10915808B2 (en) * 2016-07-05 2021-02-09 International Business Machines Corporation Neural network for chemical compounds
JP6783646B2 (en) * 2016-12-21 2020-11-11 ブリヂストンスポーツ株式会社 Selection support device, selection support system and selection support method
US10817042B2 (en) * 2018-09-27 2020-10-27 Intel Corporation Power savings for neural network architecture with zero activations during inference
US11743272B2 (en) * 2020-08-10 2023-08-29 International Business Machines Corporation Low-latency identification of network-device properties
WO2022183098A1 (en) * 2021-02-26 2022-09-01 Ge Wang Machine learning for individual moral decision-making
US11270214B1 (en) 2021-03-29 2022-03-08 Isidore Samuel Sobkowski Providing the basis for ethical AI through explanations by coupling non-interpretable and interpretable systems
CN115331732B (en) * 2022-10-11 2023-03-28 之江实验室 Gene phenotype training and predicting method and device based on graph neural network

Family Cites Families (21)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4730259A (en) * 1985-03-01 1988-03-08 Gallant Stephen I Matrix controlled expert system producible from examples
JPS6326726A (en) * 1986-07-21 1988-02-04 Toshiba Corp Information processor
US4860213A (en) * 1987-10-01 1989-08-22 General Electric Company Reasoning system for reasoning with uncertainty
US5092343A (en) * 1988-02-17 1992-03-03 Wayne State University Waveform analysis apparatus and method using neural network techniques
US5014219A (en) * 1988-05-06 1991-05-07 White James A Mask controled neural networks
US5214745A (en) * 1988-08-25 1993-05-25 Sutherland John G Artificial neural device utilizing phase orientation in the complex number domain to encode and decode stimulus response patterns
US5003490A (en) * 1988-10-07 1991-03-26 Hughes Aircraft Company Neural network signal processor
US4958939A (en) * 1988-11-14 1990-09-25 Honeywell Inc. Centering scheme for pattern recognition
JPH02136904A (en) * 1988-11-18 1990-05-25 Hitachi Ltd Motion controller containing its own producing function for action series
US4914603A (en) * 1988-12-14 1990-04-03 Gte Laboratories Incorporated Training neural networks
US5073867A (en) * 1989-06-12 1991-12-17 Westinghouse Electric Corp. Digital neural network processing elements
JP2786275B2 (en) * 1989-11-22 1998-08-13 株式会社日立製作所 Optical neural network optical system
JP2671158B2 (en) * 1990-02-09 1997-10-29 株式会社 東洋製作所 Artificial snowfall equipment
JPH03250243A (en) * 1990-02-27 1991-11-08 Toshiba Corp Neural circuit network arithmetic unit
AU7563191A (en) * 1990-03-28 1991-10-21 John R. Koza Non-linear genetic algorithms for solving problems by finding a fit composition of functions
US5052043A (en) * 1990-05-07 1991-09-24 Eastman Kodak Company Neural network with back propagation controlled through an output confidence measure
US5218529A (en) * 1990-07-30 1993-06-08 University Of Georgia Research Foundation, Inc. Neural network system and methods for analysis of organic materials and structures using spectral data
US5121467A (en) * 1990-08-03 1992-06-09 E.I. Du Pont De Nemours & Co., Inc. Neural network/expert system process control system and method
US5241620A (en) * 1991-01-03 1993-08-31 Promised Land Technologies, Inc. Embedding neural networks into spreadsheet applications
US5226092A (en) * 1991-06-28 1993-07-06 Digital Equipment Corporation Method and apparatus for learning in a neural network
US5251268A (en) * 1991-08-09 1993-10-05 Electric Power Research Institute, Inc. Integrated method and apparatus for character and symbol recognition

Also Published As

Publication number Publication date
US5396580A (en) 1995-03-07
US5438644A (en) 1995-08-01
WO1993005473A1 (en) 1993-03-18
AU2505892A (en) 1993-04-05

Similar Documents

Publication Publication Date Title
US5438644A (en) Translation of a neural network into a rule-based expert system
Mitra et al. Neuro-fuzzy rule generation: survey in soft computing framework
Inuiguchi et al. Rough set theory and granular computing
JP5395241B2 (en) A method for determining whether an input vector is known or unknown by a neuron
Balakrishnan et al. Intelligent diagnosis systems
Yasdi Combining rough sets learning-and neural learning-method to deal with uncertain and imprecise information
Ganivada et al. Fuzzy rough granular neural networks, fuzzy granules, and classification
Su et al. Application of neural networks incorporated with real-valued genetic algorithms in knowledge acquisition
Boutsinas et al. Artificial nonmonotonic neural networks
Rabuñal et al. A new approach to the extraction of ANN rules and to their generalization capacity through GP
Ledesma et al. Feature selection using artificial neural networks
Hartley Accuracy-based fitness allows similar performance to humans in static and dynamic classification environments
Basak et al. A connectionist model for category perception: theory and implementation
Healy A topological semantics for rule extraction with neural networks
Boz Converting a trained neural network to a decision tree dectext-decision tree extractor
El-Sharkawi Neural network and its ancillary techniques as applied to power systems
Bull Model-based evolutionary computing: A neural network and genetic algorithm architecture
Avner Extraction of comprehensive symbolic rules from a multi-layer perceptron
Avner Discovery of comprehensible symbolic rules in a neural network
Yoo Symbolic rule extraction from artificial neural networks
Zhao Inducing NNC-trees with the R/sup 4/-rule
Singal et al. Incorporation of Fuzziness in ID3 and Generation of Network Architerture
Etchells Rule Extraction from Neural Networks: A practical and efficient approach
De Mingo et al. Hierarchical function approximation with a neural network model
Dazeley To The Knowledge Frontier and Beyond

Legal Events

Date Code Title Description
EEER Examination request
FZDE Discontinued