WO2005020132A1 - Neural network for processing arrays of data with existent topology, such as images, and application of the network - Google Patents

Neural network for processing arrays of data with existent topology, such as images, and application of the network Download PDF

Info

Publication number
WO2005020132A1
WO2005020132A1 PCT/EP2004/051821 EP2004051821W WO2005020132A1 WO 2005020132 A1 WO2005020132 A1 WO 2005020132A1 EP 2004051821 W EP2004051821 W EP 2004051821W WO 2005020132 A1 WO2005020132 A1 WO 2005020132A1
Authority
WO
WIPO (PCT)
Prior art keywords
knot
value
output
net
function
Prior art date
Application number
PCT/EP2004/051821
Other languages
French (fr)
Inventor
Paolo Massimo Buscema
Original Assignee
Semeion
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Family has litigation
First worldwide family litigation filed litigation Critical https://patents.darts-ip.com/?family=34043032&utm_source=google_patent&utm_medium=platform_link&utm_campaign=public_patent_search&patent=WO2005020132(A1) "Global patent litigation dataset” by Darts-ip is licensed under a Creative Commons Attribution 4.0 International License.
Application filed by Semeion filed Critical Semeion
Priority to EP04766522A priority Critical patent/EP1656632B1/en
Priority to US10/569,354 priority patent/US7877342B2/en
Priority to JP2006523636A priority patent/JP2007503629A/en
Publication of WO2005020132A1 publication Critical patent/WO2005020132A1/en

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/20Image enhancement or restoration by the use of local operators
    • G06T5/60
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • G06V10/443Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components by matching or filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S706/00Data processing: artificial intelligence
    • Y10S706/902Application using ai with detail of the ai system
    • Y10S706/924Medical

Definitions

  • the present invention relates to a neural network for processing arrays of data with pertinent topology comprising a n-dimensional array of cells (Ki) corresponding to the knots of the neural network, each 15 cell having connections to the directly adjacent cells (Kj) forming the neighbourhood of the a cell (Ki) ; a) Each cell (Ki) having an input for each connection to a directly adjacent cell of the surrounding cells (Kj) ; 20 b) each cell (Ki) having an output for the connection to one or more of the directly adjacent cells (Kj); c) the connection between each cell (Ki) and the directly adjacent cells being determined by weights 25 (wij); d) each cell being characterised by an internal value defined as the activation value or function (Ai) of the cell (Ki) ; e) each cell (Ki
  • the present invention apply to the field of artificial intelligence and hence to machines having a computational unit which is able to carry out simple processes as for example learning processes from empiric experience, deductive processes, cognitive processes by means of which collected or inputted data is analysed for discovering or investigating certain relationships between the data records which at a first superficial glance may not appear evident or recognition processes by means of which voices , patterns, figures, letters or the like are recognised for further processing. All the above mentioned processes are useful in order to put the machine in a condition to be able to take decisions on certain reactions or for simple classification aims of the data collected or inputted for example for further use.
  • non supervised algorithm i.e. Self Organising Map, which as an output furnishes a grid having a certain numbers of units each one individuated by a cell and in each grid being collected a certain number of the data records belonging to a certain prototype of data record.
  • SOM Self Organising Map
  • This algorithm however are not very effective in helping for recognising relationships of certain type between data records particularly data records where the relative position of the data records in an array of data record or in a distribution of data records in a N-dimensional space, particularly a two or three dimensional space is a relevant feature of the data record.
  • the data records are passive elements of the process .
  • Different kind of traditional artificial neural networks can be used.
  • This artificial neural networks are characterised by knots.
  • the knots are processing cells which are connected with each other in order to for a network.
  • the artificial neural networks are modelled on the neuron neytworks of the brain. In this case, each knot of the network representing an artifical neuron.
  • the knots are arranged in layers .
  • a neuron Ina a simplest configuration an inut layer of knots being connected with an output layer of knots. The number of knots corresponding normally to the different datarecords or variables of a database.
  • a neuron comprises three essential parts : a neuron cell body, branching extensions called dendrides for receiving input and an axon that carries the neuron's output to the dendrides of other neurons.
  • a neuron sends its output to other neurons via its axon.
  • An axon carries information through a multitude of action potentials, or waves of current, that depends on the neuron's potential. This process is ofte modelled as a propagation rule represented by a net value.
  • a neuron collects signals at its synapses by summing all the excitatory and inhibitory influences acting on it. If the excitatory influences are dominat, then the neuron fires and sends this message to other neurons via the outgoing synapses . In this sense the neuron function can be modelled as a simple threshold function.
  • each knot has inputs connected to the output of some or each other knot of a preceding layer of knots and an output connected to some or each other knots of a subsequente layer of knots .
  • the excitation or inhibition level exercised by the outputs of other knots connected to the input of a knot is determined by a connection strength which is defined by weights.
  • connection weights defining the connection strength artifical neural network are subjected to a training process in which the data of a data base are inputted for which data the processing output data are known.
  • the network is fed with the input data and with the known aoutput data and the connection weights are computed such that the given input and output data are matched by the weights .
  • a cellular neural network can be characterized by templates of its feed-abck and feed forward operators. These operators defines the dynamic behaviour of the cellular neural network. These operators are finite constants or square matrices of coefficients , so called cloning template which defines the dynamic rule of the cellular neural network .
  • cloning template which defines the dynamic rule of the cellular neural network .
  • different kind of operators can be used which are predefined and independent of the particular values of the data of the array of data to be processed.
  • Each operatore being specificly defined in order to carry out a particular operation of the data for extracting or highlighting features from the data or relations among the data.
  • a library of such operator templates so called genes, is provided from which one or more operator templates are chosen and used to carry out the data processing desired .
  • the weight defining the signal propagation to the input of a knot from the output of a knot of the directly surrounding layer of knots are predefined and independent from the internal values of the knots which corresponds to tha activation values or to the output values of a knot in an artificial neural network.
  • the invention aims to provide for an improved artificial neural network which combines the advantages of the structure of a cellular neural network or of a cellular automata with the advantages of a dynamic definition of the operators which also take into account the information which is intrinsically contained in the relation fo the values of the data in an array data.
  • the invention aims to provide a new artificial neural network having the structure of the cellular neural network and which allows parallel processing of image data without losing the information related to the relation between the values of the pixels forming the image.
  • the invention ahieves the above goals by means of an artifical neural network comprising a n-dimensional array of cells (K ⁇ ) corresponding to the knots of the neural network, each cell having connections to the directly adjacent cells (Kj) forming the neighbourhood of the a cell (K ⁇ ) ; a) Each cell (K ⁇ ) having an input for each connection to a directly adjacent cell of the surrounding cells (K j ) ; b) each cell (K ⁇ ) having an output for the connection to one or more of the directly adjacent cells (K j ); c) the connection between each cell (K x ) and the directly adjacent cells being determined by weights d) each cell being characterised by an internal value defined as the activation value or function (Ai) of the cell (Ki)
  • the weights (w Aj ) defining the connection between each cell (K ⁇ ) and the directly adjacent cells (Kj) are determined as the function of the current values (U j ) of the variables of each data record (Pj) univocally associated to the cell (K j ) directly adjacent to the said cell (Ki) , the said function being a so called learning function or rule; 1) the current activation value (Ai) or the output value (Ui) of each cell (Ki) after a processing steps of the neural network which is considered as the current new value (Ui) for the said univocally associated data records (Pi) being determined as a function of the current output values (UJ) of the directly adjacent cells (K j ) weighted by the corresponding weight (w ⁇ j ) defining the connection of the directly adjacent cells (K j ) with the cell (K ⁇ ) .
  • the above mentioned aritifial neural network can be modified by determining the current activation value (A ⁇ ) or the output value (Ui) of each cell (Ki) after a processing steps of the neural network which is considered as the current new value (Ui) for the said univocally associated data records (Pi) as a function of the of the weights (Wi j ) defining the connection of the directly adjacent cells (K j ) with the cell (Ki) , the said function being a so called activation function or rule.
  • the current activation value (Ai) or the output value (Ui) of each cell (Ki) after a processing steps of the neural network which is considered as the current new value (Ui) for the said univocally associated data records (Pi) can be determined as a function of the current output values (U j ) of the directly adjacent cells (K j ) and of the corresponding weight (w ⁇ ) defining the connection of the directly adjacent cells (K j ) with the cell (K ⁇ ) , the said function being a so called activation function or rule.
  • the weights (w ⁇ j ) defining the connection between each cell (Ki) and the directly adjacent cells (K j ) are determined as the function of the current values (Uj) of the variables of each data record (Pj) univocally associated to the cell (Kj) directly adjacent to the said cell (K ⁇ ) and of the current value (Ui) of the variables of the data record (Pi) univocally associated to the cell (Ki) .
  • the above artificial neural network according to the present invention is based on the following theoretical founding:
  • a phenomenon is defined as each space-time set which can be measured in terms of adsorbed or emitted electromagnetic frequencies .
  • Visual phenomenon are thus a sub set of the universe of the phenomenon as defined above and which sub set varies depending on the frequencies which one desides to observe.
  • a visual phenomenon is a phenomenon having a pertinent topology. Any phenomenon can be analyzed by means of elements which are a priori determined by a model or the phenomenon can be forced to show the relations existing between minimum elements . Any element of a phenomenon is an explicit or implicit synthesis of minimum elements. Thus an element of a model can be considered as an index reducing the original information being present in the minimum elements .
  • any kind of scientific model has the aim to let appear for every phenomenon ist inherent model by means of the interaction of minimum elements of the phenomenon. Having found the minimum elements of a phenomenon the scientific model should furnish to the said elements equations by means of which the said elements interact one with the other. When by means of the said equation the minimum elements of a phenomenon are able to reconstruct the phenomenon itself in its morphological and dynamical complexity, then the equations can be said to be a good meta-model of the model which the minimum elements has caused. Without a recnstruction proof there is no validation of the scientific activity. Every phenomenon in order to be subject of scientific knowledge must be able to be quantitatively characterized. This means that its minimum elements and their relations must be able to be represented by numerical values .
  • ⁇ t -J ⁇ ⁇ ⁇ - ⁇ 2 ) 2 + ⁇ y ⁇ - 2f + ! (3 - 4)
  • a phenomenon having a pertinent topology is formed by minimum elements and by local connections among the said minimum elements.
  • Each minimum elemnt is deifned by the position which it occuopies in the phenomenon and by a quantitative value which indicates at each instant for example a charatersitic effect of the phenomeno such as for example the quantity of light being emitted or reflected by a reflector.
  • Each local connection defines the strength of the mutual influence of each minimum elemnt on each other minimum element which is directly adjacent to the said minimum element at each time instant.
  • X / y « °s is the position of the source minimum element
  • T is the time instant and k,z D d is the target minimum element.
  • W (x'ly D, Equation ($a) and (4b) describes the evolution of the minimum elements and of the connections of a phenomenon with a pertinent topology towards an attracting element which represents the group of its solutions .
  • Both equations can be presnt in two degenerated ways which request a particular attention: As a first way LA) - réelle • ' . , •"H D ⁇ ⁇ , 0 ,>> ) (4t > In this case the minimum elements of the phenomenon evolve only as a function of the connections, by starting from their original or starting quantitative value which has the only effect of an impulse of the process .
  • One particular feature of this evolution consit in the fact that the process regulating the evolution of the connections takes place in the space of the connections (equation (4b) .
  • the invention realtes also to an algorithm for recognising relationships between data of a database which algorithm can be more effective, rapid and more precise in highlighting the relationships between the data of a database.
  • the algorithm according to the invention is an algorithm for recognizing relationships between data of a database having pertinent topology, this meaning that the data are of the kind where the relative position of the data records in an array of data records or in a distribution of data records in a N-dimensional space, particularly a two or three dimensional space is a relevant feature of the data record and where the data records can be represented as an array of cells or points, each point being univocally related to a data record of the database and having a univocally defined position in the array relatively to the cells or points of the other data records , to each data record being further associated at least one variable or more variables each one having certain value, the algorithm being characterised by the fact that - each cell or point in the array of cells of points representing a data records of a database is considered to be a unit or a knot of an Artificial Neural Network; - each unit or knot formed by a cell or point of the database being successively defined as a target unit or knot and connections being defined between each target unit or knot at least to each one of the rest of
  • each unit or knot of the database successively defined as target unit or knot being calculated by means of the set of learning rules or functions defining the strength of the connection between the units or knots or the set of activation rules or functions of the artificial neural network defining the quantitative value of each unit or knot i.e. its activation value or its output value, or by means of the combination of both the set of learning rules or functions and the set of activation rules or functions of the artificial neural network as a function of the actual output of the units or knots connected to the target unit or knot and of the actual output of the said target unit or knot;
  • each unit or knot being defined as the value of the variable or as the values of the variables associated to each data record represented by a cell or a point considered as a unit or knot of the artificial neural network;
  • the new output of the target unit or knot is considered as the new value of the variable or of the variables of the data record associated to the cell or point of the array of data records corresponding to the target unit or knot;
  • a new array of data records is computed where the data record of each cell or point has a new value of the at least one variable or new values for the variables as a result of a first computation cycle of the artificial neural network according to the above steps;
  • the said computation cycle being repeated for each successive new array of data records until a certain prefixed number of repetitions of the computation cycle has been carried out and/or unless a certain maximum allowable error or discrepancy has been reached between the original values
  • the array of data records may be submitted at least twice or more times to elaboration according to the algorithm of the present invention, in a first phase being provided a first set of learning functions or rules or a first set of activation functions or rules or a combination thereof and in a second phase being provided a second set of learning functions or rules or a second set of activation functions or rules or a combination thereof and so on when more then two phases are provided while the array of data-records being used in the second or further phases of elaboration with the algorithm of the present invention with the second or further different sets of learning or activation rules or functions or with a combination thereof is the array of data records resulting respectively from the first or from the previous phase of elaboration of the array of data records .
  • the units or knots corresponding to the cells or points related to the data records at least of gradient 1 with respect of the unit or knot actually formed by the cell or point related to the target data record be the ones spatially directly surrounding the cell or point related to the said actual target data record in a special case particularly of a two or three dimensional array of data records the cells or points related to the data records of gradient 1 with respect to the cell or point of the target data record are formed by the data records associated to the cells or points of the data record array which directly surrounds the cell or point in the array of data record related to the said target data record.
  • the new output of each target data record is defined as a function of new weights characterising the connection of each target unit or knot associated to the target data record with the units or knots represented by the cells or points of data records of gradient one relatively to the cell or point of the target data record, the set of learning rules or functions defining new weights of the connections as a function of the previous weights computed or defined in a previous computation cycle and as a function of the actual outputs of the unit or knot associated to the cell or point of target data record and of the unit or knot associated to the cells or points of the data records of gradient 1 or of the data records of the cells or points directly surrounding the cells or point of the actual target data record.
  • a similar result may be reached by using sets of activation functions where these functions defines the new output of the target unit or knot corresponding to the cell or point related to the target data record basing on the net input to the said target unit or knot which is a function of the outputs of the units or knots corresponding to the cells or points associated to the data records of gradient 1 with respect to the target data record, particularly to the units or knots corresponding to the cells or points of the array of data records directly surrounding the cell or point of the target data record.
  • the learning phase starts with a fixed predetermined value of the weights for each connection while the starting value of the unit or knot is modified according to a predetermined function which is also function of the weights and of the value of the surrounding knots or units and therefore to the data records which corresponds to cells or points in the array directly surrounding the cell or point representing a certain unit or knot of the artificial neural network .
  • a new data array is thus constructed in which each data record has maintained the position with respect to the other data records in the representation of the data records as cells or points in an array while each data record has changed its value as a function of its original value and of the original values of the data records relative to the surrounding points in the array .
  • a new cycle can thus be executed in which again each point representing a data record is set as a unit or knot of an artificial neural network, a weight is defined for each connection between units and a new value in the form of an output of each unit or knot is calculated according to the certain function.
  • Several different kinds and structures of known artificial neural network using different known and/or new learning functions for the definition of the weights and different functions for the calculation of the new values of each data record associated to each unit or knot may be used. It is important to notice that since the weights of the connections between the knots of the artificial neural network are fixed at least at each cycle and eventually equal for each connection, the algorithm according to the present invention has practically no internal noise and the noise is given only by the noise of the data of the database.
  • the algorithm according to the present invention can be used for image pattern recognition and for evidencing the different kinds of structures of the material forming a body, particularly of the tissues forming a biologic body in combination with radiological, echographic or magnetic resonance imaging or the like.
  • the algorithm according to the present invention can be useful evidencing different kinds of tissues and/or for substituting contrast medium in the diagnostic imaging.
  • the different kind of revealed or enhanced information depends by the different set of learning and or activation rules applied to the neural network structure formed by the pixel array of the original image. Principally the effect obtained is similar to the one which can be obtained by using enhancing media such as contrast media in the echographic, radiographic, nuclear magnetic resonance imaging or in similar imaging methods .
  • the kind of information enhanced can furthermore be differentiated by the choice of a certain set of learning and/o activation rules.
  • the two families are summarized in the table of figure 3 in which the different sets of rules are given a name which will be explained later on in the detailed description.
  • a first family of rules provides for weight evolution and unit activation and allow to carry out an extraction of features from a source digital or digitalized image such as for example kind of tissue in an image of biological bodies or structures of tissues or objects like for example recognizing the presence of stenosis in angiographic images .
  • a second family comprises rules which principally carry out a weight optimisation. After weight optimisation a further step of unit activation can be carried out.
  • the invention relates also to a method for image pattern recognition in digital or digitalized images .
  • the invention relates to a method for the recognition and differentiation of different tissues in digital or digitalized images of biologic bodies, as for example in diagnostic images as for example digitalized radiographies or echographic images or magnetic resonance images .
  • Fig. 1 is a schematic view of a two dimensional array of cells each one representing a data record in a two dimensional array of data, for example each cell representing schematically a point (a pixel) of a digital or digitalized image.
  • Fig. 2 represents the step of transforming the nine central points or cells of figure 1 in units or knots of an artificial neural network.
  • FIG. 3 illustrates a diagram of the different families of learning and activation rules for the neural network particularly referring to a neural network structure formed by the pixel of a digital image.
  • Fig. 4A, 4B, 4C, and 4D illustrate respectively a mammography, the enlarged mammography, the digitalized mammography which has been subjected to the method for image pattern recognition according to the present invention and according to a particular learning law called CM and the digitalized mammography which has been subjected to the method for image pattern recognition according to the present invention and according to further particular learning law called IAC.
  • Figures 5A, 5B, 5C, 5D are respectively the figure 4A and a further enlargement of the figures 4B, 4C, 4D.
  • Figures 6A, 6B, 6C, 6D, 6E, 6F, 6G are respectively a photograph of bacteria and four digitalized images of fig- 6A subjected to known pattern recognition methods while fig. 6G illustrates the digitalized image of figure 6A subjected to the method according to the present invention using automata rules for the artificial neural network called IAC particularly enhancing contrast.
  • Figures 7A, 7B, 7C are a digital photo of the ventricle, the said digitalized photo subjected to a known pattern recognition method and the digitalized photo subjected to a method according to the invention called CM.
  • Figures 8A, 8B, 8C, 8D and 8E show a set of echographic images of the same body part where metastases are present and where figures 8A to 8D are respectively an image without contrast media, with contrast media in the arterials phase, an image with contrast media in balance, an image with contrast media with in a late phase, the image of the same body part, without contrast media and treated with the method according to the present invention.
  • Figures 9A, 9B, 9C, 9D are respectively a radiographic image of the lung, the digitalized image of figure 8A subjected to a known algorithm called best filter, the image of fig. 8A subjected to a treatment according to the present invention and with an rule called CM, and the image of fig.
  • FIG. 8A subjected to a treatment according to the present invention and with an automotive rule called IAC.
  • Figures 10A, 10B, 10C illustrates respectively a mammography, the mammography of fig. 10A treated with a known pattern recognition method and the mammography of figure 10A treated according to the method of the present invention with an automotive rule called High CS .
  • Figures 11A, 11B, 11C, 11D and HE illustrate respectively a normal photograph, the photographs according to fig. HA treated by two different known patter recognition methods and the photograph of fig HA treated with the method according to the present invention with two different automotive rules called IAC and IAC combined with CM.
  • Figures 12A to 12J illustrates respectively a radiographic source image of the anatomical district of the femoral arteria and the image elaborated according to the present invention by using different kind of rules and different combination thereof.
  • Figure 13 is a schematic block diagram of a typical generalised cell or knot of the neural network according to the present invention.
  • a two dimensional array of data records is schematized by an array of square cells which are identified by Pi, where i are elements of Natural Numbers.
  • the central nine square cells P 7 , P ⁇ , 9 P12, P13, Pi4, and P ⁇ 7 , P i8 , P19, are drawn with continuous lines while the surrounding further cells are drawn with discontinuous lines for representing the fact that the array can be larger then the one represented.
  • Each cell of the array is a cell for a data record which has its univoquely related cell and so a univoquely related position in the array, the relative position of each cell of data record to the other cells of data record being a relevant feature of the corresponding data record.
  • Each data record of each cell has a value Ui , where i is element of the natural numbers and where the said value can be a parametric value of only a variable of the data record or a vector comprising each data record a set of different variables .
  • the above scheme is a representation of what is called in the specific technical language a database having a pertinent topology.
  • a digital or digitalized image is a two dimensional array of pixel.
  • Each pixel represents a data record in an array of data record and the relative position of the pixel is a relevant feature of each pixel belonging to the relevant information of the image data.
  • the value of the pixel is a simple intensity value in a grey scale image, while in a coloured image the value is a vector comprising at least two parameters or variables namely the colour and the intensity, normally three parameter or variables in the actually used techniques of visualizing coloured images as the so called HVS or the so called RGB coding of coloured images.
  • Figure 2 try to schematically picture the step used by the algorithm according to the present invention which provides for treating the data records of the two dimensional array by transforming the two dimensional array of cells in an array of cells or knots of an artificial neural network.
  • the squares represents the relation to the original cells P of figure 1
  • the enclosed circle Ki represents the transformation or association of each cell i to a unit or knot K ⁇ of an array of cells or knots of an artificial neural network.
  • connection arrows w ⁇ j represents the connections between the units or knots K ⁇ and K j of the artificial neural network.
  • Figure 2 is related only to the nine central cells of figure 1 and the connection arrows are continuous for the central cell P13, which is the cell corresponding to the actual target data record, while the discontinuous arrows are related to connection to peripheral cells which are obviously used by the present algorithm and the mechanism of which derives identically from the description of the way of working of the present algorithm as described with reference to the sole central cell P i3 .
  • the said algorithm sweeps the array cell by cell each time identifying a cell as the momentary central cell and the corresponding peripheral cells and the associated connections between the central cell and the peripheral cells and when the computation has been carried out for a certain cell momentary defined as a central cell , an adiacent cell is defined as the new central cell and new peripheral cells of this new central cells are identified with the corresponding connections, the computation being carried out this time for this new central cell and the new peripheral cells with the corresponding connections and so on. This mechanism is repeated until each cell of the array has been subjected to computation by the algorithm, thus ending a first computation cycle of the artificial neural network.
  • FIG. 13 illsutrates a schematic diagram of a cell or knot of the artificial neural network according to the present invention.
  • Each knot or cell Ki comprises inputs at which the output signals U j from other knots K j are received by the knot or cell Ki.
  • each knot or cell K ⁇ has an output Ui which is connected to the input of one or more other knots or cells K j .
  • each knot or cell K has also an input for an initialisation value defining the initialisation internal value or actication value A ⁇ of the knot or cell or the initialisation output value Ui.
  • the initialisation value for the internal value or activation value A ⁇ or the first output value Ui of the knot K is set equal to the value Ui of the data record Pi univocally associated as explained above to the knot Ki.
  • Each knot or cell Kj has further an output for the computed value U" for the datarecord P after the n-th processing cycle of a sequence of a certain number m of processing cycles.
  • the said computed new value U" corresponds either to the internal value or activation value A" of the knot or cell K ⁇ after the n-th processing cycle has been carried out or to the output value u" of the knot or cell Ki after the said n-th processing cycle has been carried out.
  • the present artificial neural network comprises direct connections of a knot or cell K only with the knots or cells Kj directly adjacent to the said knot or cell Ki.
  • the set of this knots or cells K j are technically defined as the local neighbourhood of the knot or cell K ⁇ or the knots or cells of gradient 1.
  • gradient 1 refers to the stepwise difference between the position of the reference knot or cell K ⁇ and the position of the set of knots or cells Kj forming the local neighbourhood of the said knot or cell Ki. This definition is intuitively understandable and is a widely spread technical definition. So it is clear that in a discretized array where only discrete positions are possible for the elements of the array, defining a reference element the gradient 1 neughbourhood is formed by the set of elements which position differs from the position of the reference element by a step 1 in any direction of the array. This is the simplest case.
  • the term gradient 1 is more generic since it is not limited to the spatial coordinates, but can be referred to any kind of space in which the an n-dimensional database is projected.
  • connection strength is quantitatively given by weights j i or Wi j which are multiplied with the output U j of the knots or cells K j of the local neighbourhood of the knot K ⁇ , when the signals of these knots of the local neighbourhood are send to the reference knot K ⁇ connected to them.
  • weighted signals inputted to the reference knot Ki will have an effect on its internal state Ai in order to cause or inhibite the emission of an output signal n or to change the value of the said output signal Ui .
  • equations which defines the quantitative values of the weights are called learning rules
  • equations which changes the internal state or the activation value or the output value of the knots as a function of the output values of the reference knot and of the adjacent knots of the knots of the locqal neighbourhood are called activation rules .
  • computation may be carried out by only considering a set of learning rules or functions which computes new outputs of the actual centre cell Pi, by means of defining new weights Wi j of the connections as a function of the outputs of the centre cell and of the surrounding cells, this means of the values of the variable or of the variables of the data record corresponding to this cells.
  • the new output of the centre cell will be then set as the new value or the new values of the variable or of the variables of the centre cell and will be a function of the outputs of the surrounding cells, this means of the values of the variable or of the variables of the data records corresponding to the said surrounding cells and of the new weights computed by the algorithm.
  • the weights ⁇ of the connections between the cells P ⁇ and P j used as knots Ki and K of the artificial neural network are set all equal to a certain value which can be randomly defined by a value for example 1 or 0.0001.
  • the cell P1 3 working as a unit or knot K i3 of an artificial neural network is changed as a function of the values of the surrounding cells j which are also set equal to knots K j of the neural network (in figure 2 this are the cells and knots P 7 , P 8 , P9, P12, Pm, and P i7 , P 18 , P19 and K 7 , K 8 , K 9 , K12, K14, and Ki 7 , Ki ⁇ , K 19 ) and of the weights of the connections Wi j .
  • Another way of computing the new outputs of the units or knot Ki is to use a set of activation functions or rules , which defines the output of the central knot K13 as a function of the values U j (in fig. 2 U 7 , U 8 , Ug, U12, Ui4, and U 17 , U ⁇ 8 , Ui9>of the data records of the cells Pj (in fig. 2 P 7 , P 8 , P 9 , P ⁇ 2 , P i4 , and P ⁇ 7 , P ⁇ 8 , P1 9 ) surrounding the central cell Pi (in fig. 2 P ⁇ 3 ) corresponding to the a unit or knot Ki (in fig.
  • the function defines the net input of the knot K i3 by means of the outputs U 7 , U 8 , Ug, Ui2, Ui4, and U ⁇ 7 , Ui ⁇ , U19 of the surrounding knots K 7 , K 8 , K 9 , K i2 , K14, and K1 7 , K ⁇ 8 , K 19 corresponding to the values of the associated cells P 7 , P ⁇ , P 9 , Pi2 f Pi4 f and P ⁇ , P ⁇ 8 , P19 surrounding the central cell P 13 and the associated knot K i3 .
  • a new array of cells has been constructed in which each data record has maintained its cell and where each data record has a new value for the variable or for the variables computed as described before in the first cycle while also new weights for the connections has been computed.
  • a second cycle can be carried out by using this new array of data records this time using the values of the data records computed in the first or previous cycle as outputs of the knots of the artificial neural network corresponding to the cells and where the new weights for the connection between the knots corresponding to the cells are used which has been computed in the first or previous cycle as described before.
  • again new outputs of the knots and thus new values of the data records are computed thus forming again a new array of data records in which the data records has updated values .
  • corresponding new weights for the connections between knots are computed to be used in a following cycle of the algorithm if this is provided or necessary.
  • the number (n) of following cycles can be determined as a fixed parameter.
  • the number of repetitions can be determined by an error or deviance calculation of the values of the new array of data records related to the original one by setting a maximum error or deviance. This calculation can be carried out by means of usual error or deviance calculation in the field of the artificial neural networks .
  • Another criterion might consist in the fact that the difference of the sequence of new values for the variables of the data records is monitored and the repetition of the algorithm is stopped when the differences of the said values in the sequence of cycles is not anymore significant, this mean that it is less than a predetermined value. This means also that the new values does not differ substantially from the previous one.
  • a first family comprises rules which principally carry out a weight optimisation which can further be combined with a unit activation. This kind of rules helps in finding out edges between different zones of an image for recognizing boundaries, for example boundaries between different kind of tissues in images of biological bodies .
  • Wit lJ e Rti J + e -*i J i
  • Ri j is some sort of measure for the distance of the i-th unit from the j-th unit.
  • Ui are the values of the single cells Pi transformed in knots Ki of the artificial neural network and where
  • the suffix i defines the central cell or knot and the suffix j defines the cells or knots directly surrounding the said central cell or knot.
  • this general rule may be further specialized or enhanced obtaining different set of rules .
  • the above disclosed automata rule defines the connection strength between each target or reference knot or cell K ⁇ with each one of the knots or cells K j of the local neighbourhood as the non linear projection of their different value Ui and Uj.Each reference knot or cell Ki is thus connected with eight symmetric weights W j to the knots or cells K j of the local neighbourhood of gradient 1. Every bidimensional database such as a bidimensional image will so be formed by a matrix in which each element of the matrix is connected as an autonomous unit to the eight elements of its local neighbourhood, i.e.
  • Each bidimensional matrix array of data such as an image will be further formed by a specific number of symmetric connections between each element of the matrix and the elements of the local neighbourhood.
  • the quantitative strength value of the connections W j iWhich are defined by the above mentioned equations forms the constraints by means of which the data arry such as an image, being in the form of an active matrix of the connection strength will cause the evolution in time of the initialisation value of each cell or knot till to the predetermined n-th processing cycle and towards the natural actractor.
  • a set of rule II is generally indicated as CS, i.e. Constraint Satisfaction which can be developed in different variants disclosed hereinafter: 11.1. so called CS XOR
  • Ui is also defined as input of the i-th knot or cell in combination with the initialization step. This definition wants to stress out that at the first step of carrying out the neural network the i-th unit has as an output value the value of the corresponding pixel in the digitalized source image. Obviously in the starting step of the neural network this value has to be attributed to the i-th knot or cell ad is considered as an input. In the following computational cycles the value of each knot or cell is changed by the algorithm and when a certain number of repetition has been carried out the value of the i-th knot or cell corresponds or is correlated to the value of the i-th pixel in the output image. Uj are the outputs of the surrounding knots K
  • Neti is the net input of the central Knot Ki computed as a function of the outputs U j of the surrounding knots K j and the weights w ⁇ j of the connections of the the surrounding knots K j to the central Knot K ⁇ .
  • n is the number of cycle
  • a t Net i ( ⁇ - u i ) a - (5 ⁇ --) Neti >0 1
  • a t Net j ⁇ «,• • ⁇ ⁇ ( ⁇ - -) Net t ⁇ 0
  • this case represents a combination of two of the previous examples of activation functions, namely the activation functions (3) called CS CONTEST and the activation functions (4) called High CS XOR.
  • a further special set of rules differing from the CS set of rules is the set of rules III identified by the name IAC, i.e. Interaction and Activation Competition IAC
  • N ⁇ n ⁇ i ⁇ U J ⁇ W; W,j ⁇ 0
  • a , (Max -U f ): Act z - - decay ⁇ (u , - rest ); Act ⁇ > 0
  • the net input Neti of the central knot K ⁇ is defined as a combination, more precisely as a sum of two different functions for computing the net input depending on the fact if the weights Aj of the connections are positive or negative. More precisely the functions ecc ⁇ and inii are identical only a selection is made between inputs associated to positive and to negative weight connections. This is expressed by the two functions ecci and inii. This two components of Neti are related to an excitation or enhancement in the case of the component ecci and to an inhibition in the case of inii .
  • the net input Neti is a combination of the ecci and inii in which ecci and ini ⁇ are multiplied respectively by a parameter ⁇ and by a parameter ⁇ .
  • the two parameters determines the influence of the positive weigt connections and of the negative weight connections on the net input Neti • Furthermore an activation parameter Acti is provided as a function of the net input Neti to the knot Ki. This parameter is used for choosing one of two different function for computing the update value ⁇ i of the output Ui of the knot Ki depending on the fact whether Acti is negative or positive. The enhancing and inhibiting clearly appears also from the two functions for computing ⁇ i since Acti is present as a parameter in the said functions.
  • the said functions further comprises the following expressions : Max and Min which are defined as the top and the bottom range of the activation value. Decay which is a function defined as the usual decay value along the time of each unit. Rest which is defined as the default value toward each unit tends for.
  • the new output value of the knot K is defined by a particular exponential function which further enhances the excitation and inhibition mechanism of this example of activation functions .
  • the above equations defines an Active matrix of the connections in which every pixel is an agent which receives dynamically excitation and inhibition impulses from the other pixels and which modifies correspondingly its internal state.
  • the above equation system can be defined as a collective auto-coordinating system between agent-pixels.
  • the pixels having an high luminosity tend to support oneselves, while the other pixels are drawn towards low values of luminosity.
  • the fixed connection weights between each agent-pixel and its local neighbourhood act as constraints which modulate the communication between each pixel .
  • the said communication is formed by exciting and inhibiting messages.
  • the evolution causes by discrete steps transformation of the original image that the pixel which are in a situation of a sudden change of luminosity will left isolated and are drawn to low values of luminosity.
  • the entire system will highlight the edges which are present in an image without using function which is explicitly designed for this task.
  • the pixel are in competition with each other some sort of walls are visualized which we often perceive as walls between different figures .
  • the advantage of the above system of ruels lies in the fact that the edge detection effect takes place due to a local auto-coordination. Furthermore also a function as an adapve threshold filter occurs . For each edge a line is represented having a thickness which corresponds to the difference in local luminosity which the image shows in every area.
  • a second family of rules provides for weight evolution and unit activation and allow to carry out an extraction of features from a source digital or digitalized image such as for example kind of tissue in an image of biological bodies or structures of tissues or objects like for example recognizing the presence of stenosis in angiographic images .
  • Two sub groups of sets of learning and activation rules are provided.
  • a first one indicated as the set V so called CM Contractive map can be defined as a recirculation neural network and comprises a learning phase and a recall phase. (for deeper information see Massimo Buscema & Semeion Group "Reti Neurali Arificiali esseli Sociali Complessi” Volume I, Ediée Franco Angeli Milano 1999) .
  • i is the value of the central cell or knot K ⁇ U j are the values of the surrounding knots K j Wi j indicates the weights of the connection of the surrounding knots K j and the central knot K ⁇ .
  • W, j defines the weights of the connection in the n-th cycle .
  • Outi is the output value of the i-th unit corresponding to the i-th pixel which is the target unit or pixel
  • New W j i is the new weight of the connection between j-th and i-th unit
  • New Wi j is the new weight of the connection between i-th and j -th unit w is the mean value of the weight.
  • the arguments rem and quot relates to a particular function applied to the value of each pixel , this means to the output of each unit U .
  • the value of each pixel and thus the output of each corresponding unit can be divided by the number of grey scale levels . If one considers for example 256 grey scale levels than the value of each pixel can be divided by this number.
  • this division gives rise to a certain reduced number of classes to which the pixels, i.e. the units belong and furthermore to a rest. The rest is indicated by variable rem.
  • This kind of operation allow to classify the units an thus the pixels in a certain reduced number of classes and to visualize each class in the output image by giving to the class a certain uniform colour or grey level.
  • the rest this means the rem value is further used to differentiate the intensity of the colour or grey level.
  • each class can be given a colour such as red, green, blue, yellow or four grey levels.
  • the rest of each pixel can be used for varying each colour between dark red or the light or pale red and so on.
  • the rem function as a function for modulating the values given to unit output and sou to the pixels of the output image within certain value range characterising each class defined by the quot function.
  • the quot and rem function apply also to pixels values corresponding to coloured pixels, where normally the pixel value is a vector providing three parameters as for example RGB definition of colured pixel values or HSV, or the like.
  • the pixels value and correspondingly the output value of the units of the image transformed in a neural network are normalized at the beginning for maintaining the possible output values within an interval between 0 and 1.
  • the output values are again denormalized, i.e.
  • the evolution law or rule of the weights defining the connections is in this set of rules of the deterministic type and the weights are intialized with values which are proximate to zero.
  • the new values of brightness for each pixel in every process cycle are defined in the space of the weights .
  • each value of each pixel is defined by the weights defining the connection of this pixel with the pixels of the local neighbourhood.
  • a further different set of rules which is the set VI and is defined as Local Sine LS belongs also to the above mentioned second family of set of rules .
  • the local sine set can be provided according to two variants called Local Sine 1 LSI and Local Sine 2 LS2 These two sets differs only in one equation and the set of rules is described by the following equations :
  • Di j represents some sort of distance of the i-th unit form the j-th unit.
  • is the standard variation.
  • different kinds of sets of learning rules or of activation rules or of combinations thereof will lead to different outputs which will enhance particular relationships between the data records of the array with respect to other relation ships .
  • the present algorithm does not consist in a non supervised mapping algorithm but transforms an array of data records in an active artificial neural network capable of working in a supervised or non supervised manner depending on the rules or functions for learning and activating.
  • a learning phase is present which normally does not occur in non supervised mapping algorithm.
  • Any kind of known or future artificial neural network may be applied for actuating the principle of the algorithm according to the present invention which is independent from the structure of the artificial neural network itself.
  • the result so obtained is obviously again an array of data records which may be subjected once again to computation with the algorithm according to the present invention this time choosing equal or different learning rules and a different set of activation functions.
  • an elaboration of the original array of data records is carried out by means of a sequence of two algorithm according to the present invention which are different relating to the learning rules and/or to the sets of activation functions .
  • the different sets of rules applied in the first computation cycle and in a second computation cycle does not belong necessary to only one family of set of rules, but the set of learning and activation rules of both families may be combined in as two, three or more stage algorithm for elaborating digital or digitalized images .
  • the effectiveness of the algorithm according to the invention will be shown by means of some practical examples.
  • the array of data records is formed by image date of a digital image or a digitalize image .
  • Each cell or point Pi of the schematic example of figures 1 and 2 is formed by a pixel of the image and the value of each cell of point Pi is formed by the intensity parameter, i.e. by the parameter influencing the level of grey of the pixel .
  • the examples could be also carried out for coloured image in this case the value of each pixel would have be defined by a set of parameter in the form of a vector defining the intensity and the colour.
  • Figures 4A illustrate a mammography taken according to traditional techniques.
  • Figure 4B is an enlargement of figure 4A. The traditional image of analogical type was digitalized with a scanner and treated with the algorithm according to the present invention.
  • Figure 4C illustrates the image resulting from the treatment of the mammography of figure 4B where the algorithm was applied using only the set of learning rules and activation rules defined above as CM.
  • the rule CM has the effect of enhancing contrast and putting into evidence image zones having equal intensity or level of grey scale by defining clear boredom lines .
  • Equal intensities or grey scale level are defined within a certain range of intensities or grey scale level in a discretized scale of intensities or grey levels.
  • the nodulus or the zone of higher vascolarization in figure 4B is highlighted by the treatment with the algorithm according to the invention and the boredom of this zones are sharply depicted in figure 4C.
  • Figures 5A to 5D are enlargement of this zone of the nodule to better evidence the clear boredom of the image zones furnished by the treatment of traditional mammography with the algorithm according to the present invention. It is also to notice that the image resulting from the treatment with the algorithm according to the present invention is able to extract also more information from the analogical mammography. Indeed the edges of the breast in the analogical image is not clearly differentiated from the background of the image. As it appears from figure 4C the algorithm according to the invention is able to set a more defined boredom clearly differentiating the background from the image of the breast. This is important in order to make measurement of the position of the nodule if an intervention has to be carried out, particularly a reduced invasive intervention by means of microsurgical techniques.
  • FIG. 4D and the corresponding enlargement of figure 5D illustrate the image obtained by a double treatment of the analogical image of figure 4C and 5C with the algorithm according to the present invention.
  • the first treatment is carried out by applying the algorithm using only the set of functions which has been defined above as CM.
  • the image data this means the array of image data obtained by this treatment is further subjected to a treatment with the algorithm according to the present invention where use is made of a combination of learning and activation functions and more precisely of the set defined above as Automata rules AR and of the set of functions defined above as IAC.
  • a treatment with the algorithm according to the present invention where use is made of a combination of learning and activation functions and more precisely of the set defined above as Automata rules AR and of the set of functions defined above as IAC.
  • an inversion of black and white zones has been obtained with respect to the image of figures 4C and 5C and partially with respect of the images of figures 4B and 5B.
  • the background is white instead of black.
  • the nodule is white as in the original analogical image of the mammography (figures 4B and 5B) .
  • FIG. 6A illustrates a analogical image of bacteria. The image has been taken with an electronic microscope.
  • the figure 6B to 6F illustrates the images resulting from the treatment of the image of figure 6A with different known image pattern recognition algorithms .
  • Figure 6G illustrates the image resulting from the treatment of image of figure 6A by means of the algorithm according to the present invention which is applied using only the set of rules named AR and the evolutionary rule named IAC as defined above.
  • figure 7A is a digital image of the ventricle of the heart.
  • Figure 7B illustrates the image treated by a known image recognition algorithm called Snake .
  • Figure 7C illustrates the image obtained by treatment of image of figure 7A with the algorithm according to the present invention using only the set of functions defined above as CM. Clear and well defined boredoms of the image zones are obtained and the different areas are well differentiated in the image.
  • Figures 8A to 8D illustrates ecographic images of metastases taken at different instants in relation to the instant of injection of so called contrast media, i.e. of substances capable of enhancing the revelation of vascular activity by reflecting the ultrasound beam in a non linear way, namely at a frequency different from the fundamental frequency of the impinging beam, normally at a frequency of the order of the second harmonics of the fundamental frequency of the impinging beam.
  • Figure 8A is an image taken at an instant before the contrast media are present in the imaged zone.
  • Figure 8B is an image taken during the arterial phase where the contrast medium has reached the imaged zone.
  • Figure 8C is an image of the same zone of figures 8A and 8B taken at an instant where the arterial and venal phase are at a balance.
  • Figure 8D is an image taken at a late instant relating to the injection of the contrast medium.
  • a white circle illustrates the metastases which appears as a darker spot relating to the surrounding image ones .
  • Figure 8E is the image resulting form the treatment of image 8A, i.e. before the injection of the contrast medium in the zone to be imaged, with the algorithm according to the present invention using the set of functions defined as CM in the above description .
  • Contrast media imaging also known as harmonic imaging in the field of ultrasound imaging requires also longer times to be carried out since after injection of the contrast media some time is needed until the contrast media reaches the tissue to be imaged. Furthermore there is also difficult to predict when the contrast media will reach the tissue to be imaged and sometimes it happens that no image can be taken at the correct instant coinciding with the presence of the contrast media in the tissue to be imaged.
  • the algorithm according to the present invention allows to overcome all the difficulties and drawbacks related to imaging with contrast media which are more than the most important ones disclosed above.
  • Figures 9A to 9D illustrates respectively an analogic image, namely a radiography of the lung.
  • FIG 10 illustrates three images 10A, 10B, 10C put one beside the other and related to a mammography.
  • the left hand figure 10A is the original mammography
  • the right hand image 10C is the original mammography, where some nodules are identified by highlighting them with white dots and encircling in white circles .
  • the comparison of the left hand image 10A with the right hand image 10C allow to appreciate the very tiny level of differentiation of the image of the nodules from the surrounding image zone.
  • the centre image of figure 10B is the result of the treatment of the digitalized figure 10A by means of the algorithm according to the present invention using only a set of activation rules defined in the above description as High CS. It appears evident that the algorithm has revealed and clearly highlighted in the obtained treated image the nodules which has been identified by the human eye and to this a further great number of nodules which where not apparent by human eye in the original image of figures 10A and IOC.
  • FIG. 11A to HE relates to an image of a landscape and not to a diagnostic image and is choosen in order to appreciate the accuracy of the algorithm according to the present invention differentiating image zones by maintaining the relevant structure of the original image.
  • Figure HA is a seascape image with a boat in the foreground.
  • the image may be an image acquired by means of digital techniques such as digital cameras or the like or an image which has been acquired by means of analogical techniques and which has been digitalized by means of a scanner or the like .
  • Figure 11B and 11C illustrates the image obtained by treating the image according to figure HA by means of two different known image elaboration algorithms .
  • Figure 11D illustrates the image according to figure HA after treatment with the algorithm according to the present invention, where only the set of functions has been used defined as IAC in the above description. In this image the subjects of the original image can be recognized much better than in the images of figures 11B and 11C and the different image zones are very good differentiated and recognized.
  • Figure HE illustrates the image obtained by the treatment of figure HA with the algorithm according to the present invention where a set of functions defined as CM in the above description has been used in combination with a set of functions defined as IAC in the above description. The result of this treatment is that the image zones having a certain impact on the viewer are enhanced.
  • the present algorithm can be applied in combination with other algorithm such as artificial neural networks or other prediction algorithm which are trained and tested to recognize kind of tissue or structure of the material composing and image .
  • this data instead of feeding to the image recognition algorithm using the image data array as passive information for training and testing the image data of the digital or digitalized original image, this data can be subjected previously to treatment by means of the present algorithm in one of the forms disclosed above.
  • the treatment of the image data of the original image by means of the algorithm according to the present invention can be carried out several times each time using a different set of learning functions or rules or a different set of activation functions or rules or a different combination thereof or using a sequence of treatment of the original image data in which each treatment phase uses different set of learning functions or rules or a different set of activation functions or rules or a different combination thereof.
  • the prediction algorithm being applied to the image data obtained by each treatment and then the results of each prediction may be combined together or compared one with the other.
  • Figures 12A to 12J illustrates further examples of elaboration of a same source image by means of the method according to the invention in which different set or rules or combination thereof are applied.
  • Figure 12A is the source image which is a radiography of the femoral arteria.
  • a circle and two ellipse indicated respectively with 1 , 2 and 3 encircles particulars of the vessels where a stenosis could be revealed. While the stenosis identified by the circle 1 appears in a sure way also in the source image, the zones encircled by the ellipses 2 and 3 does not give sure information .
  • the first image 12B is obtained by the method according to the invention in which the set of rules V using the rem function.
  • the constriction at 2 and 3 of the right hand branch of the arterial can be seen much better while also the central branch indicated by a rectangle 4 appear visible .
  • the structure at the interior of the arterial branches does not appear very clearly.
  • Figure 12C illustrate the image obtained by elaborating the source image with the set of rules V so called CM this time using only the option quot.
  • Four classes of grey levels can be seen indicating different zones , namely three grey levels and white .
  • the zones classified by quot does not show any further structure.
  • the constrictions at 1, 2, 3, and 4 can be already identified in the output image.
  • Figure 12D shows the result of applying rules V by taking into consideration both the quot and the rem function.
  • the four image zones having one of the four grey level of the image according to figure 12C are modulated by the rem function, namely the grey scale of image 12A.
  • the four image zones having different grey levels are further modulated by the rem function and so a structure can be recognized.
  • FIG. 12E illustrates the result of elaborating the source image 12A by means of the rules according to example V and by using the true color option, where the pixel values elaborated are maintained normalized within the interval between 0 and 1. also in this case it is possible to recognize the constrictions of the arterial at 1, 2 , 3 and 4.
  • Figure 12F is the image obtained by elaborating the source image with the local sine algorithm according to the first option. Also in this case the presence of the constrictions appears evident at 1 , 2 , 3 and 4.
  • the four figures 12G to 12J illustrates the result obtained by elaborating the source image 12A by a combination of two elaboration steps or stages each one carried out by means of the set of rules according to example V or VI .
  • Image of figure 12G is obtained by first carrying out an elaboration with the set of rules according to example VI second option, namely the so called LS2 set of rules and by submitting the output image obtained by this first elaboration to a second elaboration this time using the set of rules V so called CM applying the Rem and Quot functions.
  • Figure 12H illustrates a variant of the elaboration carried out for figure 12G in which the output image obtained by the elaboration with the set of rules according to example VI second option the so called LS2 set of rules is further elaborated with the set of rules according to example V, the so called CM, this time using only the rem function.
  • Fig. 121 is a similar way of elaborating the source image 12A as in figure 12H, this time the second elaboration stage is carried out by using the set of rules of example V by applying only the quot function.
  • Fig. 121 is a similar way of elaborating the source image 12A as in figure 12H, this time the second elaboration stage is carried out by using the set of rules of example V by applying only the quot function.
  • 12J is still another variant of the two stage elaboration according to the previous examples of figure 12G to 121 , where the set of rules of example V is used for the second elaboration, this time by applying the True Colour variant.
  • the images are all shown in a grey scale palette, it is possible to define colours and to correlate the different zones of the image to a specific colour, thus obtaining an artificially coloured image which better enhances the different zones of the image and the different objects identified by the elaboration according to the method of the present invention.

Abstract

A neural network for processing arrays of data with pertinent topology comprises a n-dimensional array of cells (Ki) corresponding to the knots of the neural network! each cell having connections to the directly adjacent cells (Kj) forming the neighbourhood of a cell (Ki) ; Each cell (Ki) having inputs for each connection to directly adjacent cells; an output for the connection to one or more of the directly adjacent cells (Kj); the connection between the cells being determined by weights (wij); each cell being characterised by an internal value and being able to carry out signal processing for generating a cell output signal (ui) ; the output signal (ui) of a cell (Ki) being a function of its interna value and of the input signals form the neighbouring cells; each cell being associated univoquely to a record of a n-dimensional database (Pi) with pertinent topology and the value of each data record being the starting value of the corresponding cell;. Processing is carried out by considering the internal value or the output value (ui) of each cell (Ki) after a certain number of iterative processing steps of the neural network as the new obtained value (Ui) for the said univocally associated data records (Pi).

Description

\L NETWORK FOR PROCESSING ARRAYS OF DATA WITH EXISTENT TOPOLOGY , SUCH AS IMA AND APPLICATION OF THE NETWORK
A neural network for processing arrays of data with pertinent topology, an algorithm for recognising relationships between data of a database and a method for image pattern recognition based on the said neural network and on the said algorithm. 10 The present invention relates to a neural network for processing arrays of data with pertinent topology comprising a n-dimensional array of cells (Ki) corresponding to the knots of the neural network, each 15 cell having connections to the directly adjacent cells (Kj) forming the neighbourhood of the a cell (Ki) ; a) Each cell (Ki) having an input for each connection to a directly adjacent cell of the surrounding cells (Kj) ; 20 b) each cell (Ki) having an output for the connection to one or more of the directly adjacent cells (Kj); c) the connection between each cell (Ki) and the directly adjacent cells being determined by weights 25 (wij); d) each cell being characterised by an internal value defined as the activation value or function (Ai) of the cell (Ki) ; e) each cell (Ki) being able to carry out signal 30 processing according to a signal processing function so called transfer function for generating a cell output signal (ui) ; f) the transfer function determining the output signal (ui) of a cell (Ki) as a function of the activation value or function (Ai) of the cell (Ki) , which transfer function comprising also the identity function which puts the activation value or function (Ai) of the cell (Ki) equal to the output signal (ui) of a cell (Ki) ; g) a n-dimensional database of input data records (Pi) being provided which has to be submitted to computation by means of the neural network and in which n-dimensional database the relative position of the data records (Pi) when projected in a corresponding n- dimensional space is a relevant feature of the data records (Pi) , the data records (Pi) of the database being able to be represented by an array of points in the said n-dimensional space, each point having an univocally defined position in the said array of points and being univocally related to a data record (Pi) of the said database, each data record (Pi) of the said database comprising further at least one variable or more variables each one having a certain value (Ui) ; h) each data record (Pi) being univocally associated to a cell (Ki) of the n-dimensional array of cells forming the neural network which cells (Ki) has the same position in the n-dimensional array of cells
(Ki) as the corresponding data record (Pi) represented by a point in the said n-dimensional array of points ; i) the value (Ui) of the variables of each data record
(Pi) being considered as the initialisation value of the network being taken as the initial activation value (Ai) or the initial output value (ui) of the univocally associated cell (Ki) ; j) the activation value (Ai) or the output value (ui) of each cell (Ki) after a certain number of iterative processing steps of the neural network being considered as the new obtained value (Ui) for the said univocally associated data records (Pi) .
The present invention apply to the field of artificial intelligence and hence to machines having a computational unit which is able to carry out simple processes as for example learning processes from empiric experience, deductive processes, cognitive processes by means of which collected or inputted data is analysed for discovering or investigating certain relationships between the data records which at a first superficial glance may not appear evident or recognition processes by means of which voices , patterns, figures, letters or the like are recognised for further processing. All the above mentioned processes are useful in order to put the machine in a condition to be able to take decisions on certain reactions or for simple classification aims of the data collected or inputted for example for further use. Actually, given a database in which data are in the form of records each one being identified by related values of a certain defined number of variables , the relationships between the data records can be investigated by means of the so called "non supervised algorithms". Known non supervised algorithm are for example the so called SOM, i.e. Self Organising Map, which as an output furnishes a grid having a certain numbers of units each one individuated by a cell and in each grid being collected a certain number of the data records belonging to a certain prototype of data record. The SOM is a known algorithm which is described in more details in KOHONEN, 1995: T. Kohonen, Self Organising Maps, Springer Verlag, Berlin, Heidelberg 1995 or Massimo Buscema & Semeion Group "Reti neurali artificiali e sistemi sociali complessi", Year 199, Edizioni Franco Angeli s.r.l. Milano, Italy, chapter 12. This clustering can give information about the similarity of the records one with the other and so allow to carry out data classifications or to recognize relationships which can be used by a machine for deciding how to carry put a task or if a task has to be carried out or which kind of task has to be carried out. This algorithm however are not very effective in helping for recognising relationships of certain type between data records particularly data records where the relative position of the data records in an array of data record or in a distribution of data records in a N-dimensional space, particularly a two or three dimensional space is a relevant feature of the data record. Furthermore the data records are passive elements of the process . Different kind of traditional artificial neural networks can be used. This artificial neural networks are characterised by knots. The knots are processing cells which are connected with each other in order to for a network. The artificial neural networks are modelled on the neuron neytworks of the brain. In this case, each knot of the network representing an artifical neuron. The knots are arranged in layers . Ina a simplest configuration an inut layer of knots being connected with an output layer of knots. The number of knots corresponding normally to the different datarecords or variables of a database. In the biological case a neuron comprises three essential parts : a neuron cell body, branching extensions called dendrides for receiving input and an axon that carries the neuron's output to the dendrides of other neurons. Generally speaking a neuron sends its output to other neurons via its axon. An axon carries information through a serie of action potentials, or waves of current, that depends on the neuron's potential. This process is ofte modelled as a propagation rule represented by a net value. A neuron collects signals at its synapses by summing all the excitatory and inhibitory influences acting on it. If the excitatory influences are dominat, then the neuron fires and sends this message to other neurons via the outgoing synapses . In this sense the neuron function can be modelled as a simple threshold function. In the artificial neural networks the same model is used each knot has inputs connected to the output of some or each other knot of a preceding layer of knots and an output connected to some or each other knots of a subsequente layer of knots . The excitation or inhibition level exercised by the outputs of other knots connected to the input of a knot is determined by a connection strength which is defined by weights. If the sum of the signals inputted to a knot excides a certain threshold value the knot will fire and the output will send out a signal. The internal state or value of a knot is defined as an activation function. By processing data with this traditional kind of artificial neural networks, the data are fed to the knots of the input layer and the result of the process is furnished at the outputs of the knots of the output layer . For better and deeper understanding of the structure of artificial neural networks see "Reti Neurali Artificiali e sistemi social! complessi" Volume I, by Massimo Buscema & Semeion Group, Semeion Centro Ricerche, Franco Angeli Milano 1999 (ISBN 88-464-1682- 1) . For determining the weights defining the connection strength artifical neural network are subjected to a training process in which the data of a data base are inputted for which data the processing output data are known. The network is fed with the input data and with the known aoutput data and the connection weights are computed such that the given input and output data are matched by the weights . When considering processing of a database which has a pertinent topology, this that the data can be projected as points in a n-dimensional space , where the relative position of the points representing the data is a relevant feature of the data themselves , such as for example the two dimensional array of pixels forming an image, the above mentioned traditional algorithm do not consider the said topologic feature, for example the position of the pixel in the image in relation to the other pixels and furthermore the processing is not carried out in parallel for each pixel . A solution of this problem has been attempted by using so called cellular automata or their improvement as cellular neural networks. Document US 5,140,670 and document "Cellular Neural Networks: Application " by Leon o. Chua and Ling Yang, I.E.E.E. Trans. On Circuits & Systema vol. 35 (1988) Oct., No. 10, New York, NY, US discloses a combination of a so called cellular automata and neural networks which show the features of the artificial neural network disclosed at the beginning. This new kind of information-processing system is a large scale non linear analog circuit like neural networks , which circuits processes signal in real time. Like cellular automata it is made of a massive aggregate of regulalrly spaced circuits clones, called cells, which communicate with each other directly only through its nearest neighbors . Cells not directly connected together may affect each other indirectly because of the propagation effects of the continuous-time dynamics of cellular neural network. The cellular neural networks are able to carry out feed-back and feed-forward operations. The connection among cells are uniform and local . This means that a cellular neural network can be characterized by templates of its feed-abck and feed forward operators. These operators defines the dynamic behaviour of the cellular neural network. These operators are finite constants or square matrices of coefficients , so called cloning template which defines the dynamic rule of the cellular neural network . Thus in a cellular neural network different kind of operators can be used which are predefined and independent of the particular values of the data of the array of data to be processed. Each operatore being specificly defined in order to carry out a particular operation of the data for extracting or highlighting features from the data or relations among the data. Normally a library of such operator templates, so called genes, is provided from which one or more operator templates are chosen and used to carry out the data processing desired . So for example when considering a two dimensional image an operator or a gene can be provided for detecting and highlighting edges, a further operator or gene can be provided for sharpening and so one. The peratirs can be sequentially used for processing the data in order to obtain a combination of their effects on the output image . From the above it is clear that although the known cellular automata take into consideration the fact that the data are topologically pertinent as better defined above, nevertheless the operators are made by constants and are completely independent from the values of the data to be processed. Comparing this behaviour to a neural network, this means that the weight defining the signal propagation to the input of a knot from the output of a knot of the directly surrounding layer of knots, are predefined and independent from the internal values of the knots which corresponds to tha activation values or to the output values of a knot in an artificial neural network. Thus the intrinsic information contained in the array of data due to their topologic relation ship and to their values is lost or completely not considered. The invention aims to provide for an improved artificial neural network which combines the advantages of the structure of a cellular neural network or of a cellular automata with the advantages of a dynamic definition of the operators which also take into account the information which is intrinsically contained in the relation fo the values of the data in an array data. As a special case the invention aims to provide a new artificial neural network having the structure of the cellular neural network and which allows parallel processing of image data without losing the information related to the relation between the values of the pixels forming the image. The invention ahieves the above goals by means of an artifical neural network comprising a n-dimensional array of cells (K±) corresponding to the knots of the neural network, each cell having connections to the directly adjacent cells (Kj) forming the neighbourhood of the a cell (K±) ; a) Each cell (K±) having an input for each connection to a directly adjacent cell of the surrounding cells (Kj) ; b) each cell (K±) having an output for the connection to one or more of the directly adjacent cells (Kj); c) the connection between each cell (Kx) and the directly adjacent cells being determined by weights
Figure imgf000010_0001
d) each cell being characterised by an internal value defined as the activation value or function (Ai) of the cell (Ki) ; e) each cell (K±) being able to carry out signal processing according to a signal processing function so called transfer function for generating a cell output signal (u±) ; f) the transfer function determining the output signal (ui) of a cell (Ki) as a function of the activation value or function (Ai) of the cell (Ki) , which transfer function comprising also the identity function which puts the activation value or function (Ai) of the cell (Ki) equal to the output signal (u±) of a cell (Ki) ; g) a n-dimensional database of input data records (Pi) being provided which has to be submitted to computation by means of the neural network and in which n-dimensional database the relative position of the data records (P±) when projected in a corresponding n- dimensional space is a relevant feature of the data records (Pi) , the data records (Pi) of the database being able to be represented by an array of points in the said n-dimensional space, each point having an univocally defined position in the said array of points and being univocally related to a data record (Pi) of the said database, each data record (Pi) of the said database comprising further at least one variable or more variables each one having a certain value (Ui) ; h) each data record (Pi) being univocally associated to a cell (K±) of the n-dimensional array of cells forming the neural network which cells (Ki) has the same position in the n-dimensional array of cells (Ki) as the corresponding data record (Pi) represented by a point in the said n-dimensional array of points; i) the value (Ui) of the variables of each data record (Pi) being considered as the initialisation value of the network being taken as the initial activation value (A) or the initial output value (UJ.) of the univocally associated cell (K±) ; j) the activation value (A±) or the output value (Ui) of each cell (K) after a certain number of iterative processing steps of the neural network being considered as the new value (Ui) for the said univocally associated data records (Pi) . which artificial neural network further comprises the following features: k) for each processing step of the said certain number of iterative processing steps, the weights (wAj) defining the connection between each cell (K±) and the directly adjacent cells (Kj) are determined as the function of the current values (Uj) of the variables of each data record (Pj) univocally associated to the cell (Kj) directly adjacent to the said cell (Ki) , the said function being a so called learning function or rule; 1) the current activation value (Ai) or the output value (Ui) of each cell (Ki) after a processing steps of the neural network which is considered as the current new value (Ui) for the said univocally associated data records (Pi) being determined as a function of the current output values (UJ) of the directly adjacent cells (Kj) weighted by the corresponding weight (w±j) defining the connection of the directly adjacent cells (Kj) with the cell (K±) . As a variant the above mentioned aritifial neural network can be modified by determining the current activation value (A±) or the output value (Ui) of each cell (Ki) after a processing steps of the neural network which is considered as the current new value (Ui) for the said univocally associated data records (Pi) as a function of the of the weights (Wij) defining the connection of the directly adjacent cells (Kj) with the cell (Ki) , the said function being a so called activation function or rule. The current activation value (Ai) or the output value (Ui) of each cell (Ki) after a processing steps of the neural network which is considered as the current new value (Ui) for the said univocally associated data records (Pi) can be determined as a function of the current output values (Uj) of the directly adjacent cells (Kj) and of the corresponding weight (w±) defining the connection of the directly adjacent cells (Kj) with the cell (K±) , the said function being a so called activation function or rule. Furthermore for each processing step of the said certain number of iterative processing steps, the weights (w±j) defining the connection between each cell (Ki) and the directly adjacent cells (Kj) are determined as the function of the current values (Uj) of the variables of each data record (Pj) univocally associated to the cell (Kj) directly adjacent to the said cell (K±) and of the current value (Ui) of the variables of the data record (Pi) univocally associated to the cell (Ki) . When the database is formed by the pixels of a two dimensional image, then the above neural network forms a machine for image processing according to the present invention. Further improvements are disclosed in the following description and are subject of the dependent claims . The above artificial neural network according to the present invention is based on the following theoretical founding: As an example a phenomenon is defined as each space-time set which can be measured in terms of adsorbed or emitted electromagnetic frequencies . Visual phenomenon are thus a sub set of the universe of the phenomenon as defined above and which sub set varies depending on the frequencies which one desides to observe. A visual phenomenon is a phenomenon having a pertinent topology. Any phenomenon can be analyzed by means of elements which are a priori determined by a model or the phenomenon can be forced to show the relations existing between minimum elements . Any element of a phenomenon is an explicit or implicit synthesis of minimum elements. Thus an element of a model can be considered as an index reducing the original information being present in the minimum elements . This reduction can be critical since it is possible that the elements of the model ignore some properties which is present in the minimum elements and in their local interactions . Analysing visual phenomenons, or other phenomena having a pertinent topology, the reduction carried out by the elements of the model is serious . Indeed some index as for example mean values, variance and every other index basing on a mean value will non maintain the topological properties of the minimum elements being synthetized. For example the mean value of two rectangular triangles will not generate a rectangular triangle except in very rare cases . Further every phenomenon which we can observe as a subject of scientific knowledge is a phenomenon having a pertinent topology. Due to the only fact that a shape can exist,in a theoretically isotropic space every shape shows with its topology its specific pertinence . Any kind of scientific model has the aim to let appear for every phenomenon ist inherent model by means of the interaction of minimum elements of the phenomenon. Having found the minimum elements of a phenomenon the scientific model should furnish to the said elements equations by means of which the said elements interact one with the other. When by means of the said equation the minimum elements of a phenomenon are able to reconstruct the phenomenon itself in its morphological and dynamical complexity, then the equations can be said to be a good meta-model of the model which the minimum elements has caused. Without a recnstruction proof there is no validation of the scientific activity. Every phenomenon in order to be subject of scientific knowledge must be able to be quantitatively characterized. This means that its minimum elements and their relations must be able to be represented by numerical values . From the point of view of physics this numerical values describes "forces" and/or the "results" of the forces exercised. The interactions between these numerical values within specific equations can allow the reconstructive proof of the measure with which the eta-model (i.e. the equations) generate a model which is similar to the phenomenon which has to be understood. Thus the aim of the scientific knowledge is to define the mathematical functions being implicit in a phenomenon. When the said mathematical function is relatively complex, i.e. highly non linear, it is probable that our perception of the phenomenon oblige us to define the said phenomenon as being provided with qualitative features which cannot be reduce only to quantitative effects . This anyway only a perceptive effect due to the complexity which characterizes the function being implict in the phenomenon. It is thus possible to state that the quantitative component of every phenomenon is the perceptive effect of its highly quantitative non linearity. Considering that all the phenomenon which are present in nture evolve with highly non linear dynamics it is obvious that the perception of qualitative effects is so common to let one think that the quality is inherent in the phenomenon itself. In reality the difference between quantity and quality does not exist. Quality is the way with which the quantity expresses its virtuosity. In any phenomenon having pertinent topology, the identity and the unity of the phenomenon is guarated by its space-time cohesion. This means that every minimum elemnt of the phenomenon is contiguous and connected directly or indirectly through specific forces to the other minimum elements . The quantitative value of each minimum element of the phenomenon which is analysed is thus the result of the action of the said forces. It can be demonstrated that in a phenomenon with pertinent topology the forces connecting each minimum elemnt with any other minimum element of the local neighbourhood are sufficient to explain the space- ime cohesion of the entire phenomenon. C onsiderin a phenomenon P comprising minimum elements pi, p2, ... PM in a space having D spatial dimensions, where D=2 for simplicity. Assuming now that the local neighbourhood , N, of each minimum element is formed by a first layer of directly surrounding minimum elements a so called a gradient 1 surrounding in which 1 is referred to the a step. In a boidimensional space the local neighbourhood is then formed by N=8 minimum elements . The phenomenon P can thus be represented by a matrix having M elements with M= R(rows)X C (columns) .
Figure imgf000016_0001
Figure imgf000016_0002
In this matrix each minimum element Px,y will, have a quantitative value for example a quantity of light which a reflector is able to reflect and which we can express as u^ where we{θ,l}. The element p will
have also connections with its local neighbourhood N. Defining now fw+ +z (with k and z integer and
such that Are {-l,l} and z<={-l,l}) as the forces of its connections with the minimum elements of its local neighbourhood N. It this possible to write now
(3.1) uχJn+ = f[il « x+ k,y+Σ W^x.y.x+k.y+zln] l were |*| + |z|>0; from which the following equation can be derived:
Figure imgf000017_0001
In other words : the quantitative value of each minimum element n is a function of the quantitative
value of the 8 minimum elements surrounding the said element and of the forces which connects the said minimu element to the surrounding elements . In a more analytical language :
g3 u[n] x+ y+] n ,y,x+ϊ,y+l),g4 u[n -l,y n ,X->,y ), g5
Figure imgf000017_0002
),
If we begin with x=l and y=l recursively substituting at each n-th step ιrn*x+k,y+z the same function /() which we have calculated above for u^ than one can see that one easily reaches an index xc and y^ , which demonstrates that the quantitative value of each element p depends from all the other elements.
From the above it results that by defining the strength of the connections between each minimum element of a phenomenon P and the minimum elements of its local neighbourhood, one defines also the global connections strength of the etire phenomenon P. Obviuosly this statement is valid only if assuming that the phenomeno P has a dynamic behaviour . From this it must be argued that each phenomenon having a pertinent topology is always a temporal process or the partial result of the temporal process .
By combining the ideas of local connectivity and time dynamics which characterize the phenomemnon having pertinent topology on has to argue that bthese phenomenon has at the same time a paralle and a sequential behaviour. This kind of asynchronous parallelism can be formalized. Considering a phenomenon P with pertinent topology ina apportion of a plane having two dimensions (D=2) and which elements are squares of minimu sides L and which elements divide perfectly the plane in single tiles . In this conditions one can evaluate the function by which the two minimum elements of P being at the higher diatnce one from the other exercise a repricola influence one on the other. Assuming that each minimu elemnt of P is directly connected to ist local neighbourhood of gradient 1 (g=l) , the signal diffusion of each minimum elemnt will take place t each time istant t according to the following equations : f I(t) = (2g{l) +lY -l (3.3) S(H-1)=' + 1 where I (t) is the number of the minimum elements pf P which are influenced at the T-th temporal cycle by the signal of the refrence minimum element. These eauations represent the diffusion functions of the signal of a minimum element of a phenomenon P of any dimension D whatever . It is further possible to define the time delay Δt with which the two most distant minimum elements of P influence each other. For a bidimensional situation with D=2 and knowing
P χ γ >v e P χ γ2 > v2 the two minimum elements defined above, ' the time delay is expressed by
Δt = -J{χι -χ 2 )2 +{yι - 2f +! (3-4) Furthermore a phenomenon having a pertinent topology is formed by minimum elements and by local connections among the said minimum elements. Each minimum elemnt is deifned by the position which it occuopies in the phenomenon and by a quantitative value which indicates at each instant for example a charatersitic effect of the phenomeno such as for example the quantity of light being emitted or reflected by a reflector. The above can be expressed as Minimum element= ir x,*y,z,...D^
Where u is the quantitative value and x, y, z, ....D is the position and t is the time instant. Each local connection defines the strength of the mutual influence of each minimum elemnt on each other minimum element which is directly adjacent to the said minimum element at each time instant. Each local connection is thus defined by the position of the minimum elements connecte by it and by a quantitative value at each time instant and this can be expressed as: Local connection =
Figure imgf000020_0001
Where W is the quantitative value
X/ y« °s is the position of the source minimum element
T is the time instant and k,z Dd is the target minimum element. Considering the above an active matrix of the connections can be defined by the system of the following equations :
„P+1] y W J/] ) (4a) x,y,...,Dt J \ x,y,...,Ds k,z,...,Dd ' (x,y,...,Ds ),(k,z,...,Dd )>
W(x'ly D,
Figure imgf000020_0002
Equation ($a) and (4b) describes the evolution of the minimum elements and of the connections of a phenomenon with a pertinent topology towards an attracting element which represents the group of its solutions . Both equations can be presnt in two degenerated ways which request a particular attention: As a first way LA) - ». • '. , •"H DΛ^, 0,>>) (4t> In this case the minimum elements of the phenomenon evolve only as a function of the connections, by starting from their original or starting quantitative value which has the only effect of an impulse of the process . One particular feature of this evolution consit in the fact that the process regulating the evolution of the connections takes place in the space of the connections (equation (4b) . It is in this space that the solution are found for the evolution of the minimum elements . So in this case the starting values of the minimum elements act as simple constraint to the evolution of the connections. These last ones dynamically furnishes the values of the elements and thus the actracting element of such a process will be the new definition of any minimum elemnt as an exclusively relational element which is generated by the dynamic negotiation between its starting value and the starting value of the adjacent minimum elements. The second degenerated case is described by the following equation:
U(x,y, ,D,),(k.z, Dd )
Figure imgf000021_0001
D, ),(k,z, Dd )>U(x' ,y, ,D,)>U(k,z, ,Dd)) (4d) In this case only the values of the minimum elements will evolve while their connections remain unchanged and the said connections will act as constraints of the evolutive process . The above analysis of phenomena having pertinent topology basing on the example of electromagnetic or visual phenomenon demonstrates clealrly once again thye limits of the known neural networks and particularly fo the cellular neural networks . This network s deisclosed in the state of the art does not completley fit with the structure of phenomena having pertinent topology since the relation between interactions and quantitative values of the minimum elements is not considered, but the interactions are described by predefined models, i.e. the operator templates, which structure is not influenced by the values of the minimum elements of the phenomenon. On the contrary the artificial neural network according to the present invention is stepping closer to the structure of phenomena by considering the mutual relation of the quantitative value of the minimum elements formed by the data records of the database and the interactions between this minimum elements. In the following description there are shown and described in more details different sets of rules for describing the way with which the interactions between cells or knots of the present neural network and the quantitative value of the single cells or knots, i.e. their internal value or activation value or output value are related onte to the other in order to guide the computational process of the network. The invention realtes also to an algorithm for recognising relationships between data of a database which algorithm can be more effective, rapid and more precise in highlighting the relationships between the data of a database. The algorithm according to the invention is an algorithm for recognizing relationships between data of a database having pertinent topology, this meaning that the data are of the kind where the relative position of the data records in an array of data records or in a distribution of data records in a N-dimensional space, particularly a two or three dimensional space is a relevant feature of the data record and where the data records can be represented as an array of cells or points, each point being univocally related to a data record of the database and having a univocally defined position in the array relatively to the cells or points of the other data records , to each data record being further associated at least one variable or more variables each one having certain value, the algorithm being characterised by the fact that - each cell or point in the array of cells of points representing a data records of a database is considered to be a unit or a knot of an Artificial Neural Network; - each unit or knot formed by a cell or point of the database being successively defined as a target unit or knot and connections being defined between each target unit or knot at least to each one of the rest of the units or knots formed by the rest of cells or points of the database which are at least of gradient 1 relatively to the corresponding target unit or knot;
- a new output value of each unit or knot of the database successively defined as target unit or knot being calculated by means of the set of learning rules or functions defining the strength of the connection between the units or knots or the set of activation rules or functions of the artificial neural network defining the quantitative value of each unit or knot i.e. its activation value or its output value, or by means of the combination of both the set of learning rules or functions and the set of activation rules or functions of the artificial neural network as a function of the actual output of the units or knots connected to the target unit or knot and of the actual output of the said target unit or knot;
- The actual outputs of each unit or knot being defined as the value of the variable or as the values of the variables associated to each data record represented by a cell or a point considered as a unit or knot of the artificial neural network; - And the new output of the target unit or knot is considered as the new value of the variable or of the variables of the data record associated to the cell or point of the array of data records corresponding to the target unit or knot; - By carrying out the said steps for computing a new output of a target unit or knot for at least part or for each cell or point of the array of data records a new array of data records is computed where the data record of each cell or point has a new value of the at least one variable or new values for the variables as a result of a first computation cycle of the artificial neural network according to the above steps; - The said computation cycle being repeated for each successive new array of data records until a certain prefixed number of repetitions of the computation cycle has been carried out and/or unless a certain maximum allowable error or discrepancy has been reached between the original values of the variable or of the variables of the original array of data records with respect to the values of the variable or variables of the array of data records according to the one computed in the last cycle and/or unless the difference between the value of the variable or the values of the variables of data records in the sequence of array of data records computed in the sequence of cycles is lower than a predetermined maximum rate . Many different sets of known learning functions or rules may be used or many different sets of known activation functions or rules may be used which may also be combined each one of the said sets or combination thereof can enhance or bring to evidence certain kind of relations in a better way with respect to other kinds of relations between the data records. According to a further improvement, the array of data records may be submitted at least twice or more times to elaboration according to the algorithm of the present invention, in a first phase being provided a first set of learning functions or rules or a first set of activation functions or rules or a combination thereof and in a second phase being provided a second set of learning functions or rules or a second set of activation functions or rules or a combination thereof and so on when more then two phases are provided while the array of data-records being used in the second or further phases of elaboration with the algorithm of the present invention with the second or further different sets of learning or activation rules or functions or with a combination thereof is the array of data records resulting respectively from the first or from the previous phase of elaboration of the array of data records . Although there is no need that the units or knots corresponding to the cells or points related to the data records at least of gradient 1 with respect of the unit or knot actually formed by the cell or point related to the target data record be the ones spatially directly surrounding the cell or point related to the said actual target data record in a special case particularly of a two or three dimensional array of data records the cells or points related to the data records of gradient 1 with respect to the cell or point of the target data record are formed by the data records associated to the cells or points of the data record array which directly surrounds the cell or point in the array of data record related to the said target data record. According to a further feature in the case the new array of data records computed by the algorithm is based only on a set of learning functions or rules, the new output of each target data record is defined as a function of new weights characterising the connection of each target unit or knot associated to the target data record with the units or knots represented by the cells or points of data records of gradient one relatively to the cell or point of the target data record, the set of learning rules or functions defining new weights of the connections as a function of the previous weights computed or defined in a previous computation cycle and as a function of the actual outputs of the unit or knot associated to the cell or point of target data record and of the unit or knot associated to the cells or points of the data records of gradient 1 or of the data records of the cells or points directly surrounding the cells or point of the actual target data record. A similar result may be reached by using sets of activation functions where these functions defines the new output of the target unit or knot corresponding to the cell or point related to the target data record basing on the net input to the said target unit or knot which is a function of the outputs of the units or knots corresponding to the cells or points associated to the data records of gradient 1 with respect to the target data record, particularly to the units or knots corresponding to the cells or points of the array of data records directly surrounding the cell or point of the target data record. The learning phase starts with a fixed predetermined value of the weights for each connection while the starting value of the unit or knot is modified according to a predetermined function which is also function of the weights and of the value of the surrounding knots or units and therefore to the data records which corresponds to cells or points in the array directly surrounding the cell or point representing a certain unit or knot of the artificial neural network . A new data array is thus constructed in which each data record has maintained the position with respect to the other data records in the representation of the data records as cells or points in an array while each data record has changed its value as a function of its original value and of the original values of the data records relative to the surrounding points in the array . A new cycle can thus be executed in which again each point representing a data record is set as a unit or knot of an artificial neural network, a weight is defined for each connection between units and a new value in the form of an output of each unit or knot is calculated according to the certain function. Several different kinds and structures of known artificial neural network using different known and/or new learning functions for the definition of the weights and different functions for the calculation of the new values of each data record associated to each unit or knot may be used. It is important to notice that since the weights of the connections between the knots of the artificial neural network are fixed at least at each cycle and eventually equal for each connection, the algorithm according to the present invention has practically no internal noise and the noise is given only by the noise of the data of the database. Different examples of sets of learning rules or functions and of known sets of activation rules or functions and of combinations thereof are disclosed in the following detailed description of some examples . As it will appear more clearly in the following detailed examples different learning rules for determining the weights of the connections of the artificial neural network or different functions for calculating the outputs of the units or knots of the artificial neural network which corresponds to the new values of the data records can be useful for enhancing certain kind of features of the data records. The algorithm according to the present invention can find a simple and effective use in evaluating digital o digitalized images in order to recognize or put in evidence certain eatures of the said images , as for example tissue differentiation, image pattern recognition and contrast enhancement. Particularly the algorithm according to the present invention can be used for image pattern recognition and for evidencing the different kinds of structures of the material forming a body, particularly of the tissues forming a biologic body in combination with radiological, echographic or magnetic resonance imaging or the like. As it will appear more clearly from the detailed examples the algorithm according to the present invention can be useful evidencing different kinds of tissues and/or for substituting contrast medium in the diagnostic imaging. In particular the different kind of revealed or enhanced information depends by the different set of learning and or activation rules applied to the neural network structure formed by the pixel array of the original image. Principally the effect obtained is similar to the one which can be obtained by using enhancing media such as contrast media in the echographic, radiographic, nuclear magnetic resonance imaging or in similar imaging methods . The kind of information enhanced can furthermore be differentiated by the choice of a certain set of learning and/o activation rules. In general it is possible to recognize two families of this sets of rules . The two families are summarized in the table of figure 3 in which the different sets of rules are given a name which will be explained later on in the detailed description. A first family of rules provides for weight evolution and unit activation and allow to carry out an extraction of features from a source digital or digitalized image such as for example kind of tissue in an image of biological bodies or structures of tissues or objects like for example recognizing the presence of stenosis in angiographic images . A second family comprises rules which principally carry out a weight optimisation. After weight optimisation a further step of unit activation can be carried out. This kind of rules helps in finding out edges between different zones of an image for recognizing boundaries, for example boundaries between different kind of tissues in images of biological bodies . The above example is not the only possible field of appliance of the algorithm according to the invention. Another example may consist in evaluating the relations and dependence between genes relatively to their common activation. The invention relates also to a method for image pattern recognition in digital or digitalized images . Particularly the invention relates to a method for the recognition and differentiation of different tissues in digital or digitalized images of biologic bodies, as for example in diagnostic images as for example digitalized radiographies or echographic images or magnetic resonance images . It is important to notice that although the present description of the examples is referred particularly to a two dimensional array of data records the algorithm and the method according to the invention are not limited to such a space but can be applied also on three and N dimensional arrays of data records . The algorithm according to the invention and the method using the said algorithm will appear more clearly form the following description of some examples by means of the annexed drawings in which: Fig. 1 is a schematic view of a two dimensional array of cells each one representing a data record in a two dimensional array of data, for example each cell representing schematically a point (a pixel) of a digital or digitalized image. Fig. 2 represents the step of transforming the nine central points or cells of figure 1 in units or knots of an artificial neural network. Fig. 3 illustrates a diagram of the different families of learning and activation rules for the neural network particularly referring to a neural network structure formed by the pixel of a digital image. Fig. 4A, 4B, 4C, and 4D illustrate respectively a mammography, the enlarged mammography, the digitalized mammography which has been subjected to the method for image pattern recognition according to the present invention and according to a particular learning law called CM and the digitalized mammography which has been subjected to the method for image pattern recognition according to the present invention and according to further particular learning law called IAC. Figures 5A, 5B, 5C, 5D are respectively the figure 4A and a further enlargement of the figures 4B, 4C, 4D. Figures 6A, 6B, 6C, 6D, 6E, 6F, 6G are respectively a photograph of bacteria and four digitalized images of fig- 6A subjected to known pattern recognition methods while fig. 6G illustrates the digitalized image of figure 6A subjected to the method according to the present invention using automata rules for the artificial neural network called IAC particularly enhancing contrast. Figures 7A, 7B, 7C are a digital photo of the ventricle, the said digitalized photo subjected to a known pattern recognition method and the digitalized photo subjected to a method according to the invention called CM. Figures 8A, 8B, 8C, 8D and 8E show a set of echographic images of the same body part where metastases are present and where figures 8A to 8D are respectively an image without contrast media, with contrast media in the arterials phase, an image with contrast media in balance, an image with contrast media with in a late phase, the image of the same body part, without contrast media and treated with the method according to the present invention. Figures 9A, 9B, 9C, 9D are respectively a radiographic image of the lung, the digitalized image of figure 8A subjected to a known algorithm called best filter, the image of fig. 8A subjected to a treatment according to the present invention and with an rule called CM, and the image of fig. 8A subjected to a treatment according to the present invention and with an automotive rule called IAC. Figures 10A, 10B, 10C illustrates respectively a mammography, the mammography of fig. 10A treated with a known pattern recognition method and the mammography of figure 10A treated according to the method of the present invention with an automotive rule called High CS . Figures 11A, 11B, 11C, 11D and HE illustrate respectively a normal photograph, the photographs according to fig. HA treated by two different known patter recognition methods and the photograph of fig HA treated with the method according to the present invention with two different automotive rules called IAC and IAC combined with CM. Figures 12A to 12J illustrates respectively a radiographic source image of the anatomical district of the femoral arteria and the image elaborated according to the present invention by using different kind of rules and different combination thereof. Figure 13 is a schematic block diagram of a typical generalised cell or knot of the neural network according to the present invention. Referring now to figure 1, a two dimensional array of data records is schematized by an array of square cells which are identified by Pi, where i are elements of Natural Numbers. The central nine square cells P7, Pβ, 9 P12, P13, Pi4, and Pχ7, Pi8, P19, are drawn with continuous lines while the surrounding further cells are drawn with discontinuous lines for representing the fact that the array can be larger then the one represented. Each cell of the array is a cell for a data record which has its univoquely related cell and so a univoquely related position in the array, the relative position of each cell of data record to the other cells of data record being a relevant feature of the corresponding data record. Each data record of each cell has a value Ui , where i is element of the natural numbers and where the said value can be a parametric value of only a variable of the data record or a vector comprising each data record a set of different variables . The above scheme is a representation of what is called in the specific technical language a database having a pertinent topology. This means that not only the values of the data records are relevant feature of each data record but also their relative position in a projection of the datarecord of the data base in a n- dimensional space or in an arrangement of the datarecords of the database in a n-dimensional matrix form which is just a different formalism for indicating such a projection. For example a digital or digitalized image is a two dimensional array of pixel. Each pixel represents a data record in an array of data record and the relative position of the pixel is a relevant feature of each pixel belonging to the relevant information of the image data. The value of the pixel is a simple intensity value in a grey scale image, while in a coloured image the value is a vector comprising at least two parameters or variables namely the colour and the intensity, normally three parameter or variables in the actually used techniques of visualizing coloured images as the so called HVS or the so called RGB coding of coloured images. Figure 2 try to schematically picture the step used by the algorithm according to the present invention which provides for treating the data records of the two dimensional array by transforming the two dimensional array of cells in an array of cells or knots of an artificial neural network. The squares represents the relation to the original cells P of figure 1 , while the enclosed circle Ki represents the transformation or association of each cell i to a unit or knot K± of an array of cells or knots of an artificial neural network. The connection arrows w±j represents the connections between the units or knots K± and Kj of the artificial neural network. Figure 2 is related only to the nine central cells of figure 1 and the connection arrows are continuous for the central cell P13, which is the cell corresponding to the actual target data record, while the discontinuous arrows are related to connection to peripheral cells which are obviously used by the present algorithm and the mechanism of which derives identically from the description of the way of working of the present algorithm as described with reference to the sole central cell Pi3. Also the fact that the arrows are provided at both the ends of the connections represents the fact that by describing the way of working with reference of a peripheral cell to the centre cell P13, this actual centre cell P13 will become a peripheral cell of another cell which will be the centre cell . So by applying the algorithm according to the present invention to the array of cells the said algorithm sweeps the array cell by cell each time identifying a cell as the momentary central cell and the corresponding peripheral cells and the associated connections between the central cell and the peripheral cells and when the computation has been carried out for a certain cell momentary defined as a central cell , an adiacent cell is defined as the new central cell and new peripheral cells of this new central cells are identified with the corresponding connections, the computation being carried out this time for this new central cell and the new peripheral cells with the corresponding connections and so on. This mechanism is repeated until each cell of the array has been subjected to computation by the algorithm, thus ending a first computation cycle of the artificial neural network. The following cycles are identical with the difference that they are applied to the array of cells which has been previously treated by the algorithm. Figure 13 illsutrates a schematic diagram of a cell or knot of the artificial neural network according to the present invention. Each knot or cell Ki comprises inputs at which the output signals Uj from other knots Kj are received by the knot or cell Ki. Furthermore each knot or cell K± has an output Ui which is connected to the input of one or more other knots or cells Kj . The output Ui of each cell or knot is a function of an internal value i or state of the knot KL also called activation value which can be also an identity function namely Ai=Ui. In the present artificial neural network each knot or cell K has also an input for an initialisation value defining the initialisation internal value or actication value A± of the knot or cell or the initialisation output value Ui. In this case, referring to the example of figures 1, 2 and 13, the initialisation value for the internal value or activation value A± or the first output value Ui of the knot K is set equal to the value Ui of the data record Pi univocally associated as explained above to the knot Ki. Each knot or cell Kj. has further an output for the computed value U" for the datarecord P after the n-th processing cycle of a sequence of a certain number m of processing cycles. The said computed new value U" corresponds either to the internal value or activation value A" of the knot or cell K± after the n-th processing cycle has been carried out or to the output value u" of the knot or cell Ki after the said n-th processing cycle has been carried out. As it appears from figure 2, the present artificial neural network comprises direct connections of a knot or cell K only with the knots or cells Kj directly adjacent to the said knot or cell Ki. The set of this knots or cells Kj are technically defined as the local neighbourhood of the knot or cell K± or the knots or cells of gradient 1. The term gradient 1 refers to the stepwise difference between the position of the reference knot or cell K± and the position of the set of knots or cells Kj forming the local neighbourhood of the said knot or cell Ki. This definition is intuitively understandable and is a widely spread technical definition. So it is clear that in a discretized array where only discrete positions are possible for the elements of the array, defining a reference element the gradient 1 neughbourhood is formed by the set of elements which position differs from the position of the reference element by a step 1 in any direction of the array. This is the simplest case. The term gradient 1 is more generic since it is not limited to the spatial coordinates, but can be referred to any kind of space in which the an n-dimensional database is projected. The connections between the knots Ki and the knots of the local neighbourhood Kj are defined by their strength. The connection strength is quantitatively given by weights ji or Wij which are multiplied with the output Uj of the knots or cells Kj of the local neighbourhood of the knot K±, when the signals of these knots of the local neighbourhood are send to the reference knot K± connected to them. As explained in the introduction of the description these weighted signals inputted to the reference knot Ki will have an effect on its internal state Ai in order to cause or inhibite the emission of an output signal n or to change the value of the said output signal Ui . As already explained in the introduction such a neural network does not only consider as a basis for the processing the values of the knots or cells but also their relations and furthermore the interactions between the said values of the knots or cells and their relations . Two extremes has been generally discussed in the introduction one of which considers the evolution of the internal value Ai or of the output value Ui of the knots or cells Ki only as a function of the connections wji' by starting from their original or starting quantitative value Ui which are the values of the datarecords Pi univocally associated to a corresponding knot Ki . This initial or starting value Ui has the only effect of an impulse of the process. One particular feature of this evolution consit in the fact that the process regulating the evolution of the connections w takes place in the space of the connections (equation
(4b) . It is in this space that the solution are found for the evolution of the knots or cells . T he starting values of the knots or cells KL act as simple constraint to the evolution of the connections w±j . These last ones dynamically furnishes the values of the knots or cells and thus the actracting element of such a process will be the new definition of any knot or cells K i as an exclusively relational element which is generated by the dynamic negotiation between its starting value U i and the starting value Uj of the adjacent knots or cells Kj of the local neighbourhood. The weights W which quantitatively defines the conncetions are the function of the internal values of the knots Kj of the local neighbourhood and/or of the reference knot Ki. In a second extreme case as discussed form a general point of view in the introduction only the values of the knots or cells either as activation values Ai and j or in the form of output values Ui and Uj will evolve while their connections as represented by the weights Wji remain unchanged and the said connections will act as constraints of the evolutive process . Thus the way for determining the weights w and the values Ui of the knots are given by set of equations called also rules. Generally speaking, equations which defines the quantitative values of the weights are called learning rules , while equations which changes the internal state or the activation value or the output value of the knots as a function of the output values of the reference knot and of the adjacent knots of the knots of the locqal neighbourhood are called activation rules . So as already explained computation may be carried out by only considering a set of learning rules or functions which computes new outputs of the actual centre cell Pi, by means of defining new weights Wij of the connections as a function of the outputs of the centre cell and of the surrounding cells, this means of the values of the variable or of the variables of the data record corresponding to this cells. The new output of the centre cell will be then set as the new value or the new values of the variable or of the variables of the centre cell and will be a function of the outputs of the surrounding cells, this means of the values of the variable or of the variables of the data records corresponding to the said surrounding cells and of the new weights computed by the algorithm. During computation in a first cycle the weights ± of the connections between the cells P± and Pj used as knots Ki and K of the artificial neural network are set all equal to a certain value which can be randomly defined by a value for example 1 or 0.0001. Basing on this weights the value of the actual central cell Pi, in the example of figure 2 the cell P13 working as a unit or knot Ki3 of an artificial neural network is changed as a function of the values of the surrounding cells j which are also set equal to knots Kj of the neural network (in figure 2 this are the cells and knots P7, P8, P9, P12, Pm, and Pi7, P18, P19 and K7, K8, K9, K12, K14, and Ki7, Kiβ, K19) and of the weights of the connections Wij. Another way of computing the new outputs of the units or knot Ki is to use a set of activation functions or rules , which defines the output of the central knot K13 as a function of the values Uj (in fig. 2 U7, U8, Ug, U12, Ui4, and U17, Uι8, Ui9>of the data records of the cells Pj (in fig. 2 P7, P8, P9, Pα2, Pi4, and Pι7, Pι8, P19) surrounding the central cell Pi (in fig. 2 Pχ3) corresponding to the a unit or knot Ki (in fig. 2 K13) to be activated and as a function of the weights w±j . In this case new weights w^j are not computed. There is also the possibility to apply a combined computation using a certain set of learning rules computing new weights w±j for the connection at each cycle and a certain set of activation rules computing new output values at each cycle . Thus at the end of a cycle new outputs , this means new values of the variables for each data record of the array of data record has been computed and also new weight of the connections has be computed, both which new values of the variable or of the variables of the data records and of the weights of the connections are used in the following computation cycle. Relating to the example of figure 2, the function defines the net input of the knot Ki3 by means of the outputs U7, U8, Ug, Ui2, Ui4, and Uι7, Uiβ, U19 of the surrounding knots K7, K8, K9, Ki2, K14, and K17, Kχ8, K19 corresponding to the values of the associated cells P7, Pβ, P9, Pi2f Pi4f and Pι , Pι8, P19 surrounding the central cell P13 and the associated knot Ki3. A further function is provided which basing on the differences between the previous values of the outputs and the new values calculated by the algorithm, defines new weights Wij for the connections between the central knot Ki=Kι3 in fig. 2 and each one of the peripheral knots Kj equal to respectively K7, K8, Kg, Ki2, K14, and K , K , K19 in figure 2. At the end of the first cycle when all the cells has been treated according to the above mentioned steps a new array of cells has been constructed in which each data record has maintained its cell and where each data record has a new value for the variable or for the variables computed as described before in the first cycle while also new weights for the connections has been computed. Thus a second cycle can be carried out by using this new array of data records this time using the values of the data records computed in the first or previous cycle as outputs of the knots of the artificial neural network corresponding to the cells and where the new weights for the connection between the knots corresponding to the cells are used which has been computed in the first or previous cycle as described before. In the second or following cycle again new outputs of the knots and thus new values of the data records are computed thus forming again a new array of data records in which the data records has updated values . Similarly corresponding new weights for the connections between knots are computed to be used in a following cycle of the algorithm if this is provided or necessary. The number (n) of following cycles can be determined as a fixed parameter. Alternatively the number of repetitions can be determined by an error or deviance calculation of the values of the new array of data records related to the original one by setting a maximum error or deviance. This calculation can be carried out by means of usual error or deviance calculation in the field of the artificial neural networks . Another criterion might consist in the fact that the difference of the sequence of new values for the variables of the data records is monitored and the repetition of the algorithm is stopped when the differences of the said values in the sequence of cycles is not anymore significant, this mean that it is less than a predetermined value. This means also that the new values does not differ substantially from the previous one. The meaning of substantial difference has to be related to the value of the variables of the data records and to the kind of data records since in some cases very small differences might have great influence on the information of the data record while in other cases not. In the following some examples of rules for updating the output of the knots of the artificial neural network and thus the values of the corresponding data records and for updating the weights of the connections are given. This rules which are functions are defined by a name which will help in simplifying the identification in the description of a particular rule. In particular the different kind of revealed or enhanced information depends by the different set of learning and or activation rules applied to the neural network structure formed by the pixel array of the original image . Principally the effect obtained is similar to the one which can be obtained by using enhancing media such as contrast media in the echographic, radiographic, nuclear magnetic resonance imaging or in similar imaging methods . The kind of information enhanced can furthermore be differentiated by the choice of a certain set of learning and/o activation rules. In general it is possible to recognize two families of this sets of rules . The two families are summarized in the table of figure 3 in which the different sets of rules are identified by a name explained later on . A first family comprises rules which principally carry out a weight optimisation which can further be combined with a unit activation. This kind of rules helps in finding out edges between different zones of an image for recognizing boundaries, for example boundaries between different kind of tissues in images of biological bodies . A general kind of set of rules is defined as Automata Rule (A or AR) and comprises functions for determining optimized weights, so called learning rules the following two set of rules can be used: σ k = >N k uj -k u , y
Rti = e U E y [0,2*]
Figure imgf000044_0001
e J - e
Wit = lJ e Rti J + e -*i Ji
where is a parameter which can freely defined by the user and which renders the algorithm more or less sensible to differences in the image.
Rij is some sort of measure for the distance of the i-th unit from the j-th unit.
Ui are the values of the single cells Pi transformed in knots Ki of the artificial neural network and where
The suffix i defines the central cell or knot and the suffix j defines the cells or knots directly surrounding the said central cell or knot. According to the diagram of figure 3 this general rule may be further specialized or enhanced obtaining different set of rules . The above disclosed automata rule defines the connection strength between each target or reference knot or cell K± with each one of the knots or cells Kj of the local neighbourhood as the non linear projection of their different value Ui and Uj.Each reference knot or cell Ki is thus connected with eight symmetric weights Wj to the knots or cells Kj of the local neighbourhood of gradient 1. Every bidimensional database such as a bidimensional image will so be formed by a matrix in which each element of the matrix is connected as an autonomous unit to the eight elements of its local neighbourhood, i.e. to the matrix elements directly adjacent to the reference one. Each bidimensional matrix array of data such as an image will be further formed by a specific number of symmetric connections between each element of the matrix and the elements of the local neighbourhood. The quantitative strength value of the connections WjiWhich are defined by the above mentioned equations forms the constraints by means of which the data arry such as an image, being in the form of an active matrix of the connection strength will cause the evolution in time of the initialisation value of each cell or knot till to the predetermined n-th processing cycle and towards the natural actractor. A set of rule II is generally indicated as CS, i.e. Constraint Satisfaction which can be developed in different variants disclosed hereinafter: 11.1. so called CS XOR
Center Neighbors State 1 1 1 1 0 0 0 1 0 0 0 1
Initialization : w, = input
Figure imgf000045_0001
Δ . = Net, (1 - u, ) α Net, > 0 Δ,. = Net, u, a Net, < 0
Figure imgf000046_0001
where Ui is the output of the central knot Ki
Ui is also defined as input of the i-th knot or cell in combination with the initialization step. This definition wants to stress out that at the first step of carrying out the neural network the i-th unit has as an output value the value of the corresponding pixel in the digitalized source image. Obviously in the starting step of the neural network this value has to be attributed to the i-th knot or cell ad is considered as an input. In the following computational cycles the value of each knot or cell is changed by the algorithm and when a certain number of repetition has been carried out the value of the i-th knot or cell corresponds or is correlated to the value of the i-th pixel in the output image. Uj are the outputs of the surrounding knots K
Neti is the net input of the central Knot Ki computed as a function of the outputs Uj of the surrounding knots Kj and the weights w±j of the connections of the the surrounding knots Kj to the central Knot K± . n is the number of cycle
<x± is the update value of the output U± of the central knot Ki for computing the new output value for the next cycle as a function of Net input Neti and the actual output Ui of the central knot K± α is a constant. II.2 so called CS AND Center Neighbors State 1 1 1 1 0 0 0 1 0 0 0 0
Initialization : u, - input, Net = ( \ Uj Wy ) •
9 1 At = Neti (\ - ui) a - (5 Δ --) Neti >0 1 At = Netj «,• •α (σ - -) Nett < 0
Figure imgf000047_0001
in this case same definitions apply as in the previous case. The difference between this activation function and the previous lie in the fact that the function for computing <x± comprises a further term defined as 2 and this term is defined above as a function of the mean of the outputs of all knots except the central knot K± activated. II.3 so called CS CONTEST
Center Neighbors State 1 1 1 1 0 0 0 1 1 0 0 0 Initialization : u, = input, Nett =(∑ujwij) a
9 1 Δj- =Neti l-Ui) a σ --) Nett >0 1 9 At = Net uf α ' (~-σ ) Afef/ < 0 [n+1] tt = Mπ];
u =
Figure imgf000048_0001
Also here apply the same definitions as in the previous examples of activation functions (1) and (2) . The difference lie in the fact that the term a± is computed in a different way for the case Net±<0. II.4 so called HIGH CS XOR Initialization : u, = input, e^ = (J UJ wϋ σ2(wjk)) a k≡[l,N];k≠j j A, = Net, • (1 - u,) a Net, > 0 Δ;. = Net, -u, Net, < 0
Figure imgf000048_0002
again the symbols are defined as in the previous examples of activation functions In this case the function for computing the net input Neti to the knot Ki is different as in the previous cases . The function α2 is used in the computation of the net input. In this case the function 2 is applied to the weigths w±j of the connections as it appear clearly from the above equations. The functions for computing the update a± value of the output Ui of the central knot Ki are identical to the case of the first example of activation function' II.5 so called HIGH CS AND Initialization : u, = input,
Nett = ( Y u j w;j σ 2 ( Wjk )) k ε [1, N]; k ≠ j j ? 1 Δ,- = Nett (1 - «,-) « - (σ - -) Nett > 0 7 1 Δz- = Net " ui ' α ' (σ — ) Net( < 0
Figure imgf000049_0001
N N uic (u - Uk ? - = A≠i_ _ Q 2 = k≠i N ' N As it will appear clearly from the comparison of this set of equations with the example of activation function (4) called High CS XOR and activation functions (2) called CS AND, this is a combination of the two set of activation functions. II.6 so called HIGH CS CONTEST
Initialization : u, = input,
Neti = (∑ uj wtj σ 2 (wjk))• k e [1, N] ; k ≠ j j 9 1 Δ1-=M?fr(l-κI-)-α-(σ --) Net t > 0 1 9 Δ,- = Nett «,■ ■oc (--σ ) Nett < 0
Figure imgf000050_0001
Also this case represents a combination of two of the previous examples of activation functions, namely the activation functions (3) called CS CONTEST and the activation functions (4) called High CS XOR.
The above disclosed different variation of set of rules follow a common general concept. The basic idea as explained by means of the example of a two dimensional image consist in considering each pixel of the matrix as an hypotesis which can be gradually true or false in relation to the luminosity of each pixel. Also this system inherits the local connections between each pixel and its local neighbourhood the strength of the said connection being expressed by means of the weights as computed by the above disclosed automata rule. This conncections work as constraints during the evolution of the active matrix of the connections. The function of cost of the above discloses system of equation consist in attempting to make true each hypotesis of the matrix and thus to bring luminosity of each pixel to a maximum. During this process the weights defining the connection which has been previously computed will act as constraints . As an example above there is shown the behaviour of the pixels in some border line cases :
Figure imgf000051_0001
A further special set of rules differing from the CS set of rules is the set of rules III identified by the name IAC, i.e. Interaction and Activation Competition IAC
M _ TΛ ιl Sigma
MaxPixelRa nge = 2 ; w,- e [θ,l] ; α = β = M ; N = Intorno
Max =\;Min =0;rest = 0.\;decay =0.1; N ecc i = UJ Wi w,j > 0
N ιnι i =∑UJ W; W,j < 0
Net j - (ecc ,• α ) + (ini , ■ β ) e Net l ; - e -Net ' ;
Act i - Neti + e-Netf
A , = (Max -Uf): Act z- - decay (u , - rest ); Act ι > 0
Δ , = (u, - Min ) Act t - decay u, - rest ); Act i < 0
[n + \] [n] , e l -e ' e ; l + , e -Δ l;
Here the net input Neti of the central knot K± is defined as a combination, more precisely as a sum of two different functions for computing the net input depending on the fact if the weights Aj of the connections are positive or negative. More precisely the functions ecc± and inii are identical only a selection is made between inputs associated to positive and to negative weight connections. This is expressed by the two functions ecci and inii. This two components of Neti are related to an excitation or enhancement in the case of the component ecci and to an inhibition in the case of inii . The net input Neti is a combination of the ecci and inii in which ecci and ini± are multiplied respectively by a parameter α and by a parameter β. The two parameters determines the influence of the positive weigt connections and of the negative weight connections on the net input Neti • Furthermore an activation parameter Acti is provided as a function of the net input Neti to the knot Ki. This parameter is used for choosing one of two different function for computing the update value αi of the output Ui of the knot Ki depending on the fact whether Acti is negative or positive. The enhancing and inhibiting clearly appears also from the two functions for computing αi since Acti is present as a parameter in the said functions. The said functions further comprises the following expressions : Max and Min which are defined as the top and the bottom range of the activation value. Decay which is a function defined as the usual decay value along the time of each unit. Rest which is defined as the default value toward each unit tends for. The new output value of the knot K is defined by a particular exponential function which further enhances the excitation and inhibition mechanism of this example of activation functions . Generally speaking as referred to an example of an image formed by an arry of pixels as the database to be processed, the above equations defines an Active matrix of the connections in which every pixel is an agent which receives dynamically excitation and inhibition impulses from the other pixels and which modifies correspondingly its internal state. The above equation system can be defined as a collective auto-coordinating system between agent-pixels. The pixels having an high luminosity tend to support oneselves, while the other pixels are drawn towards low values of luminosity. The fixed connection weights between each agent-pixel and its local neighbourhood act as constraints which modulate the communication between each pixel . The said communication is formed by exciting and inhibiting messages. The evolution causes by discrete steps transformation of the original image that the pixel which are in a situation of a sudden change of luminosity will left isolated and are drawn to low values of luminosity. Thus, the entire system will highlight the edges which are present in an image without using function which is explicitly designed for this task. Where the pixel are in competition with each other some sort of walls are visualized which we often perceive as walls between different figures . The advantage of the above system of ruels lies in the fact that the edge detection effect takes place due to a local auto-coordination. Furthermore also a function as an adapve threshold filter occurs . For each edge a line is represented having a thickness which corresponds to the difference in local luminosity which the image shows in every area. A further alternative of the above mentioned sets of rules is the set IV so called Pixel Mexican Hat PmH MaxPixelRange = 2M ; u, e [θ,l] ; N = Surroundings', Max = 1; Min = 0; rest - 0.1; decay = 0.1;
Figure imgf000055_0001
j = ((l _ _ £ϊ_) . β 1+ώ»'* ) - ((l -y e l+ecc> 1 + ini i 1 + ecc,■
Figure imgf000055_0002
Δj- = (Max - ) - Acti ~ decay ut - rest); Act > υ Δ = i -Min) Acti ~ decay ui - rest); Acti < 0 u
Figure imgf000055_0003
Here the same definition apply as to the previous set III called IAC. It is important to notice that the more specialized sets II.1 to II.6, III, and IV differs from the general one called AR by the fact that not only a weight optimization is carried out but also a units activation. This system of rules act as a magnifying lens having a local and adaptive threshold. The effect referred to an image is that all the pixels are highlighted which are in a minimal conflict relating to luminosity but ehich have a regular and persiten luminosity so that the visual effect is to illuminate the edges which often are not visible with the eyes. A second family of rules provides for weight evolution and unit activation and allow to carry out an extraction of features from a source digital or digitalized image such as for example kind of tissue in an image of biological bodies or structures of tissues or objects like for example recognizing the presence of stenosis in angiographic images . Two sub groups of sets of learning and activation rules are provided. A first one indicated as the set V so called CM, Contractive map can be defined as a recirculation neural network and comprises a learning phase and a recall phase. (for deeper information see Massimo Buscema & Semeion Group "Reti Neurali Arificiali e Sistemi Sociali Complessi" Volume I, Edizione Franco Angeli Milano 1999) . The set of rules V so called CM comprises the following functions : a) Learning u e [0,1] ; C = neighbors ; wy = 0.0001 Initialization
Figure imgf000056_0001
"T 3 = ] + Δw,
Figure imgf000056_0002
Outt (rem, quot) = mod(-^ MaxPixel,MaxPixel) N b) Recall New Wjj = NewWy = w^ - w
where i is the value of the central cell or knot K± Uj are the values of the surrounding knots Kj Wij indicates the weights of the connection of the surrounding knots Kj and the central knot K±.
W,j defines the weights of the connection in the n-th cycle .
Wy defines the weight of the connections in the n+l-th cycle . α Wi is the value which has to be added to the weights
Wij in order to update then for the next cycle. Outi is the output value of the i-th unit corresponding to the i-th pixel which is the target unit or pixel
New Wji is the new weight of the connection between j-th and i-th unit
New Wij is the new weight of the connection between i-th and j -th unit w is the mean value of the weight. The arguments rem and quot relates to a particular function applied to the value of each pixel , this means to the output of each unit U . Considering only a grey scale image, the value of each pixel and thus the output of each corresponding unit can be divided by the number of grey scale levels . If one considers for example 256 grey scale levels than the value of each pixel can be divided by this number. By considering only solutions of the division belonging to the integer numbers this division gives rise to a certain reduced number of classes to which the pixels, i.e. the units belong and furthermore to a rest. The rest is indicated by variable rem. This kind of operation allow to classify the units an thus the pixels in a certain reduced number of classes and to visualize each class in the output image by giving to the class a certain uniform colour or grey level. The rest, this means the rem value is further used to differentiate the intensity of the colour or grey level. For example considering four classes for the pixel or units each class can be given a colour such as red, green, blue, yellow or four grey levels. The rest of each pixel can be used for varying each colour between dark red or the light or pale red and so on. These allow to further differentiate the pixels of the output image which belong to the same class . It is possible to compare the rem function as a function for modulating the values given to unit output and sou to the pixels of the output image within certain value range characterising each class defined by the quot function. Obviously the quot and rem function apply also to pixels values corresponding to coloured pixels, where normally the pixel value is a vector providing three parameters as for example RGB definition of colured pixel values or HSV, or the like. As it might appear evident from the above listed equations , the pixels value and correspondingly the output value of the units of the image transformed in a neural network are normalized at the beginning for maintaining the possible output values within an interval between 0 and 1. During definition of the classification by means of the quot function and the corresponding rem function the output values are again denormalized, i.e. rescaled from the interval between 0 lnd 1 to the original interval . Not using the classification of the functions quot and the modulation of the function rem and not denormalizing the output values of the units of the neural network, this means the pixel values, particularly the pixel values of the output image and by defining a scale of greys or of colours which scale is correlated to the values of the units of the neural network or the above mentioned pixel values of the output image it is possible to obtain a modified rule which is defined as "CM true color", while the other rules are defined as "CM quot" , when only the classification with the quot function is applied or "CM quot + rem" when also the rem function is applied in combination with the quot function. Referred to the example of the image made of an array of pixels, the basic idea of the above disclosed set of rule is that processing starts from the initialisation value of each pixel : the said value will be left invariant through the cycles . Indeed the evolution takes place on the weight defining the connections and at each processing cycle each pixel takes its new value of brightness as a function of the said weights which characterise the pixel at the current instant. Here connections between the reference pixel and the pixel of its local neighbourhood are bidirectional. Each pixel has a connection with which it communicates with another pixel and a second connection for receiving communications form this last pixel . The evolution law or rule of the weights defining the connections is in this set of rules of the deterministic type and the weights are intialized with values which are proximate to zero. The new values of brightness for each pixel in every process cycle are defined in the space of the weights . Thus each value of each pixel is defined by the weights defining the connection of this pixel with the pixels of the local neighbourhood. A further different set of rules which is the set VI and is defined as Local Sine LS belongs also to the above mentioned second family of set of rules . The local sine set can be provided according to two variants called Local Sine 1 LSI and Local Sine 2 LS2 These two sets differs only in one equation and the set of rules is described by the following equations :
,A Wjj = π ;
Figure imgf000060_0001
dij = {ui -uj - wij F ij = dϋ ' Neti ' C0S(WJ ' w )" uj ; (LSI) ij = dij ' Neti ' C0S(W7 ' wij )' -sin(" / • wij I (LS2)
W, [t+l] [t] , Λ y y ?/ '
Figure imgf000061_0001
wt = N wMax =
Figure imgf000061_0002
}, WMin = Min{wif },
, W, Scale = Offset = Min WMax ~ WMin WMax ~ WMin
Out = Scale • Wj + Offset;
Most of the variables or functions of the above equations have already been defined in the disclosure of the previous set of rules I to V. Dij represents some sort of distance of the i-th unit form the j-th unit. α is the standard variation. As already disclosed it is possible to use only a set of learning rules, only a set of activation rules or both a set of learning rules and of activation rules for applying the algorithm to the array of data records . As it will appear clearly form the following examples illustrated also in the figures , different kinds of sets of learning rules or of activation rules or of combinations thereof will lead to different outputs which will enhance particular relationships between the data records of the array with respect to other relation ships . It is important to notice that the present algorithm does not consist in a non supervised mapping algorithm but transforms an array of data records in an active artificial neural network capable of working in a supervised or non supervised manner depending on the rules or functions for learning and activating. In any case a learning phase is present which normally does not occur in non supervised mapping algorithm. Any kind of known or future artificial neural network may be applied for actuating the principle of the algorithm according to the present invention which is independent from the structure of the artificial neural network itself. Furthermore it is possible to subject an array of data records to a first learning phase according to a certain known set of learning rules and then a first time to a first set of activation functions . The result so obtained is obviously again an array of data records which may be subjected once again to computation with the algorithm according to the present invention this time choosing equal or different learning rules and a different set of activation functions. In this way an elaboration of the original array of data records is carried out by means of a sequence of two algorithm according to the present invention which are different relating to the learning rules and/or to the sets of activation functions . The different sets of rules applied in the first computation cycle and in a second computation cycle does not belong necessary to only one family of set of rules, but the set of learning and activation rules of both families may be combined in as two, three or more stage algorithm for elaborating digital or digitalized images . The effectiveness of the algorithm according to the invention will be shown by means of some practical examples. In this examples the array of data records is formed by image date of a digital image or a digitalize image . Each cell or point Pi of the schematic example of figures 1 and 2 is formed by a pixel of the image and the value of each cell of point Pi is formed by the intensity parameter, i.e. by the parameter influencing the level of grey of the pixel . It must be noticed that the examples could be also carried out for coloured image in this case the value of each pixel would have be defined by a set of parameter in the form of a vector defining the intensity and the colour. Figures 4A illustrate a mammography taken according to traditional techniques. Figure 4B is an enlargement of figure 4A. The traditional image of analogical type was digitalized with a scanner and treated with the algorithm according to the present invention. Figure 4C illustrates the image resulting from the treatment of the mammography of figure 4B where the algorithm was applied using only the set of learning rules and activation rules defined above as CM. As it appears clearly from figure 4C, the rule CM has the effect of enhancing contrast and putting into evidence image zones having equal intensity or level of grey scale by defining clear boredom lines . Equal intensities or grey scale level are defined within a certain range of intensities or grey scale level in a discretized scale of intensities or grey levels. As it appears clearly from the comparison of figures 4B and 4C, the nodulus or the zone of higher vascolarization in figure 4B is highlighted by the treatment with the algorithm according to the invention and the boredom of this zones are sharply depicted in figure 4C. Figures 5A to 5D are enlargement of this zone of the nodule to better evidence the clear boredom of the image zones furnished by the treatment of traditional mammography with the algorithm according to the present invention. It is also to notice that the image resulting from the treatment with the algorithm according to the present invention is able to extract also more information from the analogical mammography. Indeed the edges of the breast in the analogical image is not clearly differentiated from the background of the image. As it appears from figure 4C the algorithm according to the invention is able to set a more defined boredom clearly differentiating the background from the image of the breast. This is important in order to make measurement of the position of the nodule if an intervention has to be carried out, particularly a reduced invasive intervention by means of microsurgical techniques. The use only of the set of functions defined above as CM for applying the algorithm according to the present invention to the analogical image somehow inverts the black and white zones of the image for certain zones and being set for enhancing contrast gives some sort of very discretized image defining sharply differentiated image zones . Figure 4D and the corresponding enlargement of figure 5D illustrate the image obtained by a double treatment of the analogical image of figure 4C and 5C with the algorithm according to the present invention. In this case the first treatment is carried out by applying the algorithm using only the set of functions which has been defined above as CM. The image data, this means the array of image data obtained by this treatment is further subjected to a treatment with the algorithm according to the present invention where use is made of a combination of learning and activation functions and more precisely of the set defined above as Automata rules AR and of the set of functions defined above as IAC. At first it is to be noticed that an inversion of black and white zones has been obtained with respect to the image of figures 4C and 5C and partially with respect of the images of figures 4B and 5B. Particularly, the background is white instead of black. Furthermore the nodule is white as in the original analogical image of the mammography (figures 4B and 5B) . A more miniaturized discretization of the image is obtained and also the internal structure of the nodule, which in the image of figures 4C and 5C was not subjected to a differentiation in relation to the appearance of this structure in the original mammography except for the inversion of black and white, is subjected to a differentiation and the more important structures are highlighted. Figures 6A illustrate a analogical image of bacteria. The image has been taken with an electronic microscope. The figure 6B to 6F illustrates the images resulting from the treatment of the image of figure 6A with different known image pattern recognition algorithms . Figure 6G illustrates the image resulting from the treatment of image of figure 6A by means of the algorithm according to the present invention which is applied using only the set of rules named AR and the evolutionary rule named IAC as defined above. The difference and superiority of the image obtained by the algorithm according to the present invention appears clearly and without doubts form the comparison of the image according to figure 6G with the images according to figures 6B to 6F. again it is important to notice how the particular set of learning rules applied enhances the contrasts and sets clear boredom of the different image zones . The same effect can be observed in relation to figures 7A to 7C. Here figure 7A is a digital image of the ventricle of the heart. Figure 7B illustrates the image treated by a known image recognition algorithm called Snake . Figure 7C illustrates the image obtained by treatment of image of figure 7A with the algorithm according to the present invention using only the set of functions defined above as CM. Clear and well defined boredoms of the image zones are obtained and the different areas are well differentiated in the image. The about resulting capability of the algorithm according to the present invention of evidencing contrasts and setting clear boredom to different image zones and differentiating the image zones has a surprising effect. This effect can be appreciated by comparing the images of figures 8A to 8E. Figures 8A to 8D illustrates ecographic images of metastases taken at different instants in relation to the instant of injection of so called contrast media, i.e. of substances capable of enhancing the revelation of vascular activity by reflecting the ultrasound beam in a non linear way, namely at a frequency different from the fundamental frequency of the impinging beam, normally at a frequency of the order of the second harmonics of the fundamental frequency of the impinging beam. Figure 8A is an image taken at an instant before the contrast media are present in the imaged zone. Figure 8B is an image taken during the arterial phase where the contrast medium has reached the imaged zone. Figure 8C is an image of the same zone of figures 8A and 8B taken at an instant where the arterial and venal phase are at a balance. Figure 8D is an image taken at a late instant relating to the injection of the contrast medium. In figure 8D a white circle illustrates the metastases which appears as a darker spot relating to the surrounding image ones . Figure 8E is the image resulting form the treatment of image 8A, i.e. before the injection of the contrast medium in the zone to be imaged, with the algorithm according to the present invention using the set of functions defined as CM in the above description . It appears clearly how the metastases has been highlighted, particularly the ones enclosed by the white circle of figure 8D. Also some more little dots above the bigger spot in the white circle and at the right hand side of the said bigger spot are clearly highlighted. This spots can be seen with different evidence in the zone encircled by the white circle in the different images of figures 8B to 8D. In the original image according to figure 8A the bigger spots and dots can be seen very badly and the smaller ones are practically very difficult to be identified as alterations of the tissue. Thus the treatment of a digital or digitalized image with the algorithm according to the present invention can substitute with a high reliance and precision contrast media imaging. This is a very important advantage since on the first hand it is possible to avoid invasive interventions during imaging. Furthermore the use of contrast media during imaging requires the presence of a specialized medical assistant for injecting the contrast medium in a patient. Contrast media imaging, also known as harmonic imaging in the field of ultrasound imaging requires also longer times to be carried out since after injection of the contrast media some time is needed until the contrast media reaches the tissue to be imaged. Furthermore there is also difficult to predict when the contrast media will reach the tissue to be imaged and sometimes it happens that no image can be taken at the correct instant coinciding with the presence of the contrast media in the tissue to be imaged. Thus the algorithm according to the present invention allows to overcome all the difficulties and drawbacks related to imaging with contrast media which are more than the most important ones disclosed above. This can be appreciated by every expert of the field of diagnostic imaging, where a lot of arrangements must be provided in the imaging apparati for allowing to acquire images with contrast media, particularly real time images . Figures 9A to 9D illustrates respectively an analogic image, namely a radiography of the lung. The image obtained by treatment with the algorithm according to the present invention respectively using only a set of functions defined above as CM (figure 9C) and a treatment of the image of figure 9A by means of the algorithm according to the present invention using a combination of the above mentioned set of functions defined as CM and of the set of functions defined above as IAC (figure 9D) are compared with the analogical radiographic image of figure 9A and with an image obtained by treating the image of figure 9A with a known filtering algorithm called best filter. The already effects of the algorithm according to the present invention disclosed several times above can be appreciated. Particularly the image of figure 9D shows how the filament structure present in the lung is highlighted. In this case the to roughly discretized image of figure 9C does not allow to highlight this filament structure, which is depicted in figure 9D as black traces in the lower half of the lungs and at the sides of the two lungs facing each other. While the filament structure can be seen in the image according to figure 9A, the image according to figure 9B does not show this structure. In the image according to figure 9C the structure is present however the differentiation of the image zones is to sharp and rough. Figure 10 illustrates three images 10A, 10B, 10C put one beside the other and related to a mammography. The left hand figure 10A is the original mammography, the right hand image 10C is the original mammography, where some nodules are identified by highlighting them with white dots and encircling in white circles . The comparison of the left hand image 10A with the right hand image 10C allow to appreciate the very tiny level of differentiation of the image of the nodules from the surrounding image zone. The centre image of figure 10B is the result of the treatment of the digitalized figure 10A by means of the algorithm according to the present invention using only a set of activation rules defined in the above description as High CS. It appears evident that the algorithm has revealed and clearly highlighted in the obtained treated image the nodules which has been identified by the human eye and to this a further great number of nodules which where not apparent by human eye in the original image of figures 10A and IOC. The example of figures 11A to HE relates to an image of a landscape and not to a diagnostic image and is choosen in order to appreciate the accuracy of the algorithm according to the present invention differentiating image zones by maintaining the relevant structure of the original image. This means that the objects depicted in the image can still be recognized after treatment of the image by means of the algorithm according to the invention. Figure HA is a seascape image with a boat in the foreground. The image may be an image acquired by means of digital techniques such as digital cameras or the like or an image which has been acquired by means of analogical techniques and which has been digitalized by means of a scanner or the like . Figure 11B and 11C illustrates the image obtained by treating the image according to figure HA by means of two different known image elaboration algorithms . It appears evident that the subjects of the seascape can hardly be recognized in the images according to figures 11B and llC. Figure 11D illustrates the image according to figure HA after treatment with the algorithm according to the present invention, where only the set of functions has been used defined as IAC in the above description. In this image the subjects of the original image can be recognized much better than in the images of figures 11B and 11C and the different image zones are very good differentiated and recognized. Figure HE illustrates the image obtained by the treatment of figure HA with the algorithm according to the present invention where a set of functions defined as CM in the above description has been used in combination with a set of functions defined as IAC in the above description. The result of this treatment is that the image zones having a certain impact on the viewer are enhanced. So the land in the background and the boat in the foreground are enhanced with respect to the see. It is further to be noticed that the present algorithm can be applied in combination with other algorithm such as artificial neural networks or other prediction algorithm which are trained and tested to recognize kind of tissue or structure of the material composing and image . In this case instead of feeding to the image recognition algorithm using the image data array as passive information for training and testing the image data of the digital or digitalized original image, this data can be subjected previously to treatment by means of the present algorithm in one of the forms disclosed above. In order to ensure greater accuracy of the prediction the treatment of the image data of the original image by means of the algorithm according to the present invention can be carried out several times each time using a different set of learning functions or rules or a different set of activation functions or rules or a different combination thereof or using a sequence of treatment of the original image data in which each treatment phase uses different set of learning functions or rules or a different set of activation functions or rules or a different combination thereof. The prediction algorithm being applied to the image data obtained by each treatment and then the results of each prediction may be combined together or compared one with the other. Figures 12A to 12J illustrates further examples of elaboration of a same source image by means of the method according to the invention in which different set or rules or combination thereof are applied. Figure 12A is the source image which is a radiography of the femoral arteria. In the source image a circle and two ellipse indicated respectively with 1 , 2 and 3 encircles particulars of the vessels where a stenosis could be revealed. While the stenosis identified by the circle 1 appears in a sure way also in the source image, the zones encircled by the ellipses 2 and 3 does not give sure information . The first image 12B is obtained by the method according to the invention in which the set of rules V using the rem function. Here The constriction at 2 and 3 of the right hand branch of the arterial can be seen much better while also the central branch indicated by a rectangle 4 appear visible . The structure at the interior of the arterial branches does not appear very clearly. Figure 12C illustrate the image obtained by elaborating the source image with the set of rules V so called CM this time using only the option quot. Four classes of grey levels can be seen indicating different zones , namely three grey levels and white . The zones classified by quot does not show any further structure. In an y case the constrictions at 1, 2, 3, and 4 can be already identified in the output image. Figure 12D shows the result of applying rules V by taking into consideration both the quot and the rem function. In this case the four image zones having one of the four grey level of the image according to figure 12C are modulated by the rem function, namely the grey scale of image 12A. The four image zones having different grey levels are further modulated by the rem function and so a structure can be recognized. The constrictions at 1 , 2, 3 and 4 appear more clearly than in the previous images. Figure 12E illustrates the result of elaborating the source image 12A by means of the rules according to example V and by using the true color option, where the pixel values elaborated are maintained normalized within the interval between 0 and 1. also in this case it is possible to recognize the constrictions of the arterial at 1, 2 , 3 and 4. Figure 12F is the image obtained by elaborating the source image with the local sine algorithm according to the first option. Also in this case the presence of the constrictions appears evident at 1 , 2 , 3 and 4. The four figures 12G to 12J illustrates the result obtained by elaborating the source image 12A by a combination of two elaboration steps or stages each one carried out by means of the set of rules according to example V or VI . Image of figure 12G is obtained by first carrying out an elaboration with the set of rules according to example VI second option, namely the so called LS2 set of rules and by submitting the output image obtained by this first elaboration to a second elaboration this time using the set of rules V so called CM applying the Rem and Quot functions. Figure 12H illustrates a variant of the elaboration carried out for figure 12G in which the output image obtained by the elaboration with the set of rules according to example VI second option the so called LS2 set of rules is further elaborated with the set of rules according to example V, the so called CM, this time using only the rem function. Fig. 121 is a similar way of elaborating the source image 12A as in figure 12H, this time the second elaboration stage is carried out by using the set of rules of example V by applying only the quot function. Fig. 12J is still another variant of the two stage elaboration according to the previous examples of figure 12G to 121 , where the set of rules of example V is used for the second elaboration, this time by applying the True Colour variant. Although the images are all shown in a grey scale palette, it is possible to define colours and to correlate the different zones of the image to a specific colour, thus obtaining an artificially coloured image which better enhances the different zones of the image and the different objects identified by the elaboration according to the method of the present invention. Furthermore it has to be stressed out that although the example described is limited to applying the algorithm according to the present invention to an array of image data, this is not the only field where this algorithm can be applied, since treatment is permitted of every kind of database where the data are of the kind that can be represented as an array of points or cells, each point or cell being univocally related to a data record of the database and the position of each point or cell in the array relatively to the other points or cell being a relevant feature of the data record associated to the said point in order to reveal any sort of relation between the data records of the database .

Claims

1. An artificial neural network comprising a n- dimensional array of cells (Ki) corresponding to the knots of the neural network, each cell having connections to the directly adjacent cells (Kj) forming the neighbourhood of the a cell (K±) ; a) Each cell (Ki) having an input for each connection to a directly adjacent cell of the surrounding cells (Kj) ; b) each cell (Ki) having an output for the connection to one or more of the directly adjacent cells (Kj); c) the connection between each cell (K±) and the directly adjacent cells being determined by weights
( ^) ; d) each cell being characterised by an internal value defined as the activation value or function (Ai) of the cell (Ki) ; e) each cell (K±) being able to carry out signal processing according to a signal processing function so called transfer function for generating a cell output signal (u±) ; f) the transfer function determining the output signal (U ) of a cell (Ki) as a function of the activation value or function (A±) of the cell (Ki) , which transfer function comprising also the identity function which puts the activation value or function
(Ai) of the cell (K ) equal to the output signal (u±) of a cell (K±) ; g) a n-dimensional database of input data records (Pi) being provided which has to be submitted to computation by means of the neural network and in which n-dimensional database the relative position of the data records (Pi) when projected in a corresponding n- dimensional space is a relevant feature of the data records (Pi) , the data records (P±) of the database being able to be represented by an array of points in the said n-dimensional space, each point having an univocally defined position in the said array of points and being univocally related to a data record (Pi) of the said database, each data record (Pi) of the said database comprising further at least one variable or more variables each one having a certain value (Ui) ; h) each data record (Pi) being univocally associated to a cell (K) of the n-dimensional array of cells forming the neural network which cells (Ki) has the same position in the n-dimensional array of cells
(Ki) as the corresponding data record (P±) represented by a point in the said n-dimensional array of points; i) the value (Ui) of the variables of each data record
(Pi) being considered as the initialisation value of the network being taken as the initial activation value (Ai) or the initial output value (u±) of the univocally associated cell (Kj.) ; j) the activation value (Ai) or the output value (u±) of each cell (Ki) after a certain number of iterative processing steps of the neural network being considered as the new value (Ui) for the said univocally associated data records (Pi) .
characterised in that k) for each processing step of the said certain number of iterative processing steps, the weights (Wij) defining the connection between each cell (Ki) and the directly adjacent cells (Kj) are determined as the function of the current values (Uj) of the variables of each data record (Pj) univocally associated to the cell (Kj) directly adjacent to the said cell (Ki) , the said function being a so called learning function or rule; 1) the current activation value (A.±) or the output value (Ui) of each cell (Ki) after a processing steps of the neural network which is considered as the current new value (Ui) for the said univocally associated data records (Pi) being determined as a function of the current output values (Uj) of the directly adjacent cells (Kj) weighted by the corresponding weight (Wij) defining the connection of the directly adjacent cells (Kj) with the cell (K±) .
2. A neural network as claimed in claim 1, characterised in that it is modified by determining the current activation value (Ai) or the output value (Ui) of each cell (Ki) after a processing steps of the neural network which is considered as the current new value (Ui) for the said univocally associated data records (Pi) as a function of the of the weights (v±j) defining the connection of the directly adjacent cells (Kj) with the cell (Ki) , the said function being a so called activation function or rule.
3. A neural network as claimed in claim 1 , characterised in that the current activation value (A) or the output value (u) of each cell (Ki) after a processing steps of the neural network which is considered as the current new value (Ui) for the said univocally associated data records (P±) being determined as a function of the current output values (Uj) of the directly adjacent cells (Kj) and of the corresponding weight (Wj) defining the connection of the directly adjacent cells (Kj) with the cell (Ki) , the said function being a so called activation function or rule.
4. A neural network as claimed in claims 1 or 2 , characterised in that for each processing step of the said certain number of iterative processing steps, the weights (Wij) defining the connection between each cell (Ki) and the directly adjacent cells (Kj) are determined as the function of the current values (Uj) of the variables of each data record (Pj) univocally associated to the cell (Kj) directly adjacent to the said cell (K±) and of the current value (Ui) of the variables of the data record (Pi) univocally associated to the cell (Ki) ;
5. A neural network as lcaime in one or more fo the preceding claims characterised in that only the following learning rules are used:
k = ° N -(k-uj —k-u j)'
R L*a/ == e uu eE [ r0o,,22 NJV - ]
s eRV - e~^
Wii = e J + e J
where α is a parameter which can freely defined by the user and which renders the algorithm more or less sensible to differences in the image. Rij is some sort of measure for the distance of the i-th unit from the j-th unit.
U are the values of the single cells P± transformed in knots K± of the artificial neural network and where The suffix i defines the central cell or knot and the suffix j defines the cells or knots directly surrounding the said central cell or knot.
6. A neural network according to one or more of the preceding claims 1 to 4 characterised in that the following rules are used: a) Learning u e [0,1] ; C = neighbors ; wtJ = 0.0001 Initialization W Awtj = ui - (uj - wij) (l - -^-) Uj w y[n+l] = w u[n] + Aw v.,
Figure imgf000080_0001
Outi (rem, auot) = mod(-^ MaxPixel,MaxPixel) N b) Recall
Figure imgf000080_0002
where
Ui is the value of the central cell or knot Ki Uj are the values of the surrounding knots Kj Wj indicates the weights of the connection of the surrounding knots Kj and the central knot K± . Wtj defines the weights of the connection in the n-th cycle .
Wy defines the weight of the connections in the n+l-th cycle . α Wij is the value which has to be added to the weights Wij in order to update then for the next cycle. Outi is the output value of the i-th unit corresponding to the i-th pixel which is the target unit or pixel New Wji is the new weight of the connection between j-th and i-th unit
New Wij is the new weight of the connection between i-th and j-th unit w is the mean value of the weight. The argument quot relating to the integer quotient of the output value of each unit and the number of steps in a scale of representing the said value. The argument rem relating to the rest of the above mentioned quotient.
7. A neural network according to one or more of the preceding claims 1 to 4 characterised in that only the following activation rules are used:
Center Neighbors State 1 1 1 1 0 0 0 1 0 0 0 1 Initialization : u, = input,
Figure imgf000081_0001
Δ . = Net, (1 - u , ) • Net, > 0 Δ,. = Net, u, a Net, < 0
where Ui is the input at the first step and the output of the central knot K± at the following steps Uj are the outputs of the surrounding knots Kj Neti is the net input of the central Knot Ki computed as a function of the outputs Uj of the surrounding knots Kj and the weights Wij of the connections of the the surrounding knots Kj to the central Knot K± . n is the number of cycle i is the update value of the output O± of the central knot Ki for computing the new output value for the next cycle as a function of Net input Net± and the actual output Ui of the central knot Ki α is a constant.
8. A Neural Network according to one or more of the preceding claims 1 to 4 , characterised in that only the following activation rules are used:
Center Neighbors State 1 1 1 1 0 0 0 1 0 0 0 0
Initialization : u, = input,
Figure imgf000082_0001
9 1 At = Nett (l - MI- ) - α - (σ — ) Nett > 0 9 1 Δ; = Nett «,- α (σ - -) Nefz- < 0
Figure imgf000083_0001
where
Ui is the input at the first step and the output of the central knot K± at the following steps
U are the outputs of the surrounding knots Kj Νeti is the net input of the central Knot Ki computed as a function of the outputs Uj of the surrounding knots K and the weights wAj of the connections of the the surrounding knots Kj to the central Knot K± . n is the number of cycle αi is the update value of the output U of the central knot Ki for computing the new output value for the next cycle as a function of Net input Neti and the actual output Ui of the central knot K± α is a constant.
9. A Nerual Network according to one or more of the preceding claims 1 to 4 , characterised in that only the following activation rules are used: Center Neighbors State 1 1 1 1 0 0 0 1 1 0 0 0
Initialization : u, = input, eti =(∑UjWij)-V> j 9 1 At =Neti \-Ui) a <5Δ — ) Nett >0 1 9 At =Neti-ui- -( — σ ) Nett<0
Figure imgf000084_0001
where
Ui is the input at the first step and the output of the central knot Ki at the following steps
Uj are the outputs of the surrounding knots K3 Νeti is the net input of the central Knot Ki computed as a function of the outputs Uj of the surrounding knots Kj and the weights Wij of the connections of the the surrounding knots Kj to the central Knot K±. n is the number of cycle αi is the update value of the output U± of the central knot Ki for computing the new output value for the next cycle as a function of Net input Neti and the actual output Ui of the central knot Ki α is a constant.
10 A neural Network according to one or more of the preceding claims 1 to 4, characterised in that only the following activation rules are used: Initialization : u, = input,
Neti =
Figure imgf000085_0001
Uj wtj σ 2 ( w, )) α k <≡ [1, N] ; k ≠ j j A, = Net, (1 - w.) Net, > 0 Δ . = Net, - u, a Net, < 0
Figure imgf000085_0002
where
Ui is the input at the first step and the output of the central knot K± at the following steps Uj are the outputs of the surrounding knots Kj
Neti is the net input of the central Knot Ki computed as a function of the outputs Uj of the surrounding knots Kj and the weights w j of the connections of the the surrounding knots Kj to the central Knot K±. n is the number of cycle αi is the update value of the output Ui of the central knot Ki for computing the new output value for the next cycle as a function of Net input Net± and the actual output Ui of the central knot K±
11. A Neural Network according to one or more of the preceding claims 1 to 4, characterised in that only the following activation rules are used: Initialization : u, = input,
Nett =
Figure imgf000086_0001
2(wjk)) a k e il>N]>' k ≠ j j 7 1 Δ, = Neti (1 - "/) oc (σ - -) Nett > 0 9 1 Δj = Neti M,- α (σ - -) Λføz- < 0
Figure imgf000086_0002
2 , k≠j W j = . σ (wik ) = 1 N > y l* N
Figure imgf000086_0003
where
U is the input at the first step and the output of the central knot Ki at the following steps
Uj are the outputs of the surrounding knots Kj Neti is the net input of the central Knot Ki computed as a function of the outputs Uj of the surrounding knots Kj and the weights w±j of the connections of the the surrounding knots Kj to the central Knot Ki. n is the number of cycle a± is the update value of the output Ui of the central knot Ki for computing the new output value for the next cycle as a function of Net input Net± and the actual output Ui of the central knot K±
12. A Neural Network according to one or more of the preceding claims 1 to 4 , characterised in that only the following activation rules are used: Initialization : u, - input,
Neti = (∑u j wij • σ 2 wjk )) α k ≡ [\,N];k ≠ j j 9 1 Ai = Nett - l - Ui) -(σ --) Neti >0 1 9 Ai = Neti «,• α (--σ ) Nett < 0
Figure imgf000087_0001
where
Ui is the input at the first step and the output of the central knot K± at the following steps Uj are the outputs of the surrounding knots Kj
Neti is the net input of the central Knot Ki computed as a function of the outputs Uj of the surrounding knots Kj and the weights wi3 of the connections of the the surrounding knots Kj to the central Knot K± . n is the number of cycle
<x± is the update value of the output O± of the central knot K for computing the new output value for the next cycle as a function of Net input Net± and the actual output Ui of the central knot Ki α is a constant.
13. A Neural Network according to one or more of the preceding claims 1 to 4, characterised in that only the following activation rules are used:
MaxPixelRange = 2M ; u, e [θ,l] ; α = β = l8 , ™° ; N = Intorno ;
Max = l;Min = 0;rest - 0.1; decay - 0.1;
Figure imgf000088_0001
J N ini, =uj -wϋ> wij < 0
Net i - (ecc , ) + (ini, β ) eNeti _ e-Neti
Act 1 ~ e N"e"tt' ++ re~N ~ etil Δ7- = (Max - u, ) Act, - decay ■ (u, - rest); Act, > 0
A, = (u, - Min) Act, - decay (u, - rest); Act, < 0
Figure imgf000088_0002
U is the output of the central knot Ki Uj are the outputs of the surrounding knots Kj Neti is the net input of the central Knot Ki computed as a function of the outputs Uj of the surrounding knots Kj and the weights w±j of the connections of the the surrounding knots Kj to the central Knot Ki. N is the number of cycle αi is the update value of the output Ui of the central knot Ki for computing the new output value for the next cycle as a function of Net input ti and the actual output Ui of the central knot K± α is a constant. And where the functions ecCi and inii are identical only a selection is made between inputs associated to positive and to negative weight connections and where an activation parameter Acti is provided as a function of the net input Neti to the knot Ki ; this parameter being used for choosing one of two different function for computing the update value OL± of the output Ui of the knot Ki depending on the fact whether Acti is negative or positive two functions being provided for computing αi comprising the following expressions: Max and Min which are defined as the top and the bottom range of the activation value. Decay which is a function defined as the usual decay value along the time of each unit. Rest which is defined as the default value toward each unit tends for. 14. A Neural Network according to one or more of the preceding claims 1 to 4 , characterised in that only the following activation rules are used: MaxPixelRange - 2M ; u, e [0,l] ; N = Surroundings;
Max - \;Min = 0; rest = 0 A decay = 0.1; N ecci = Λ ui - Uj Wjj ) ; Wy > 0 j N ini - \ (U - Uj Wjj ) ; Wy < 0 ecc ; ecc • Net i = ((l - —±-) - e 1+inij ini ,- 1 + ini { 1 + ec→ l ecc i c : e 1 +
Figure imgf000090_0001
Ai = (Max - Ui) - Acti ~ decay («; - rest); Acti > 0 Δ i = ui - Min) Acti ~ decay (Wj- - rest); Acti < 0
Δi -Δ,- [Λ+1] [«] , e l - e l u - u + 1 l e Δ, l- Λ ,- e -Δ l,- Ui is the output of the central knot Ki
U are the outputs of the surrounding knots Kj Neti is the net input of the central Knot Ki computed as a function of the outputs Uj of the surrounding knots Kj and the weights w j of the connections of the the surrounding knots Kj to the central Knot K±. N is the number of cycle αi is the update value of the output O± of the central knot Ki for computing the new output value for the next cycle as a function of Net input Neti and the actual output Ui of the central knot K± a is a constant. And where the functions ecc and inii are identical only a selection is made between inputs associated to positive and to negative weight connections and where an activation parameter Acti is provided as a function of the net input Neti to the knot KiJ; this parameter being used for choosing one of two different function for computing the update value i± of the output Ui of the knot K± depending on the fact whether Acti is negative or positive two functions being provided for computing <χ± comprising the following expressions: Max and Min which are defined as the top and the bottom range of the activation value. Decay which is a function defined as the usual decay value along the time of each unit. Rest which is defined as the default value toward each unit tends for.
14. A Neural Network to one or more of the preceding claims 1 to 4, characterised in that only the following activation rules are used: σ 2M w,j = π ;
Figure imgf000091_0001
tj = dy Nett cos(uj wtj )- uj ; or ij = dy Neti ' ∞siu j ' wij ) -sin(tt y w{J },
Figure imgf000091_0002
Figure imgf000091_0003
Whtax = Afø*H I WMin = Min{wij },
Scale =
Figure imgf000092_0001
Out = Scale Wf + Offset; where
Ui is the value of the central cell or knot K
Uj are the values of the surrounding knots Kj
Wij indicates the weights of the connection of the surrounding knots Kj and the central knot K± .
Outi is the output value of the i-th unit corresponding to the i-th pixel which is the target unit or pixel New Wji is the new weight of the connection between j-th and i-th unit w is the mean value of the weight.
15. A Neural Network according to one or more of the preceding claims characterized in that at each computation cycle a combination of a set or learning rules or functions and of a set of activation functions according to one or more of the preceding claims 5 to 14 is used for providing a new array of data records in which the value of each data record has been changed as a function of the outputs of the actual target data record and of the data records of the surrounding cells or points and as a function of the weights defining the connections between the target data record and the data records of the surrounding cells or points and for providing new values for the weights of the said connections which new array of data record and which new weights for the connections are used for carrying out the following computation cycle.
16. An artificial neural network according to one ro more of the preceding claims characterized in tha it is a image processing machine, the data records of the database univoquely associated or acting as knots of an array of knots (K±) being formed by the pixels (P) while tha initialisation values of the knots being formed by the pixels values (Ui.
17. An artificial neural Network as claimed in claim 16, characterised in that the pixel value is a scalar elemnt or a vector elemnt the pixel being characterised by different variables forming a component of the vector each variable being related to a physical and/or visual feature characterising the pixie and its visual aspect.
18. A Neural Network as claimed in claim 17, characterized in that the pixel is of the grey scale image and is characterized by its brightness.
19. A neural Network as claimed in claims 16 or 17, characterized in that the pixels are of a colour image and the variables characterising the pixel are at least three corresponding to a HSV or a RGB or another conventiona coding of pixel appearance.
20. An algorithm for recognizing relationships between data of a database the data being of the kind where the relative position of the data records in an array of data records or in a distribution of data records in a N-dimensional space, particularly a two or three dimensional space is a relevant feature of the data record and where the data records can be represented as an array of cells or points, each point being univocally related to a data record of the database and having a univocally defined position in the array relatively to the cells or points of the other data records, to each data record being further associated at least one variable or more variables each one having a certain value, the algorithm being characterised by the fact that - each cell or point in the array of cells of points representing a data records of a database is considered to be a unit or a knot of an Artificial Neural Network.
21. An artificial neural network according to claim 20 , in which - each unit or knot formed by a cell or point of the database being successively defined as a target unit or knot and connections being defined between each target unit or knot at least to each one of the rest of the units or knots formed by the rest of cells or points of the database which are at least of gradient 1 relatively to the corresponding target unit or knot; - a new output value of each unit or knot of the database successively defined as target unit or knot being calculated by means of the set of learning rules or functions or the set of activation rules or functions of the artificial neural network or by means of the combination of both the set of learning rules or functions and the set of activation rules or functions of the artificial neural network as a function of the actual output of the units or knots connected to the target unit or knot and of the actual output of the said target unit or knot; - The actual outputs of each unit or knot being defined as the value of the variable or as the values of the variables associated to each data record represented by a cell or a point considered as a unit or knot of the artificial neural network;
- And the new output of the target unit or knot is considered as the new value of the variable or of the variables of the data record associated to the cell or point of the array of data records corresponding to the target unit or knot;
- By carrying out the said steps for computing a new output of a target unit or knot for at least part or for each cell or point of the array of data records a new array of data records is computed where the data record of each cell or point has a new value of the at least one variable or new values for the variables as a result of a first computation cycle of the artificial neural network according to the above steps ;
- The said computation cycle being repeated for each successive new array of data records until a certain prefixed number of repetitions of the computation cycle has been carried out and/or unless a certain maximum allowable error or discrepancy has been reached between the original values of the variable or of the variables of the original array of data records with respect to the values of the variable or variables of the array of data records according to the one computed in the last cycle and/or unless the difference between the value of the variable or the values of the variables of data records in the sequence of array of data records computed in the sequence of cycles is lower than a predetermined maximum rate.
22. An algorithm according to claim 20 or 21, characterised in that the array of data records is submitted at least twice or more times to elaboration with the said algorithm in a first elaboration phase being provided a first set of learning functions or rules or a first set of activation functions or rules or a combination thereof and in a second elaboration phase being provided a second set of learning functions or rules or a second set of activation functions or rules or a combination thereof and so on when more then two phases are provided while the array of data records being used in the second or further elaboration phases with the algorithm with the second or further different sets of learning or activation rules or functions or with a combination thereof is the array of data records resulting respectively from the first or from the previous elaboration phase of the array of data records .
23. An algorithm according to one or more of the preceding claims 20 to 22 , characterised in that particularly in a two or three dimensional array of data records , the cells or points related to the data records of gradient 1 with respect to the cell or point of the target data record are formed by the data records associated to the cells or points of the data record array which directly surrounds the cell or point in the array of data record related to the said target data record.
24. An algorithm according to one or more of the preceding claims 20 to 23, characterised in that the new array of data records computed by the algorithm is based only on a set of learning functions or rules for optimizing the eights of te connections, the new output of each target data record being defined as a function of new weights characterising the connection of each target unit or knot associated to the target data record with the units or knots represented by the cells or points of data records of gradient one relatively to the cell or point of the target data record, the set of learning rules or functions defining new weights of the connections as a function of the previous weights computed or defined in a previous computation cycle and as a function of the actual outputs of the unit or knot associated to the cell or point of target data record and of the unit or knot associated to the cells or points of the data records of at least of gradient 1 or of the data records of the cells or points directly surrounding the cells or point of the actual target data record.
25. An algorithm according to one or more of the preceding claims 20 to 23, characterised in that the new array of data records computed by the algorithm is based only on a set of activation functions where these functions defines the new output of the target unit or knot corresponding to the cell or point related to the target data record basing on the net input to the said target unit or knot which is a function of the outputs of the units or knots corresponding to the cells or points associated to the data records at least of gradient 1 with respect to the target data record, particularly to the units or knots corresponding to the cells or points of the array of data records directly surrounding the cell or point of the target data record.
26. An algorithm according to one or more of the preceding claims 20 to 25 characterised in that computation in the first computation cycle starts with a fixed predetermined value of the weights for each connection while the starting value of the unit or knot is modified according to a predetermined function which is also function of the weights and of the value of the surrounding knots or units and therefore to the data records which corresponds to cells or points in the array directly surrounding the cell or point representing a certain unit or knot of the artificial neural network.
27. An algorithm according to one or more of the preceding claims 20 to 26 characterised in that only the following learning rules are used:
,N
Figure imgf000098_0001
e Ri Ji — e ~Ri Jj w. y + e V
where α is a parameter which can freely defined by the user and which renders the algorithm more or less sensible to differences in the image.
Rij is some sort of measure for the distance of the i-th unit from the j-th unit. Ui are the values of the single cells P± transformed in knots Ki of the artificial neural network and where The suffix i defines the central cell or knot and the suffix j defines the cells or knots directly surrounding the said central cell or knot.
28. An algorithm according to one or more of the preceding claims 20 to 27 characterised in that the following learning rules are used: a) Learning w e [0,1] ; C = neighbors ; wt] . = 0.0001 Initialization wij Awtj = Ui - (Uj Wij ) (l - -—) Uj
W y^ = w u!»] + Δw V..
Figure imgf000099_0001
Outi (rem, quot) — mod(^^ MaxPixel,MaxPixel) N b) Recall
Figure imgf000099_0002
where
U is the value of the central cell or knot K Uj are the values of the surrounding knots Kj Wij indicates the weights of the connection of the surrounding knots Kj and the central knot K±.
W~ defines the weights of the connection in the n-th cycle .
Wtj defines the weight of the connections in the n+l-th cycle . α ij is the value which has to be added to the weights
Wij in order to update then for the next cycle.
Outi is the output value of the i-th unit corresponding to the i-th pixel which is the target unit or pixel New Wji is the new weight of the connection between j-th and i-th unit
New ij is the new weight of the connection between i-th and j-th unit w is the mean value of the weight. The argument quot relating to the integer quotient of the output value of each unit and the number of steps in a scale of representing the said value. The argument rem relating to the rest of the above mentioned quotient.
29. An Algorithm according to one or more of the preceding claims 20 to 28 characterised in that only the following activation rules are used:
Center Neighbors State 1 1 1 1 0 0 0 1 0 0 0 1
Initialization : u, = input ,
Net, = ( UjWy ) a
A, = Net, (l - u,) Net, > 0 Δ . = Net, u, a Net, < 0 W . — Ui + ;
where Ui is the input at the first step and the output of the central knot K± at the following steps
Uj are the outputs of the surrounding knots Kj
Neti is the net input of the central Knot K± computed as a function of the outputs Uj of the surrounding knots K and the weights w±j of the connections of the the surrounding knots Kj to the central Knot Ki. n is the number of cycle αi is the update value of the output TJ± of the central knot Ki for computing the new output value for the next cycle as a function of Net input Neti and the actual output Ui of the central knot K± α is a constant.
30. An Algorithm according to one or more of the preceding claims 20 to 27, characterised in that only the following activation rules are used:
Center Neighbors State 1 1 1 1 0 0 0 1 0 0 0 0 Initialization : u, = input,
Neti = (∑UjWij ) - U
9 1 Aj = Net (l - Ui ) - a (σ z - -) Nett > 0 9 1 Ai = Neti M,- - α (σ - -) Nett < 0 w ,f[»+"]J _ = , κ}B]J + Δ. u =
Figure imgf000102_0001
where Ui is the input at the first step and the output of the central knot K± at the following steps Uj are the outputs of the surrounding knots Kj Neti is the net input of the central Knot Ki computed as a function of the outputs Uj of the surrounding knots Kj and the weights Wij of the connections of the the surrounding knots Kj to the central Knot Ki. n is the number of cycle αi is the update value of the output Ui of the central knot Ki for computing the new output value for the next cycle as a function of Net input Neti and the actual output Ui of the central knot K± α is a constant.
31. An Algorithm according to one or more of the preceding claims 20 to 27 , characterised in that only the following activation rules are used:
Center Neighbors State 1 1 1 1 0 0 0 1 1 0 0 0
Initialization : u, = input, Neti = (∑UjWij ) -a
9 1 At = Neti l - Ui) (a - -) Nett >0 1 9 Δj = Nett Uj a (—-σ ) Nett < 0 u ["+1]=MW+Δ.
« =
Figure imgf000103_0001
where
Ui is the input at the first step and the output of the central knot Ki at the following steps
Uj are the outputs of the surrounding knots Kj Neti is the net input of the central Knot K± computed as a function of the outputs Uj of the surrounding knots Kj and the weights w±j of the connections of the the surrounding knots Kj to the central Knot K± . n is the number of cycle αi is the update value of the output Ui of the central knot Ki for computing the new output value for the next cycle as a function of Net input Neti and the actual output Ui of the central knot Ki α is a constant.
32. An Algorithm according to one or more of the preceding claims 20 to 27 , characterised in that only the following activation rules are used: Initialization : u, = input,
Neti = ∑UjWij σ 2(wjk)) a k e [l,N];k ≠ j A, = Net, • (1 - u,) Net, > 0 Δf = Net, u Net, < 0
Figure imgf000104_0001
where
Ui is the input at the first step and the output of the central knot Ki at the following steps
Uj are the outputs of the surrounding knots Kj
Neti is the net input of the central Knot Ki computed as a function of the outputs Uj of the surrounding knots Kj and the weights Wij of the connections of the the surrounding knots K to the central Knot Ki. n is the number of cycle α is the update value of the output Ui of the central knot K for computing the new output value for the next cycle as a function of Net input Neti and the actual output Ui of the central knot K±
33. An Algorithm according to one or more of the preceding claims 20 to 27 , characterised in that only the following activation rules are used: Initialization : u, = input,
Figure imgf000104_0002
j i 1 Δ, = Neti (1 - M,- ) « (σ - -) Nett > 0 1 Δz- = Nett ut a Δ - -) Neti < 0 U ,[»+!] n\ = U, [»] + Δ,
Figure imgf000105_0001
where
Ui is the input at the first step and the output of the central knot Ki at the following steps
Uj are the outputs of the surrounding knots Kj Neti is the net input of the central Knot K± computed as a function of the outputs Uj of the surrounding knots Kj and the weights wiS of the connections of the the surrounding knots Kj to the central Knot Ki . n is the number of cycle i is the update value of the output Ui of the central knot Ki for computing the new output value for the next cycle as a function of Net input Neti and the actual output Ui of the central knot Ki
34. An Algorithm according to one or more of the preceding claims 20 to 27 , characterised in that only the following activation rules are used: Initialization : u, = input,
Neti = (∑UjWij G 2(wjk)) a k e [l,N]',k ≠ j
Figure imgf000106_0001
1 9 Ai = Neti ui a - -<5 ) Nett < 0
U[n+M = u [n + Δ _
Figure imgf000106_0002
where
U is the input at the first step and the output of the central knot K± at the following steps
Uj are the outputs of the surrounding knots Kj Νeti is the net input of the central Knot K± computed as a function of the outputs Uj of the surrounding knots K and the weights Wij of the connections of the the surrounding knots Kj to the central Knot K±. n is the number of cycle αi is the update value of the output Ui of the central knot Ki for computing the new output value for the next cycle as a function of Net input Neti and the actual output Ui of the central knot Ki α is a constant.
35. An Algorithm according to one or more of the preceding claims 20 to 27 , characterised in that only the following activation rules are used:
MaxPixelRange = 2M ; u, e [θ,l] ; α = β = lgU ; N = Intorno ; 2 Max - l;Min = 0;rest = 0.1; decay = 0.1; N ecc, = «, Wj, ; w„ > 0 J N iniι = ∑uj w„ Wij < 0
Net , - (ecc , cc ) + (ini, β )
Figure imgf000107_0001
A, = (Max - u, ) • Act, - decay (u, - rest); Act, > 0
A, = (u, -Min) - Act, - decay (u, - rest); Act, < 0
Figure imgf000107_0002
Ux is the output of the central knot Kx U-, are the outputs of the surrounding knots K-,
Neti is the net input of the central Knot Kx computed as a function of the outputs U-, of the surrounding knots K-, and the weights wι;) of the connections of the the surrounding knots K to the central Knot Kτ. N is the number of cycle αx is the update value of the output Uj. of the central knot Kj. for computing the new output value for the next cycle as a function of Net input Neti and the actual output Ux of the central knot Kx α is a constant. And where the functions ecci and inij. are identical only a selection is made between inputs associated to positive and to negative weight connections and where an activation parameter Actx is provided as a function of the net input Netx to the knot Kxj, this parameter being used for choosing one of two different function for computing the update value a± of the output Ui of the knot K± depending on the fact whether Acti is negative or positive two functions being provided for computing a± comprising the following expressions: Max and Min which are defined as the top and the bottom range of the activation value. Decay which is a function defined as the usual decay value along the time of each unit. Rest which is defined as the default value toward each unit tends for.
36. An Algorithm according to one or more of the preceding claims 20 to 27 , characterised in that only the following activation rules are used: MaxPixelRange - 2M ; u, e [θ,l] ; N - Surroundings;
Max = I; Min = 0;rest = A;decay = 0.1; N ecci = (Wj -Uj Wij) ; Wy > 0 j N inii = ∑ («ι -uj wij)2,' wij < 0 ecc , ini, Net, =((1- ) - e l+Mi )-((!- ) - e +ecc i 1 + ini , 1 + ecc i
Figure imgf000108_0001
Δ,- = (Max -ut)- Act{ ~ decay (wz- - rest); Acti > u
Δ = (« - Min) Actt ~ decay (ut - rest); Actt < 0
Figure imgf000109_0001
Ui is the output of the central knot K± Uj are the outputs of the surrounding knots Kj Neti is the net input of the central Knot Ki computed as a function of the outputs Uj of the surrounding knots Kj and the weights Wj of the connections of the the surrounding knots Kj to the central Knot K± . N is the number of cycle a± is the update value of the output Ui of the central knot Ki for computing the new output value for the next cycle as a function of Net input Neti and the actual output Ui of the central knot Ki α is a constant. And where the functions ecci and inii are identical only a selection is made between inputs associated to positive and to negative weight connections and where an activation parameter Acti is provided as a function of the net input Neti to the knot KiJ; this parameter being used for choosing one of two different function for computing the update value a± of the output Ui of the knot K± depending on the fact whether Acti is negative or positive two functions being provided for computing αi comprising the following expressions : Max and Min which are defined as the top and the bottom range of the activation value. Decay which is a function defined as the usual decay value along the time of each unit. Rest which is defined as the default value toward each unit tends for.
37. An Algorithm according to one or more of the preceding claims 20 to 27 , characterised in that only the following activation rules are used: k = >M
Figure imgf000110_0001
N Ne Σshϊ(uj-Wi ) k; [04}
Figure imgf000110_0002
Atj = dy Nett COS(UJ wϋ)- Uj ; or
Ay = dtj Neti • cos[uj wtj ) -sin(wy- • wtj \
Figure imgf000110_0003
wMax = MaxyWy \ WMin = Min^Vy \
, M Scale = ; Offset = - WMin-l WMaχ-WMin WMax ~ WMin
Outi = Scale - Wj + Offset; where
Ui is the value of the central cell or knot K± Uj are the values of the surrounding knots Kj ij indicates the weights of the connection of the surrounding knots K and the central knot Ki. Outi is the output value of the i-th unit corresponding to the i-th pixel which is the target unit or pixel
New Wji is the new weight of the connection between j-th and i-th unit w is the mean value of the weight.
38. An Algorithm according to one or more of the preceding claims 20 to 37 characterized in that at each computation cycle a combination of a set or learning rules or functions and of a set of activation functions according to one or more of the preceding claims 5 to 16 is used for providing a new array of data records in which the value of each data record has been changed as a function of the outputs of the actual target data record and of the data records of the surrounding cells or points and as a function of the weights defining the connections between the target data record and the data records of the surrounding cells or points and for providing new values for the weights of the said connections which new array of data record and which new weights for the connections are used for carrying out the following computation cycle.
39. A method for image processing in which the image is formed by a two or three dimensional array of pixels and in which each pixel of the array forms a unit or knot of an artificial neural network, the input and the output of the artificial neural network being formed by the original values of the pixels corresponding to each unit and by the computed value of each pixel , the computation of the output value of each one of the knots being carried out as a function of the values at least of the pixels surrounding the said knot.
40. A method according to claim 39 , characterised in that between each knot or unit of the artificial neural network corresponsing univoquely to a pixel of the array of pixels forming the image and the related units or knots of at least gradient 1 weighted connections are provided.
41. A method according to claims 39 and 40, characterised in that a weight optimization is carried out.
42. A method according to claim 41 , characterised in that unit activation is carried out after weight optimization .
43. A method according to one or more of the preciding claims 39 to 42, characterized in that weight evolution and unit activation is carried out.
44. A method according to one or more of the preceding claims 39 to 43 characterized in that it is carried out by means of an artificial neural network according to claims 1 to 19 eventually implemented as an algorithm according to claims 20 to 38.
45 A method according to one or more of claims 39 to 44 , characterised in that it is a method for image pattern recognition in which the image is a digital image or an analogical image which has been digitalized, the array of image data being formed by a finite number of points or cells each one corresponding to an image unitary element so called pixel or voxel and each one pixel or voxel being related to a value of a parametric variable describing the intensity or the level of grey of the pixels in a grey scale image or each pixel or voxel being related to a vector each component of the said vector being a parametric variable describing the intensity of the pixel or voxel and the colour of the pixel or voxel , characterised in that the said array of image data is subjected to processing with an artificial neural network according to one or more of the preceding claims 1 to 19, being the cells or points of the data records of gradient 1 formed by the pixels or voxels directly surrounding a target pixel or voxel, the said algorithm being applied a certain number of times for carrying out a certain number of repetitions of the computation cycle, which number is a fixed defined number, or is computed on the basis of a variance or difference of the output values for the pixels at a certain repetition of the computation cycle and of the original values of the image data array, or which number of repetitions of the computation cycle is determined as the number of repetition at which the following computation cycles furnishes an output for the image data array which difference from the output for the image data array of the previous computation cycle is less than a certain predetermined difference .
46. A method for image pattern recognition according to claim 45 , in which the artificial neural network is implemented as an algorithm according to claims 20 to 39.
47 A method according to claim 46, charaterised in that the algorithm according to claims 20 to 39 is applied two times in succession one of the other at each time being used a different set of learning rules or functions or a different set of activation rules or functions or a different combination thereof.
48. A method or image pattern recognition according to one or more of the preceding claims 45 to 47, characterised in that the algorithm according to claims 20 to 39 is applied in combination with a predictive algorithm learned for recognising features of different zones of an imaged subject.
49. A method according to one or more of the preceding claims 45 to 47 wherein the predictive algorithm is an artificial neural network and the learning and testing database is formed by a set of images of identical subjects, the different imaged zones of the said image subject being univoquely identified for each image of the training and testing database, while the prediction algorithm is carried out on the image data array obtained as an output of the elaboration carried out with the algorithm according to one or more of claims 20 to 39.
50. The method according to claim 49, in which the prediction algorithm is carried out also using the original image data array as an input data while the prediction results obtained by the carrying out of the prediction algorithm on the image data array computed with the algorithm according to one or more of the preceding claims 20 to 39 and the prediction results obtained carrying out the prediction algorithm on the original image data array are combined or compared.
51. A method for carrying out contrast imaging or harmonic imaging in biological tissues without providing the presence of contrast media, characterised in that an ultrasound or MRI or radiographic image of a certain body or of a certain part of a body is acquired and the acquired image data array is subjected to a method according to one or more of the preceding claims 39 to 50 using a neural network according to one or more of the preceding claims 1 to 19 eventually implemented as an algorithm according to one or more of the preceding claims 20 to 39.
52. A method according to claim 51 , characterized in that the neural network according to claims 1 to 4 is provided with a set of rules according to claim 5 eventually implemented as an algorithm according to one or more of the preceding claims 20 to 27 in combination with at least a set of learning rules according to claim 28.
53. A method according to claim 52 , characterised in that the neural network neural network according to claims 1 to 4 is provided with a combination of rules according to claim 5 with a set of rules according to one of the claims 6 to 19 and eventually being the said neural network implemented as an algorithm according to one or more of the preceding claims 20 to 27 in combination with a set of learning rules or functions according to claim 28 and with a set of learning rules or functions according to one of the claims 29 to 17. 54 A method according to claim 52 , characterised in that it uses a set of learning and or activation rules according to one or more of the preceding claims 28 to 38. 55. A method according to claim 54 , characterised in that at least two image elaboration stages are provided the first elaboration stage being carried out with one set of learning and /or activation rules while the second elaboration stage is carried out by submitting the pixel values of the output image elaborated in the first stage to a second elaboration with a second set of learning and/or activation rules. 56 A method according to claims 54 or 55, characterised in that at least a third or more elaboration stages are provided each one carried out with a different set of learning and/or activation rules . 57 A method according to one or more of the claims 54 to 56, characterised in that the learning and/or activation rules are according to claims 5 to 19 or 27 to 38. 58. A method for helping in identifying tumoral tissues characterised in that it provides the steps of acquiring a digital image or a digitalized analogical image of the anatomical district containing the tumoral tissues; elaborating the said digital or digitalize image by means of the method according to one or more of the preceding claims 39 to 57. 59. A method for helping in identifying stenosis in blood vessels characterised in that it provides the steps of acquiring a digital image or a digitalized analogical image of the anatomical district containing the tumoral tissue; elaborating the said digital or digitalize image by means of the method according to one or more of the preceding claims 39 to 57. 60. A method for helping in identifying calcifications in biological tissues characterised in that it provides the steps of acquiring a digital image or a digitalized analogical image of the anatomical district containing the tumoral tissue; elaborating the said digital or digitalize image by means of the method according to one or more of the preceding claims 39 to 57.
PCT/EP2004/051821 2003-08-22 2004-08-18 Neural network for processing arrays of data with existent topology, such as images, and application of the network WO2005020132A1 (en)

Priority Applications (3)

Application Number Priority Date Filing Date Title
EP04766522A EP1656632B1 (en) 2003-08-22 2004-08-18 Neural network for processing arrays of data with existent topology, such as images, and application of the network
US10/569,354 US7877342B2 (en) 2003-08-22 2004-08-18 Neural network for processing arrays of data with existent topology, such as images and application of the network
JP2006523636A JP2007503629A (en) 2003-08-22 2004-08-18 Neural network for processing a data array having existing phase information such as an image and method of using the network

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
EP03425559A EP1508872A1 (en) 2003-08-22 2003-08-22 An algorithm for recognising relationships between data of a database and a method for image pattern recognition based on the said algorithm
EP03425559.6 2003-08-22

Publications (1)

Publication Number Publication Date
WO2005020132A1 true WO2005020132A1 (en) 2005-03-03

Family

ID=34043032

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2004/051821 WO2005020132A1 (en) 2003-08-22 2004-08-18 Neural network for processing arrays of data with existent topology, such as images, and application of the network

Country Status (5)

Country Link
US (1) US7877342B2 (en)
EP (2) EP1508872A1 (en)
JP (1) JP2007503629A (en)
CN (1) CN1839397A (en)
WO (1) WO2005020132A1 (en)

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1712182A1 (en) 2005-04-14 2006-10-18 Esaote S.p.A. Method of ultrasonic detection and localization of contrast agent microbubbles and method for local drug administration by using microbubble carriers
WO2009007285A1 (en) * 2007-07-06 2009-01-15 Bracco Imaging Spa Image processing using neural network
JP2009528117A (en) * 2006-03-03 2009-08-06 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Identifying image feature sets for assessing image similarity
US20100217145A1 (en) * 2006-06-09 2010-08-26 Bracco Spa Method of processing multichannel and multivariate signals and method of classifying sources of multichannel and multivariate signals operating according to such processing method
CN105469423A (en) * 2015-11-16 2016-04-06 北京师范大学 Online target tracking method based on continuous attractor neural network

Families Citing this family (38)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8167803B2 (en) * 2007-05-16 2012-05-01 Verathon Inc. System and method for bladder detection using harmonic imaging
US20090143668A1 (en) * 2007-12-04 2009-06-04 Harms Steven E Enhancement of mri image contrast by combining pre- and post-contrast raw and phase spoiled image data
US8855387B2 (en) 2007-12-28 2014-10-07 Exini Diagnostics Ab System for detecting bone cancer metastases
JP2012503501A (en) 2008-09-25 2012-02-09 シーエーイー ヘルスケア インク Simulation of medical image diagnosis
US8812414B2 (en) 2011-05-31 2014-08-19 International Business Machines Corporation Low-power event-driven neural computing architecture in neural networks
US8909576B2 (en) 2011-09-16 2014-12-09 International Business Machines Corporation Neuromorphic event-driven neural computing architecture in a scalable neural network
CN102663428B (en) * 2012-03-29 2014-11-26 中国科学院上海光学精密机械研究所 Neutral network mode identification system and mode identification method thereof
CA2894791C (en) * 2012-12-13 2023-10-31 Gagan SIDHU Processing multidimensional signals
US9558443B2 (en) 2013-08-02 2017-01-31 International Business Machines Corporation Dual deterministic and stochastic neurosynaptic core circuit
US11049006B2 (en) 2014-09-12 2021-06-29 Microsoft Technology Licensing, Llc Computing system for training neural networks
US11080587B2 (en) * 2015-02-06 2021-08-03 Deepmind Technologies Limited Recurrent neural networks for data item generation
WO2018081354A1 (en) 2016-10-27 2018-05-03 Progenics Pharmaceuticals, Inc. Network for medical image analysis, decision support system, and related graphical user interface (gui) applications
CN110574041B (en) 2017-02-23 2023-06-20 诺基亚技术有限公司 Collaborative activation for deep learning domain
US10783432B2 (en) * 2017-04-14 2020-09-22 International Business Machines Corporation Update management for RPU array
EP3400878B1 (en) 2017-05-10 2020-03-25 Esaote S.p.A. Method for postural independent location of targets in diagnostic imagines acquired by multimodal acquisitions and system for carrying out the said method
CN107480780A (en) * 2017-09-03 2017-12-15 胡明建 A kind of more threshold values polygamma functions select the design method of output artificial neuron more
CN107491810A (en) * 2017-09-23 2017-12-19 胡明建 A kind of design method of more threshold values feedback artificial neurons
CN107633300A (en) * 2017-09-28 2018-01-26 胡明建 A kind of design method of graded potential formula artificial neuron
CN107657315A (en) * 2017-09-30 2018-02-02 胡明建 A kind of graded potential formula selects the design method of end output artificial neuron
CN107977662B (en) * 2017-11-06 2020-12-11 清华大学深圳研究生院 Layered calculation method for realizing high-speed processing of computer visual image
US10973486B2 (en) 2018-01-08 2021-04-13 Progenics Pharmaceuticals, Inc. Systems and methods for rapid neural network-based image segmentation and radiopharmaceutical uptake determination
IT201800001148A1 (en) * 2018-01-17 2019-07-17 Semeion Centro Ricerche METHOD OF IMAGE PROCESSING AND SYSTEM FOR IMAGE PROCESSING
KR102060895B1 (en) * 2018-06-01 2020-02-11 연세대학교 산학협력단 Method for producing medical image and device for producing medical image
US10685261B2 (en) * 2018-06-11 2020-06-16 GM Global Technology Operations LLC Active segmention of scanned images based on deep reinforcement learning for OCR applications
WO2020144134A1 (en) 2019-01-07 2020-07-16 Exini Diagnostics Ab Systems and methods for platform agnostic whole body image segmentation
CN113710159A (en) 2019-04-24 2021-11-26 普罗热尼奇制药公司 System and method for automated and interactive analysis of bone scan images to detect metastasis
KR102254971B1 (en) * 2019-07-24 2021-05-21 가톨릭대학교 산학협력단 Method and apparatus for converting contrast enhanced image and non-enhanced image using artificial intelligence
US11564621B2 (en) 2019-09-27 2023-01-31 Progenies Pharmacenticals, Inc. Systems and methods for artificial intelligence-based image analysis for cancer assessment
US11900597B2 (en) 2019-09-27 2024-02-13 Progenics Pharmaceuticals, Inc. Systems and methods for artificial intelligence-based image analysis for cancer assessment
CN110852993B (en) * 2019-10-12 2024-03-08 拜耳股份有限公司 Imaging method and device under action of contrast agent
CN111310377B (en) * 2020-02-21 2022-03-15 北京航空航天大学 Non-probability reliability topological optimization design method for continuum structure
JP7407618B2 (en) * 2020-02-25 2024-01-04 謙一 井上 information processing system
US11321844B2 (en) 2020-04-23 2022-05-03 Exini Diagnostics Ab Systems and methods for deep-learning-based segmentation of composite images
US11386988B2 (en) 2020-04-23 2022-07-12 Exini Diagnostics Ab Systems and methods for deep-learning-based segmentation of composite images
US11721428B2 (en) 2020-07-06 2023-08-08 Exini Diagnostics Ab Systems and methods for artificial intelligence-based image analysis for detection and characterization of lesions
IT202000023257A1 (en) 2020-10-02 2022-04-02 Esaote Spa SYSTEM AND METHOD FOR DIAGNOSTIC IMAGING OF WATER AND FAT
EP3985409A1 (en) 2020-10-19 2022-04-20 Esaote S.p.A. Method for correcting inhomogeneity of the static magnetic field particularly of the static magnetic field generated by the magnetic structure of a machine for acquiring nuclear magnetic resonance images and mri system for carrying out such method
CN112927217B (en) * 2021-03-23 2022-05-03 内蒙古大学 Thyroid nodule invasiveness prediction method based on target detection

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5140670A (en) * 1989-10-05 1992-08-18 Regents Of The University Of California Cellular neural network

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5912986A (en) * 1994-06-21 1999-06-15 Eastman Kodak Company Evidential confidence measure and rejection technique for use in a neural network based optical character recognition system
US5790758A (en) * 1995-07-07 1998-08-04 The United States Of America As Represented By The Secretary Of The Navy Neural network architecture for gaussian components of a mixture density function
DE69627154D1 (en) * 1996-10-17 2003-05-08 St Microelectronics Srl Modular fuzzy logic neural network
US6028956A (en) * 1997-04-04 2000-02-22 Kofile Inc. Object location and span determination method and apparatus which determines a location and span of an object in an image
JP3069058B2 (en) * 1997-04-07 2000-07-24 株式会社エイ・ティ・アール人間情報通信研究所 Neurocellular automata and optimizers
US6011295A (en) * 1997-07-22 2000-01-04 Foveonics, Inc. Neural network active pixel cell
US6801655B2 (en) * 2001-05-10 2004-10-05 The United States Of America As Represented By The Secretary Of The Navy Spatial image processor

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5140670A (en) * 1989-10-05 1992-08-18 Regents Of The University Of California Cellular neural network

Non-Patent Citations (5)

* Cited by examiner, † Cited by third party
Title
CHUA L O ET AL: "CELLULAR NEURAL NETWORKS: APPLICATIONS", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS, IEEE INC. NEW YORK, US, vol. 35, no. 10, 1 October 1988 (1988-10-01), pages 1273 - 1290, XP000119052 *
GUZELIS C ET AL: "Recurrent perceptron learning algorithm for completely stable cellular neural networks", CELLULAR NEURAL NETWORKS AND THEIR APPLICATIONS, 1994. CNNA-94., PROCEEDINGS OF THE THIRD IEEE INTERNATIONAL WORKSHOP ON ROME, ITALY 18-21 DEC. 1994, NEW YORK, NY, USA,IEEE, 18 December 1994 (1994-12-18), pages 177 - 182, XP010131320, ISBN: 0-7803-2070-0 *
L. DIAPPI, M. BUSCEMA ET AL.: "The urban sprawl dynamics: does a Neural Network understand the spatial logic better than a Cellular Automata?", 42ND ERSA CONGRESS, 27 August 2002 (2002-08-27) - 31 August 2002 (2002-08-31), Dortmund, pages 1 - 20, XP002267589, Retrieved from the Internet <URL:http://www.raumplanung.uni-dortmund.de/rwp/ersa2002/cd-rom/papers/033.pdf> [retrieved on 20040121] *
SAHOTA P ET AL: "Training genetically evolving cellular automata for image processing", SPEECH, IMAGE PROCESSING AND NEURAL NETWORKS, 1994. PROCEEDINGS, ISSIPNN '94., 1994 INTERNATIONAL SYMPOSIUM ON HONG KONG 13-16 APRIL 1994, NEW YORK, NY, USA,IEEE, 13 April 1994 (1994-04-13), pages 753 - 756, XP010121398, ISBN: 0-7803-1865-X *
WEICKERT J: "FOUNDATIONS AND APPLICATIONS OF NONLINEAR ANISOTROPIC DIFFUSION FILTERING", ZEITSCHRIFT FUER ANGEWANDTE MATHEMATIK UND MECHANIK, VDI VERLAG, BERLIN, DE, vol. 76, no. 1, 3 July 1995 (1995-07-03), pages 283 - 286, XP008024946, ISSN: 0044-2267 *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1712182A1 (en) 2005-04-14 2006-10-18 Esaote S.p.A. Method of ultrasonic detection and localization of contrast agent microbubbles and method for local drug administration by using microbubble carriers
US7981040B2 (en) 2005-04-14 2011-07-19 Esaote, S.P.A. Method of ultrasonic detection and localization of contrast agent microbubbles and method for local drug administration by using microbubble carriers
JP2009528117A (en) * 2006-03-03 2009-08-06 コーニンクレッカ フィリップス エレクトロニクス エヌ ヴィ Identifying image feature sets for assessing image similarity
US20100217145A1 (en) * 2006-06-09 2010-08-26 Bracco Spa Method of processing multichannel and multivariate signals and method of classifying sources of multichannel and multivariate signals operating according to such processing method
WO2009007285A1 (en) * 2007-07-06 2009-01-15 Bracco Imaging Spa Image processing using neural network
CN105469423A (en) * 2015-11-16 2016-04-06 北京师范大学 Online target tracking method based on continuous attractor neural network
CN105469423B (en) * 2015-11-16 2018-06-22 北京师范大学 A kind of online method for tracking target based on continuous attraction sub-neural network

Also Published As

Publication number Publication date
EP1656632A1 (en) 2006-05-17
US7877342B2 (en) 2011-01-25
US20070233624A1 (en) 2007-10-04
CN1839397A (en) 2006-09-27
EP1508872A1 (en) 2005-02-23
EP1656632B1 (en) 2012-10-03
JP2007503629A (en) 2007-02-22

Similar Documents

Publication Publication Date Title
EP1656632B1 (en) Neural network for processing arrays of data with existent topology, such as images, and application of the network
Wang et al. Simulating human saccadic scanpaths on natural images
Haehn et al. Evaluating ‘graphical perception’with CNNs
Deco et al. Attention, short-term memory, and action selection: a unifying theory
Coen-Cagli et al. Cortical surround interactions and perceptual salience via natural scene statistics
Sharma et al. Brain tumor segmentation using genetic algorithm and artificial neural network fuzzy inference system (ANFIS)
Antolik et al. Development of maps of simple and complex cells in the primary visual cortex
Gahukar et al. Segmentation of MRI brain image using fuzzy c means for brain tumor diagnosis
Sewell et al. Exemplifying “Us”: Integrating social identity theory of leadership with cognitive models of categorization
Alomoush et al. Fully automatic grayscale image segmentation based fuzzy C-means with firefly mate algorithm
Huber et al. Benchmarking perturbation-based saliency maps for explaining Atari agents
Heena et al. Machine learning based biomedical image processing for echocardiographic images
Młynarski et al. Efficient coding theory of dynamic attentional modulation
US20030228054A1 (en) Neurodynamic model of the processing of visual information
Walshe et al. Detection of occluding targets in natural backgrounds
Govinda et al. Artificial neural networks in UWB image processing for early detection of breast cancer
Fruend Constrained sampling from deep generative image models reveals mechanisms of human target detection
Gupta An Integrated Computational Model of Visual Search Combining Eccentricity, Bottom-up, and Top-down Cues
Chatterjee et al. Computational Analysis Reveals Sensitivity to Contrast in Simple Cells of Primary Visual Cortex
Lindsay et al. Deep Learning Networks and Visual Perception
Nguyen Explainable Artificial Intelligence for Image Quality Assessment
Fruend Simple, biologically informed models, but not convolutional neural networks describe target detection in naturalistic images
Jang Exploring the robust nature of human visual object recognition through comparisons with convolutional neural networks
Henderson et al. Human frontoparietal cortex represents behaviorally-relevant target status during invariant object recognition
Suja et al. EAD-HNN: ENHANCED ALZHEIMER'S DISEASE EARLY DETECTION USING HYBRID NEURAL NETWORKS

Legal Events

Date Code Title Description
WWE Wipo information: entry into national phase

Ref document number: 200480024189.2

Country of ref document: CN

AK Designated states

Kind code of ref document: A1

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BR BW BY BZ CA CH CN CO CR CU CZ DE DK DM DZ EC EE EG ES FI GB GD GE GH GM HR HU ID IL IN IS JP KE KG KP KR KZ LC LK LR LS LT LU LV MA MD MG MK MN MW MX MZ NA NI NO NZ OM PG PH PL PT RO RU SC SD SE SG SK SL SY TJ TM TN TR TT TZ UA UG US UZ VC VN YU ZA ZM ZW

AL Designated countries for regional patents

Kind code of ref document: A1

Designated state(s): GM KE LS MW MZ NA SD SL SZ TZ UG ZM ZW AM AZ BY KG KZ MD RU TJ TM AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HU IE IT LU MC NL PL PT RO SE SI SK TR BF BJ CF CG CI CM GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
WWE Wipo information: entry into national phase

Ref document number: 2004766522

Country of ref document: EP

WWE Wipo information: entry into national phase

Ref document number: 2006523636

Country of ref document: JP

WWP Wipo information: published in national office

Ref document number: 2004766522

Country of ref document: EP

DPEN Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed from 20040101)
WWE Wipo information: entry into national phase

Ref document number: 10569354

Country of ref document: US

Ref document number: 2007233624

Country of ref document: US

WWP Wipo information: published in national office

Ref document number: 10569354

Country of ref document: US