US 8116571 B2 Abstract An input pattern feature amount is decomposed into element vectors. For each of the feature vectors, a discriminant matrix obtained by discriminant analysis is prepared in advance. Each of the feature vectors is projected into a discriminant space defined by the discriminant matrix and the dimensions are compressed. According to the feature vector obtained, projection is performed again by the discriminant matrix to calculate the feature vector, thereby suppressing reduction of the feature amount effective for the discrimination and performing effective feature extraction.
Claims(1) 1. A pattern feature extraction method comprising:
dividing via a processor an input image into sets of a plurality of equally partitioned block images, wherein each dividing step divides said input image into said sets of said plurality of equally partitioned block images by a divisor, the divisor for each set differing from one set to another;
extracting via a processor a Fourier amplitude of each of the input image and the block images that are obtained in each dividing step for generating a multiblock Fourier amplitude vector composed of the thus extracted Fourier amplitudes of the input image and the block images to thereby extracting a feature amount of the input image;
projecting the multiblock Fourier amplitude vector using a basis matrix to obtain a projection vector; and
normalizing the projection vector to obtain a normalized vector,
wherein the basis matrix is a basis matrix specified by a transformation matrix for extracting principal components of the multiblock Fourier amplitude vector and by a discriminant matrix obtained by discriminant analysis on the principal components.
Description Conventionally, in the field of pattern recognition, the similarity between patterns such as characters or human faces has been determined by extracting feature vectors from input patterns, extracting feature vectors effective for identification from the feature vectors, and comparing the feature vectors obtained from the respective patterns. In the case of face verification, for example, pixel values of a facial image normalized with the positions of the eyes or the like are raster-scanned to transform the pixel values into a one-dimensional feature vector, and the principal component analysis is performed by using this feature vector as an input feature vector (non-patent reference 1: Moghaddam et al., “Probabilistic Visual Learning for Object Detection”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 17, No. 7, pp. 696-710, 1997) or linear discriminant analysis is performed on the principal components of the feature vector (non-patent reference 2: W. Zhao et al., “Discriminant Analysis of Principal Components for Face Recognition”, Proceedings of the IEEE Third International Conference on Automatic Face and Gesture Recognition, pp. 336-341, 1998), thereby reducing dimensions and performing personal identification or the like based on faces by using obtained feature vectors. In these methods, covariance matrices, within-class covariance matrices, and between-class covariance matrices are calculated with respect to prepared learning samples, and basis vectors are obtained as solutions to the eigenvalue problems in the covariance matrices. The features of input feature vectors are then transformed by using these basis vectors. Linear discriminant analysis will be described in more detail below. Linear discriminant analysis is a method of obtaining a transformation matrix W which maximizes the ratio of a between-class covariance matrix S
In this equation, the within-class covariance matrix Σ
If a priori probability P(ω The transformation matrix W which maximizes equation (1) can be obtained as a set of generalized eigenvectors corresponding to M large eigenvalues of equation (6) as the eigenvalue problem of a column vector w
Note that a conventional linear discriminant analysis method is disclosed in, for example, non-patent reference 5: Richard O. Duda et al., “Pattern Recognition” (supervised/translated by Morio Onoue, Shingijutu Communications, 2001, pp. 113-122). Assume that the number of dimensions of the input feature vector x is especially large. In this case, if small learning data is used, Σ As described in patent reference 1: Japanese Patent Laid-Open No. 7-296169, it is known that a high-order component with a small eigenvalue in a covariance matrix includes a large parameter estimation error, which adversely affects recognition precision. According to the above article by W. Zhao et al., the principal component analysis is performed on input feature vectors, and discriminant analysis is applied to principal components with large eigenvalues. More specifically, as shown in According to the computation scheme for feature transformation matrices described in patent reference 1: Japanese Patent Laid-Open No. 7-296169, the number of dimensions is reduced by deleting high-order eigenvalues of total covariance matrix Σ The principal component analysis using the total covariance matrix Σ Assume that the feature vector x is comprised of three elements (x=(x This phenomenon will be described with reference to the accompanying drawings. Assume that In the prior art, this is a phenomenon which cannot be avoided by the principal component analysis and the technique of deleting spaces with small eigenvalues in (total) covariance matrices. The present invention has been made in consideration of the above problems in the prior art, and has as its object to provide a feature vector transformation technique for suppressing a reduction in feature amount effective for discrimination and performing efficient feature extraction when a feature vector effective for discrimination is to be extracted from an input pattern feature vector and feature dimensions are to be compressed. A pattern feature extraction method according the present invention is characterized by comprising the step of expressing one of a pattern feature and a feature from an image by using a plurality of feature vectors x This pattern feature extraction method is characterized in that the step of performing linear transformation comprises the step of compressing the number of feature dimensions by transforming a feature vector of a pattern. In addition, the method is characterized in that the step of expressing comprises the step of dividing a pattern feature into a plurality of feature vectors x In addition, the method is characterized by further comprising the step of calculating in advance a matrix w specified by the discriminant matrices W The above pattern feature extraction method is characterized in that the step of expressing comprises the step of extracting a feature vector x This pattern feature extraction method is characterized in that the step of obtaining in advance the discriminant matrix W The method is characterized by further comprising the step of calculating in advance a matrix W specified by the discriminant matrices W The above pattern feature extraction method is characterized in that the step of expressing comprises the step of segmenting an image into a plurality of preset local regions, and expressing a feature amount as a feature vector x This pattern feature extraction method is characterized in that the step of obtaining in advance the discriminant matrix W The method is characterized by further comprising the step of calculating in advance a matrix W specified by the discriminant matrices W The above pattern feature extraction method is characterized by further comprising the step of performing a two-dimensional Fourier transform for an image, wherein the step of expressing comprises the step of extracting a real component and an imaginary component of a two-dimensional Fourier transform as a feature vector x This pattern feature extraction method is characterized in that in the step of performing linear transformation, a feature amount is extracted from an image by transforming a feature vector x This pattern feature extraction method is characterized by further comprising the step of calculating a discriminant feature of principal components of a feature vector x This pattern feature extraction method is characterized in that the step of expressing further comprises the step of segmenting an image into a plurality of regions, and in the step of extracting the feature vector x In addition, the method is characterized in that in the step of segmenting, a region is segmented into regions having different sizes in a multiple manner. In addition, the method is characterized in by further comprising the step of reducing feature dimensions by performing feature extraction by kernel discriminant analysis on an obtained two-dimensional Fourier power spectrum and extracting an effective feature amount. The method is characterized by further comprising the step of reducing feature dimensions by performing linear transformation using a discriminant matrix obtained in advance by linear discriminant analysis with respect to an obtained two-dimensional Fourier power spectrum. The method is characterized in that the step of obtaining in advance the discriminant matrix W This pattern feature extraction method is characterized in that the step of expressing further comprises the step of calculating a power spectrum of a two-dimensional Fourier transform, the step of segmenting an image into a plurality of regions and calculating a power spectrum of a two-dimensional Fourier transform for each of the regions, and the step of extracting a vector obtained combining the respective power spectra as a feature vector x A pattern feature extraction device according to the present invention is a pattern feature extraction device for compressing feature dimensions of a pattern feature by using linear transformation, characterized by comprising basis matrix storage means for storing a basis matrix specified by a discriminant matrix W A computer-readable storage medium according to the present invention is a computer-readable storage medium which records a program for causing a computer to execute pattern feature extraction to compress feature dimensions of a pattern feature by using linear transformation, the program being characterized by including a program for executing a function of expressing a pattern feature by a plurality of feature vectors x An image feature extraction method according to the present invention is characterized by comprising the step of obtaining a Fourier spectrum vector by calculating a Fourier spectrum of an input normalized image by using a predetermined mathematic expression, the step of extracting a multiblock Fourier amplitude vector from a Fourier amplitude of a partial image of the normalized image, the step of obtaining normalized vectors of a Fourier spectrum vector and the multiblock intensity vector by performing projection of feature vectors with respect to the Fourier spectrum vector and the multiblock intensity vector by using a basis matrix, the step of combining the normalized vectors to form a combined Fourier vector and obtaining a projection vector of the coupled value by using a second basis matrix, and the step of extracting a Fourier feature by quantizing the projection vector. An embodiment of the present invention will be described in detail with reference to the accompanying drawings. The pattern feature extraction device will be described in detail below. As shown in The input feature vectors x The discriminant matrix storage means As described above, discriminant matrices may be obtained by calculating a within-class covariance matrix Σ Discriminant matrices can be obtained in advance by selecting eigenvectors W When M In this case, the sizes of the matrices W The numbers of feature dimensions can be efficiently reduced by greatly reducing dimension counts M For this reason, the dimension counts M The third linear transformation means
As in one case of equation (7), the Lth-dimensional feature vector y is projected according to equation (9) by using the basis matrix W In this manner, each feature vector is divided, and linear discriminant analysis is performed on learning samples of feature vectors with small dimension counts, thereby suppressing estimation errors, which tend to occur in high-dimensional feature components, and obtaining features effective for discrimination. In the above case, the three linear transformation means are provided to perform processing concurrently and stepwisely. However, since a linear discrimination means can be basically realized by a product-sum computing unit, one linear transformation means can be commonly used by switching discriminant matrices to be read out in accordance with the input feature vector to be linearly transformed. The size of a necessary computing unit can be reduced by using one linear transformation means in this manner. As is obvious from equations (7), (8), and (9), the computation of an output feature vector z can be expressed as:
That is, linear transformations using the respective discriminant matrices can be integrated into linear transformation using one matrix. In stepwise computation processing, the number of times of product-sum computation is L×(M According to the above case, when different kinds of features, e.g., directional features and density features, are to be joined together, discriminant analysis is repeatedly performed on a feature vector having undergone discriminant analysis for each feature. However, a plurality of elements corresponding to one feature may be divided into a plurality of vectors, discriminant analysis may be performed on each element set as an input feature, and the corresponding projected vector may be further subjected to discriminant analysis. In the second embodiment, a facial image feature extraction device will be described. As shown in Techniques of extracting features from facial images include a method of positioning facial images at the eye position or the like and setting their density values as vector features, as disclosed in the above article by W. Zhao et al. In the second invention as well, pixel density values of an image are handled as an input feature, i.e., an original feature. However, an image feature has a large image size, for example, 42×54 pixels=2352 dimensions with the central positions of the left and right eyes being normalized to the coordinates (14, 23) and (29, 23). With such large feature dimensions, it is difficult to perform high-precision feature extraction by directly performing linear discriminant analysis using limited learning samples. Therefore, a deterioration in feature which is caused when the principal component analysis or the like is applied is suppressed by decomposing image feature elements, performing discriminant analysis on the decomposed features, and obtaining discriminant matrices. One of the methods of decomposing image features is to segment an image. For example, as shown in Note that letting regions have overlaps when an image is segmented makes it possible to reflect, in feature vectors, feature amounts based on the correlations between pixels in the boundary regions. Therefore, the respective regions may be sampled after being overlapped. Since the number of feature dimensions is greatly reduced to 252 as compared with the original image, a basis matrix based on discriminant analysis can be calculated with high precision by sampling several images of each of several hundred individuals, i.e., a total of several thousand facial images. If the number of feature dimensions is as large as that of the original feature (2352 dimensions), in order to obtain similar performance with features based on discriminant analysis, it is expected that facial images of several thousand individuals must be sampled. In practice, however, it is difficult to collect such a large amount of image data, and hence this technique cannot be realized. Assume that the feature in each local region is compressed to a 20-dimensional feature by a first-stage discriminant feature. In this case, the resultant output feature vectors become a feature vector of 9 regions×20 dimensions=180 dimensions. By further performing discriminant analysis on this feature vector, the number of dimensions can be efficiently reduced to about 50 dimensions. This second-stage discriminant matrix is also stored in the discriminant matrix group storage means By applying discriminant analysis locally and stepwisely, a facial feature with high identification performance can be extracted. Assume that in character recognition, for example, and are to be identified. In this case, if the principal component analysis is performed on each entire character image to extract components having large eigenvalues, the feature “{grave over ( )}” that helps to identify and tends to be lost (for this reason, similar character identification is sometimes performed by using a specific high-order feature instead of a feature of a portion with a large eigenvalue obtained by the principal component analysis). The effectiveness of segmenting an image into local regions and extracting discriminant features is similar to a phenomenon in similar character identification in character recognition. It can be thought that spatially limiting a feature that is easy to identify can ensure higher precision per unit dimension than performing discriminant analysis on principal components as a whole.In addition, the image feature decomposition means Another embodiment of the present invention will be described in detail with reference to the accompanying drawings. The facial image matching system will be described in detail below. As shown in The facial metadata creating unit When a facial image is to be registered, a facial photo or the like is input upon adjustment of the size and position of the face by using the facial image input unit The input facial image is registered in the facial image database At the time of retrieval, the facial image input unit In retrieval operation, when it is to be checked whether or not data identical to a pre-input facial image exists in the database (facial identification), the similarity between the input facial image and each data registered in the facial metadata storage unit When it is to be checked whether or not a facial image specified by an ID number or the like in advance coincides with a retrieved facial image (face verification), the facial similarity calculation unit The facial image matching system operates in the above manner. Such operation can be implemented on a computer system. For example, facial image matching can be realized by storing a metadata creation program for executing metadata creation to be described in detail next and a similarity calculation program in a memory and executing these programs using a program control processor. In addition, these programs may be recorded on a computer-readable recording medium. The operation of this facial image matching system, and more specifically, the operations of the facial metadata creating unit (1) Creation of Facial Metadata The facial metadata creating unit The region cutting means The reason why a central region of a face is cut in the above manner is that a stable feature can be extracted by cutting a range free from the influences of a hair style and the like even if the hair style changes (for example, when facial verification is used in a home robot, verification can be done even if the hair style changes before and after bathing). If a hair style and the like do not change (for example, personal identification within scenes in a video clip), since an improvement in verification performance can be expected by performing verification using images including hair styles, a large facial image including a hair style and a small facial image of a central portion of the face are cut. The facial image feature extraction means After a Fourier frequency feature is extracted with this arrangement, discriminant features of principal components are calculated for a feature vector including the real and imaginary parts of the Fourier frequency component as elements and a feature vector including a power spectrum as an element, and a discriminant feature is calculated again for a feature vector obtained by combining the above vectors, thereby calculating the feature amount of the face. Each operation will be described in more detail below. The Fourier transform means
The Fourier power calculation means calculates a Fourier power spectrum |F(u, v)| by obtaining the size of the Fourier feature F(u, v) according to equation (12).
The two-dimensional Fourier spectra F(u, v) and |F(u, v)| obtained in this manner are obtained by transforming only the images of two-dimensional real parts, the obtained Fourier frequency components become symmetrical. For this reason, these spectrum images F(u, v) and |F(u, v)| have M×N components (u=0, 1, . . . , M−1; v=0, 1, . . . , N−1), and half of the components, i.e., M×N/2 components (u=0, 1, . . . , M−1; v=0, 1, . . . , N−1) and the remaining half components are substantially equivalent. Therefore, the subsequent processing may be performed by using half components as a feature vector. Obviously, computation can be simplified by omitting computation for components which are not used as elements of a feature vector in the Fourier transform means The linear transformation means That is, a discriminant matrix φ Letting ψ It suffices if the number of dimensions to be reduced by the principal component analysis is set to about 1/10 (about 200 dimensions) of the original feature Fourier feature. Thereafter, the number of dimensions is reduced to about 70 by this discriminant matrix. This basis matrix is calculated in advance from learning samples and is used as information to be stored in the basis matrix storage means In the case of the Fourier spectrum |F(u, v)| as well, a spectrum is expressed as a one-dimensional feature vector x Calculating a principal component discriminant feature for each component of a Fourier feature in this manner makes it possible to obtain a discriminant feature y A normalization means That is, a vector y
In this manner, the normalization means is provided to normalize the feature vector y In addition, since the sizes of these vectors have already been normalized within a feature space necessary for discrimination in the process of dimension reduction, normalization robust against noise can be realized as compared with a case wherein normalization is performed in a feature space containing more deleted noise. This normalization can remove the influences of variation elements such as variation components which are proportional to the overall illumination intensity which is difficult to remove by simple linear transformation. The feature vectors y When the output feature vector z is to be quantized in five bits per element, the size of each element must be normalized in advance. For example, the size of each element is normalized in advance in accordance with the variance value of each element. That is, a standard deviation value σ In this case, normalization is the computation of multiplying each element by the reciprocal of the standard deviation. In consideration of a matrix Σ having a σ Performing normalization in this manner can perform range correction necessary for quantization. In addition, since normalization is performed by using the standard deviation, computation based on the Mahalanobis distance can be performed by only calculating a simple L2 norm in computing the norm of an inter-pattern distance at the time of collation, thereby reducing the computation amount at the time of collation. As described above, the facial image feature extraction means Note that a computer may be caused to execute the above facial metadata creation sequence by a computer program. In addition, this program may be recorded on a computer-readable recording medium. (2) Facial Similarity Calculation The operation of the facial similarity calculation unit The facial similarity calculation unit For example, a similarity is calculated by the square distance of equation (16).
Note that when a distance is used, a larger value indicates a lower similarity (the faces do not resemble each other), whereas when a cosine is used, a larger value indicates a higher similarity (the faces resemble each other). According to the above description, one facial image is registered, and a retrieval is performed by using one facial image. When, however, a plurality of images are registered for the face of one individual and a retrieval is to be performed by using one facial image, a similarity may be calculated for each of a plurality of facial metadata on the registration side. Likewise, when a plurality of images are to be registered for the face of one individual and a retrieval is to be performed by using a plurality of images, calculating a similarity by obtaining the mean or minimum value of similarity for each combination makes it possible to calculate a similarity for one facial data. This indicates that the matching system of the present invention can be applied to face verification in an image sequence by regarding the image sequence as a plurality of images. The embodiments of the present invention have been described above by referring to the accompanying drawings as need. Obviously, however, the present invention can be implemented by a computer-executable program. In addition, this program may be recorded on a computer-readable recording medium. Another embodiment of the present invention will be described in detail with reference to the accompanying drawings. The present invention is directed to an improvement in the facial metadata creating unit Since this number of dimensions is too large in general when learning data is small in amount, the principal component analysis is performed in advance to obtain in advance the basis of the principal component analysis which reduces the number of dimensions. For example, an appropriate number of dimensions is about 300. Discriminant analysis is further performed on the feature vector of this dimension count to obtain a basis which reduces the number of dimensions and corresponds to a feature axis exhibiting good discriminant performance. A basis corresponding to the principal component analysis and discriminant analysis is calculated in advance (this basis will be referred to as a PCLDA projection basis ψ). A discriminant feature z can be obtained by projecting the 5120-dimensional feature by linear computation using the projection basis ψ using this PCLDA basis. The feature amount of the face can be obtained by further performing quantization and the like for this feature. Note that the 5120-dimensional feature amount can be reduced in the number of dimensions by considering the symmetry of the Fourier power spectrum and removing and not using high-frequency components. This can realize high-speed learning, reduce the amount of data required, and realize high-speed feature extraction. Therefore, the number of dimensions is preferably reduced as needed. Segmenting a region into blocks and multiplexing Fourier spectra in this manner can sequentially obtain multiple expressions of feature amounts having translation universality and local feature amounts from a feature amount equivalent to an image feature (in the case of 1024 segmentations). A feature amount effective for identification is selected from the multiple, redundant feature expressions by discriminant analysis, thereby obtaining a compact feature amount which provides good identification performance. A Fourier power spectrum is obtained by nonlinear computation for an image, which can calculate an effective feature amount which cannot be obtained by simply applying discriminant analysis based on linear computation to the image. Although the application of linear discriminant analysis to principal components has been described above, second-stage feature extraction may be performed by using kernel discriminant analysis (discriminant analysis using a kernel technique called Kernel Fisher Discriminant Analysis, KFDA, Kernel Discriminant Analysis: KDA, or Generalized Discriminant Analysis: GDA). For a detailed description of kernel discriminant analysis, see the reference by Q. Liu et al. (non-patent reference 3: “Kernel-based Optimized Feature Vectors Selection and Discriminant Analysis for Face Recognition”, Processing of IAPR International Conference on Pattern Recognition (ICPR), Vol. II, pp. 362-365, 2002) or the reference by G. Baudat (non-patent reference 4: Generalized Discriminant Analysis Using a Kernel Approach”, Neural Computation, Vol. 12, pp. 2385-2404, 2000). By extracting a feature using kernel discriminant analysis, the effect of nonlinear feature extraction can be enhanced to allow extraction of an effective feature. In this case, however, since a large feature vector of 5120 dimensions is to be processed, a large amount of memory and a large amount of learning data are required even for the principal component analysis Referring to In this case, the principal component analysis and discriminant analysis are performed for each region by using a 1024-dimensional feature amount (512 dimensions if the number of dimensions is reduced to half in consideration of symmetry) to obtain a basis matrix ψ By performing processing for each block in this manner, the number of data and computer resources required for learning can be reduced. This makes it possible to shorten the time required for the optimization of learning. Note that high-speed computation can be realized by omitting the vector normalization processing and calculating a basis matrix for PCLDA projection and a basis matrix for LDA projection in advance. Assume that an input image f(x, y) has 32×32 pixels. In this case, as shown in In consideration of the symmetry of the Fourier power spectrum of the real image, it suffices to extract ½ of them. Alternatively, in order to avoid an increase in the size of a feature vector for discriminant analysis, a feature vector may be formed without sampling any high-frequency components for discrimination. If, for example, a feature vector is formed by sampling ¼ of spectra which correspond to low-frequency components, the number of learning samples required can be reduced or the processing time required for learning and recognition can be shortened. If the number of learning data is small, discriminant analysis may be performed after the number of feature dimensions is reduced by the principal component analysis in advance. Discriminant analysis is performed by using a feature vector x Likewise, the feature vector x Assume that the input is a facial image normalized to a size of 44×56 pixels. In this case, the above processing is applied to the 32×32 pixels of a central portion to extract a facial feature amount. In addition, facial feature amounts are also extracted from multiple segmented regions of the 44×56 pixel region of the entire face, including the entire 44×56 pixel region, four 22×28 pixel regions, and 16 11×14 pixel pixels. Another embodiment of the present invention will be described in detail with reference to the accompanying drawings. This embodiment is an embodiment of a facial feature description method using the present invention and descriptors of facial features. In this case, for a description of a facial feature named “AdvancedFaceRecognition”, elements named “FourierFeature” and “CentralFourieFeature” are provided. Each of “FourierFeature” and “CentralFourieFeature” is a 5-bit integer without a sign, representing that it can have 24-dimensional components to 63-dimensional components. Descriptors of such facial features using the present invention will be described in more detail. numofFourierFeature This field specifies the number of components of FourierFeature. The allowable range is from 24 to 63. numOfCentralFourierFeature This field specifies the number of components of CentralFourierFeature. The allowable range is from 24 to 63. FourierFeature This element represents a facial feature based on the cascaded LDA of the Fourier characteristics of a normalized face image. The normalized face image is obtained by scaling an original image into 56 lines with 46 luminance values in each line. The center positions of two eyes in the normalized face image shall be located on the 24th row and the 16th and 31st columns for the right and left eyes respectively. The FourierFeature element is derived from two feature vectors; one is a Fourier Spectrum Vector x (1) Extraction of a Fourier Spectrum Vector x (2) Extraction of a Multi-block Fourier Amplitude Vector x (3) Projections of feature vectors using PCLDA basis matrices ψ (4) Projection of a Joint Fourier Vector y (5) Quantization of the projected vector Z Step-1) Extraction of Fourier Spectrum Vector Given a normalized face image f(x, y), the Fourier spectrum F(u, v) of f(x, y) is calculated by
The dimension of x Step A multi-block Fourier Amplitude Vector is extracted from the Fourier amplitudes of partial images in the normalized face image. As the partial images, three types of images are used; (a) a holistic image, (b) quarter images, and (a) 1/16 images. (a) Holistic Image A holistic image f Quarter images are obtained by dividing the holistic image f One-sixteenth images are obtained by dividing f From these images, Fourier amplitudes |F Multi-block Fourier Amplitude Vectors is obtained by scanning low frequency regions of each amplitude |F Therefore, the Multi-block Fourier Amplitude Vector x
The dimension of x Step The Fourier Spectrum Vector x The normalized vectors y Each elements of z
The quantized elements are stored as FourierFeature. FourierFeature[0] represents the first quantized element w CentralFourierFeature This element represents a facial feature based on the cascaded LDA of the Fourier characteristics of the central part in the normalized face image. CentralFourierFeature is extracted in the similar way as FourierFeature. The central portion g(x, y) is obtained by clipping the image f(x, y) into 32×32 image starting at (7, 12) as follows:
The Fourier spectrum G(u, v) of g(x, y) is calculated by A Central Multi-block Fourier Amplitude Vectors x (a) Central Part
A Fourier amplitude |G The processing in STEP The size of CentralFourierFeature is indicated by numOfCentralFourierFeature. Facial feature description data obtained in this manner is compact in description length but exhibits high recognition performance, and hence is an expression efficient for the storage and transmission of data. Note that the present invention may be implemented by a computer-executable program. In the case of the fifth embodiment, the present invention can be implemented by describing the functions indicated by steps In addition, this program may be recorded on a computer-readable recording medium. When the example shown in According to each embodiment described above, a feature vector effective for discrimination by discriminant analysis is extracted from an input pattern feature vector for each element vector, and feature extraction is performed again for the obtained feature vector by using a discriminant matrix by discriminant analysis. This makes it possible to suppress a reduction in feature amount effective for discrimination when feature dimension reduction is performed, and to transform a feature vector for efficient feature extraction. Each embodiment described above is effective for a case wherein the number of learning samples required for discriminant analysis is limited in spite of a large pattern feature amount. That is, the number of feature dimensions can be reduced, while a loss of features effective for identification is suppressed, without necessarily using the principal component analysis. As has been described above, the image feature extraction method, the image feature extraction device, and the recording medium storing the corresponding program in the field of pattern recognition according to the present invention are suitable for the use in a feature vector transformation technique for compressing feature dimension by extracting feature vectors effective for recognition from input feature vectors. Patent Citations
Non-Patent Citations
Referenced by
Classifications
Legal Events
Rotate |