WO2001080187A1 - Processeur d'images, procede de traitement d'images et support d'enregistrement - Google Patents
Processeur d'images, procede de traitement d'images et support d'enregistrement Download PDFInfo
- Publication number
- WO2001080187A1 WO2001080187A1 PCT/JP2001/003202 JP0103202W WO0180187A1 WO 2001080187 A1 WO2001080187 A1 WO 2001080187A1 JP 0103202 W JP0103202 W JP 0103202W WO 0180187 A1 WO0180187 A1 WO 0180187A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- data
- hair
- image
- coordinate system
- hairstyle
- Prior art date
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/24—Indexing scheme for image data processing or generation, in general involving graphical user interfaces [GUIs]
Definitions
- TECHNICAL FIELD The present invention relates to an image processing apparatus, an image processing method, and a $ medium. More specifically, the present invention relates to an image processing apparatus, an image processing method, and a recording medium. TECHNICAL FIELD The present invention relates to an image processing apparatus, an image processing method, and a recording medium that are suitable for use in synthesizing a hair image represented by a hair.
- CG computer graphics
- FIG. 1A As a method of expressing human hair by computer graphics (hereinafter abbreviated as CG), as shown in FIG. 1A, an entire hairstyle is approximated using a polygon model, and the surface of the hairstyle is approximated. A method of pasting the texture of the hair and a method of rendering using an anisotropic reflection model showing a different reflection intensity depending on a light source position or a viewpoint position are used.
- a method of expressing hair by a particle system has been proposed.
- a particle system can easily represent a large number of particulate objects by giving characteristics such as the mass of the particles as parameters and generating particles according to a certain rule when handling a large number of particulate objects. How to
- the applicant of the present invention has proposed a method of approximating each individual hair as a curve as shown in FIG. 1B, for example, as Japanese Patent Application No. 10-164545.
- the method has succeeded in expressing natural hairstyle images. Kishi, Saegusa, Morishima, "Development of Hair Design System Based on Tuft Model”, Method of Designing Hairstyle by Editing Multiple Curves of Curved Approximate Hairs pp. 67-74, 1 998 ”.
- the method of expressing the head by a particle system or the method of approximating each hair by a geometrical shape can express the hairstyle of ,, and express the movement of the hair be able to.
- it is necessary to control the parameters of the particles that express the hair in order to control the hairstyle, and it is difficult to intuitively control the hairstyle. was difficult.
- the present invention has been proposed in view of the above-mentioned situation, and it is possible to change the hairstyle of each hair expressed individually by CG according to a plurality of different head shapes. The purpose is to be able to do it.
- the present invention proposed to achieve the above object is to store hairstyle data in an image processing apparatus for generating a hairstyle matching image by combining a hair image with a three-dimensional shape image of a head.
- Storage means readout means for reading hairstyle data stored in the storage means, and hair data corresponding to one hair included in the hairstyle data read out by the readout means, at a predetermined position in the three-dimensional shape image of the head
- Generating a hairstyle matching image based on the mapping result of the mapping method and the mapping result of the mapping means.
- three-dimensional curve data consisting of a plurality of control points can be used as the hair data.
- Polygon data consisting of a plurality of polygons can be used for the hair data.
- the storage means stores hairstyle data composed of a plurality of two-dimensionally arranged hair data.
- the storage means stores the hairstyle data composed of a plurality of the hair data stored in a two-dimensional array corresponding to a mapping of a three-dimensional hairstyle expressed by a cylindrical seat system to a two-dimensional coordinate system.
- the apparatus according to the present invention further includes interpolation means for interpolating the hair data at a predetermined position using the hair data in the vicinity thereof.
- the interpolation means interpolates the hair data based on the group information included in the hair data near the predetermined position.
- the mapping means includes a coordinate conversion means for converting the hair data expressed in the first coordinate system into data in a second coordinate system in which a three-dimensional shape image of the head is expressed.
- the coordinate transformation means converts the hair data expressed in the first coordinate system into data in a local coordinate system having an origin on the surface of the head, and then a three-dimensional shape image of the head is expressed. Convert to data in the second coordinate system.
- the coordinate conversion means converts the hair data expressed in the first coordinate system into data in the second coordinate system in which a three-dimensional shape image of the head is expressed, based on a random number. Execute at least one of coordinate axis rotation and origin movement.
- the present invention is an image processing method of an image processing apparatus for generating a hairstyle matching image by combining a hair image with a three-dimensional shape image of a head, and a storage step of storing hairstyle data, and a processing of a storage step
- a readout step for reading out the hairstyle data stored in step 3 and the hair data corresponding to one hair included in the hairstyle data read out in the processing of the readout step are converted into a predetermined three-dimensional shape image of the head.
- the method includes a mapping step of mapping to a position, and a generation step of generating a hairstyle matching image based on a mapping result in the processing of the matsupinda step.
- hairstyle data composed of a plurality of hair data stored in a two-dimensional array corresponding to a mapping of a three-dimensional hairstyle expressed in a cylindrical coordinate system onto a two-dimensional coordinate system is stored. Is done.
- the method of the present invention includes an interpolation step of interpolating the hair data at a predetermined position using hair data in the vicinity thereof.
- the hair data is interpolated based on the drop information included in the hair data near a predetermined position.
- the processing of the mapping step includes a coordinate conversion step of converting the hair data expressed in the first coordinate system into data in the second coordinate system in which a three-dimensional shape image of the head is expressed.
- the hair data represented in the first coordinate system is transformed into data in the local coordinate system having the origin on the surface of the head, and then a three-dimensional shape image of the head is represented. Is converted to the data of the second coordinate system.
- the present invention is a recording medium recording an image processing program for generating a hairstyle matching image by combining a hair image with a three-dimensional shape image of a head
- the program recorded on the recording medium is a hairstyle
- hairstyle data composed of a plurality of hair data arranged two-dimensionally is stored.
- the second order of the three-dimensional hairstyle expressed in the cylindrical coordinate system Hairstyle data composed of a plurality of the hair data stored in a two-dimensional array corresponding to the mapping to the original coordinate system is stored.
- the method of the present invention further includes an interpolation step of interpolating hair data at a predetermined position using hair data in the vicinity thereof.
- the hair data is interpolated based on the group information included in the hair data near the predetermined position.
- the mapping step includes a coordinate transformation step of transforming the hair data expressed in the first coordinate system into data in the second coordinate system in which a three-dimensional shape image of the head is represented. Including.
- the hair data represented in the first coordinate system is transformed into data in the local coordinate system having the origin on the surface of the head, and then the three-dimensional shape of the head is obtained. It is converted to data in the second coordinate system where the image is represented.
- the hair data represented in the first coordinate system is transformed into data in the second coordinate system representing a three-dimensional shape image of the head. Based on the random number, at least one of the coordinate axis rotation and the origin movement is executed.
- the hairstyle data is stored, the stored hairstyle data is read, and the hair data corresponding to one hair included in the read hairstyle data is stored in the head.
- the part is mapped to a predetermined position in the three-dimensional shape image, and a hairstyle matching image is generated based on the mapping result.
- the stored hairstyle data is composed of a plurality of two-dimensionally arranged hair data.
- FIG. 2 is a block diagram showing a configuration example of a hair synthesis system to which the present invention is applied.
- FIG. 3 is a flowchart illustrating the operation of the hair synthesis system.
- FIG. 4 is a diagram for explaining the operation of the hair synthesis system.
- FIG. 5 is a block diagram showing a configuration example of the hairstyle mapping section shown in FIG.
- FIG. 6 is a flowchart illustrating the processing of the hairstyle mapping processing unit.
- FIGS. 7A and 7B are diagrams for explaining hairstyle data in which hair data is two-dimensionally arranged.
- FIG. 8 is a diagram for explaining the calculation of the mating position in step S13 in FIG.
- FIG. 9 is a diagram for explaining the hair data interpolation process in step S15 shown in FIG.
- FIGS. 10A and 10B are diagrams for explaining the coordinate conversion in step S16 shown in FIG.
- FIG. 11 is a diagram showing a data structure of the hairstyle database shown in FIG.
- FIG. 12 is a diagram for explaining an interpolation process using group information included in hair data.
- FIGS. 13A and 13B are diagrams for explaining coordinate axis rotation in coordinate conversion.
- FIGS. 14A and 14B are diagrams for explaining coordinate axis movement in coordinate conversion.
- FIGS. 15A and 15B are diagrams for explaining the mapping of the arbitrarily deformed three-dimensional head model.
- FIG. 16 is a diagram showing a cylindrical coordinate system when hairstyle data is generated.
- FIG. 17 is a diagram illustrating a configuration example of a personal computer that realizes the image processing apparatus.
- BEST MODE FOR CARRYING OUT THE INVENTION A hair synthesis system to which the present invention is applied will be described with reference to FIG.
- This hair synthesis system consists of an image capture device 1 that captures the neck (upper part from the neck) of a person, generates a three-dimensional shape model of the neck based on the captured image of the neck, and creates an arbitrary hairstyle It comprises an image processing device 2 for generating a hairstyle matching image by synthesizing CG, and a display device 9 for displaying the generated hairstyle matching image.
- this hair synthesis system is installed in, for example, a hair salon or a barber shop, it can be used for purposes such as assisting customers in selecting a hairstyle.
- the imaging device 1 of the hair synthesis system is a combination of a three-dimensional shape measuring device such as a stereo camera and a range finder, and a two-dimensional image input device such as a CCD (Charge Couple Device).
- a stereo camera the three-dimensional shape measuring device of the photographing device 1 performs a predetermined image processing by photographing the neck of a person (a customer such as a beauty salon) a plurality of times from different directions. Obtain the 3D shape of the part (the neck part from which the hair part has been removed).
- the distance is measured by laser scanning the neck of the person to be the subject, and the three-dimensional shape of the face is obtained.
- the two-dimensional image input device of the imaging device 1 captures the neck of a person as a subject and obtains a texture corresponding to the three-dimensional shape of the face.
- the three-dimensional shape measuring device of the imaging device 1 acquires three-dimensional shape data of a dark portion that absorbs light of the subject, for example, a black hair head and a portion that reflects light such as metal. Can not do it. Therefore, from the imaging device 1, the three-dimensional shape data of the part of the subject's neck, such as the face excluding the hair (hereinafter referred to as the face), and the three-dimensional shape data of the face portion are associated with each other. Texture image data is output.
- the three-dimensional shape measuring device and the two-dimensional image input device constituting the photographing device 1 may be plural in order to measure the neck of a person (hereinafter, referred to as a customer) as a subject from multiple directions.
- the image processing device 2 combines the 3D shape data of the face input from the imaging device 1 with the 3D shape data of the head read from the head database 6 to construct a 3D model of the neck.
- 3D head reconstruction processing unit 3 hairstyle mapping processing unit that maps hairstyle data read from hairstyle database 5 to 3D shape model of neck4, rendering of 3D shape model of neck with hairstyle data mapped Rendering section 5, which generates hairstyle-matched image data, a head database 6 in which three-dimensional shape data of a plurality of heads are recorded in advance, and a plurality of hairstyle data in advance.
- the hairstyle data stored in the hairstyle database 7 is edited and generated using a three-dimensional graphics modeler or the like. Therefore, it is possible to generate and store hairstyle data for a hairstyle that is actually difficult to style.
- the data structure of the hairstyle database 7 will be described later with reference to FIG.
- the display device 9 is composed of an LCD (LCD), a CCD (Cathode Ray Tube), or the like, and displays a hairstyle matching image input from the image processing device 2.
- LCD LCD
- CCD Cathode Ray Tube
- step S1 the imaging apparatus 1 measures the three-dimensional shape by photographing or laser-scanning the customer's neck, which is the subject, and performs image processing on the three-dimensional shape data and texture image data of the customer's neck. Output to device 2.
- any method such as a method using a stereo camera and a method using a range finder may be used.
- the three-dimensional shape measuring device of the photographing device 1 cannot acquire the three-dimensional shape data of the hair portion.
- Shape data and texture image data (in Fig. 4, shown as measurement data in which a texture image is attached to the 3D shape data of the face) are output.
- the imaging device 1 is provided with a plurality of three-dimensional shape measurement devices and two-dimensional image input devices, a plurality of sets of three-dimensional shape data and texture image data of the face are output.
- step S2 the three-dimensional head reconstruction processing unit 2 of the image processing device 2 uses the measurement data of the face input from the imaging device 1, that is, the three-dimensional shape data and the texture image data of the customer. Construct a 3D model of the head.
- the head data selected and read from the head database 6 is combined with the measurement data of the customer's face to create a complete 3D model of the customer's head.
- the head database 6 is a tertiary of multiple heads of any shape only It is composed of the original model.
- the head data is scaled and aligned based on the size and orientation of the measurement data of the customer's face, and is combined with the customer's measurement data.
- head data and customer measurement data may cause a new polygon to be created between the release edges of both data, or the release edge part of one data will match the release edge of the other. It may be deformed as follows.
- the three-dimensional shape of the constructed customer's head is represented, for example, by a triangular polygon model.
- the constructed 3D model of the customer's head is obtained by removing the hair from the customer's real image as shown in Fig. 4.
- step S3 the operation of the user (customer or the like) for selecting the hairstyle data stored in the hairstyle database 7 is received by the controller 8, and the selected hairstyle data is read out from the hairstyle database 7 and the hairstyle is read. Supplied to the matting unit 4.
- an icon indicating the hairstyle ⁇ thumbnail or the like is displayed on the display device 9 to urge the user to make a selection.
- the selection may be made by moving a cursor or the like on the screen of the display device 9, or may be made by inputting a number or the like indicating a hairstyle.
- step S4 the hairstyle mapping processing section 4 of the image processing apparatus 2 maps the supplied hairstyle data to the customer's head 3D model supplied from the 3D head reconstruction processing section 3, Output to rendering unit 5.
- the hair data corresponding to each individual hair included in the hairstyle data is mapped on the surface of the head 3D model according to the shape of the head 3D model by the same method as the texture matting. Things. The details of the mapping process will be described later with reference to FIGS.
- step S5 the rendering unit 5 renders the three-dimensional model of the customer's head to which the hairstyle data is mapped, generates hairstyle matched image data as shown in FIG.
- Output to The rendering method is a texturing process of a polygon model for the face of the three-dimensional model of the head, and a curved model rendering process for the hair.
- Render curve model Ring processing is described in, for example, “Kobayashi, Morishima, and Harashima,“ Simulation of Filamentary Objects: Motion Model and CG ”, IEICE Technical Report, PRU90-127, pp. 15-20, 1991 The details are disclosed;
- the display device 9 displays an image corresponding to the input hairstyle matching image data.
- the controller 8 prompts the user (customer or the like) to input whether or not the hairstyle matching image displayed on the display device 9 is satisfied.
- the user can operate the controller 8 to view the hairstyle matching image while changing the viewpoint position and the illumination position.
- the rendering unit 5 executes the rendering processing of the head 3D model and the hair-shaped data again so as to correspond to the changed condition (the viewpoint position and the lighting position). And outputs it to the display device 9.
- step S5 If an input indicating that the user is not satisfied is received in step S5, the process returns to step S3, and the subsequent processing is repeated. If it is determined in step S5 that the input of the user indicating the satisfaction has been received, the process is terminated.
- the hairstyle data reading section 21 of the hairstyle mapping section 4 reads the hairstyle data selected using the controller 8 from the hairstyle database 7 and outputs the data to the mapping position calculation section 22.
- the mapping position calculation unit 22 calculates a position on the head 3D model for mapping the hair data (corresponding to one hair) constituting the hairstyle data.
- the hair data interpolation processing unit 23 obtains the hair data to be mapped by interpolating the hair data around the hairstyle data. Coordinate converter 2
- the hairstyle data reading section 2 has already been stored in the mapping position calculation section 22 of the hairstyle mapping section 4 by the processing of step S3. It is assumed that hairstyle data is supplied from 1.
- the hairstyle data supplied to the mapping position calculation unit 22 is a two-dimensional array of hair data approximated by an object having a three-dimensional shape as shown in FIG.
- one hair is approximated by a space curve such as a cubic B-spline curve.
- one hair is approximated by the polygon model.
- the addresses of the two-dimensional array correspond to the mapping coordinate values given to the head three-dimensional model whose hairstyle is mapped, as in the texture mapping process.
- mapping coordinates (u 1 ) u
- V) (0.1, 0. If 1) is given, the address in the hairstyle data array of the hair data mapped to the vertex (0.1 0.1 V ma J and ing.
- umai is the number of data in the u-direction of the hairstyle data array
- V is the number of data in the V-direction of the hairstyle data array.
- the hairstyle data is a two-dimensional array of hair data having an arbitrary three-dimensional shape, it is possible to express an arbitrary hairstyle, and it is practically difficult to cut the hairstyle. Even data can be generated. The method of generating hairstyle data will be described later with reference to FIG.
- step SI1 the mapping position calculation unit 22 initializes the vertical count value V of the hairstyle data array to 0.0, and in step S12, the horizontal force ⁇ of the hairstyle data array is initialized. Initialize the count value u to 0.0.
- step S13 the mapping position calculation unit 22 calculates a position on the head 3D model for mapping the hair data represented by the address (u, V) of the hairstyle data array.
- Figure 8 shows an arbitrary triangular polygon on the 3D head model.
- the vertices AB and C of the triangular polygon have attribute vectors A attr , B attr , and C at , respectively, which represent the attributes of the three-dimensional head model.
- the attributes consist of the object coordinate values of the 3D model of the head, the local coordinate axes at the vertices, and the mapping coordinates of the hair. If vectors forming any two sides of the triangular polygon ABC C are vector V 1 V 2, the vector V 1 V 2 belongs to each vertex. For example, it can be expressed by the following equation (1) using the sex vector.
- V l B a , tr -A a ,. R
- the coefficients a and b are the points that intersect when a line parallel to the other axis is drawn from one point P h to one axis, and an arbitrary point P h Is normalized by the length of each axis vector.
- Vertices A, B respectively a mapping coordinate values of C (u a, VJ, ( u b, v b), as the (uc, vc), equation (point P h mapping coordinate values represented by 2) Looking at the elements of, the following equation (3) holds.
- V h V a + a (V b -Va) + b (V c-V a) (3)
- Equation (3) holds for any point in the (U, V) space.
- step S14 the mapping position calculation unit 22 determines whether or not the triangular polygon of the three-dimensional model of the head to be mapped is a hair generation region. This determination is made based on a flag (positive or negative) indicating whether or not to map the hair data assigned to each triangular polygon of the three-dimensional head model. If the triangular polygon including the mapping position obtained in the processing of step S13 has a positive flag, the triangular polygon is determined to be a hair generation area, and the process proceeds to step S ⁇ 5.
- step S 1 5 hair data interpolation processing unit 2 3, using the hair data stored (u P, v P) the hair de one data to be Matsubingu represented by the hairstyle data sequence complement Ask.
- the interpolation method an average value of four neighboring points may be used, a weighted average of four neighboring points may be used, or the nearest neighbor point may be used as it is without performing interpolation.
- FIG. 9 shows a method using a weighted average value from four neighboring points.
- counter value u u P
- the hair data in each of 4 is H1 to H4.
- the hair data Hp to be mapped is interpolated from the data of the four neighboring points using the following equation (5).
- the hair data is a spatial curve such as a cubic B-spline curve as shown in Fig. 7A. If and, the hair data interpolation process can be obtained by interpolating the data of each control point. On the other hand, when the hair data is a polygon model as shown in FIG. 7B, the interpolating process of the hair data is obtained by interpolating each polygon term data.
- dm is a decimal part of m P
- dn is a decimal part of n P.
- step S16 the coordinate conversion unit 24 performs coordinate conversion of the interpolated hair data from the object coordinate system of the hair data to the object coordinate system of the three-dimensional head model.
- H be any hair data in the hairstyle data array.
- the hair data H is represented by the object coordinate system (1, J, K) of the hair data.
- the object coordinate system of the hair data (1, J, K) is arbitrary as long as it is a three-dimensional coordinate system whose origin is the starting point of the hair data H.
- the horizontal direction of the hairstyle data array is I
- the vertical direction is J
- the direction orthogonal to them is K.
- the head 3D model is represented by the object coordinate system (X, ⁇ , Z).
- the object coordinate system of the three-dimensional head model is generally a three-dimensional orthogonal coordinate system.For example, with the center of gravity of the head as the origin, the right ear direction is X, the parietal direction is Y, and the nose direction is This is a rectangular coordinate system such that
- the hair data ⁇ ⁇ expressed in the object coordinate system (1, J, ⁇ ) of the hair data is converted to the mapping position P h of the head 3D model obtained in step S13. Convert to the local coordinate system (T, S, N).
- P h (T, S, N) is arbitrary as long as a three-dimensional coordinate system to a point P h as the origin, if example embodiment, the normal direction base-vector at point P h and N, tangent obtained by the cross product of the longitudinal inside mandrel (1 0 in the Y-axis) toward the point P h from the radiation direction vector and the normal direction base-vector N of 3D head model
- An orthogonal coordinate system in which the directional vector is T and the vector orthogonal to the normal direction N and the tangential direction T is S.
- each unit vector of the local coordinate system (T, S, N) has the value of the object coordinate system (X, Y, Z) of the 3D model of the head.
- control point C i expressed by the equation (6) is expressed by the object coordinate system (X, Y, Z) of the three-dimensional head model.
- step S17 the mapping position calculation unit 22 It is determined whether the data value u has exceeded 1.0. If it is determined that the counter value u does not exceed 1.0, the process proceeds to step S18. In step S18, the mapping position calculation unit 22 increments the counter value u by the step value du: the step value du is the degree to which the hair data stored in the hair-type data array is Is a value indicating whether to perform mapping at intervals of.
- step S17 if it is determined that the counter value u has exceeded 1.0, the horizontal matching process is performed in the hairstyle data array. Means that the end of one line has been reached, so go to step S19.
- step S19 the mapping position calculation unit 22 determines whether or not the counter value V has exceeded 1.0. If it is determined that the counter value V does not exceed 1.0, the process proceeds to step S20. In step S20, the mapping position calculation unit 22 increments the counter value V by the step value dV.
- the step value d v is a value indicating at what interval the hair data stored in the hairstyle data array is vertically mapped.
- step S19 If it is determined in step S19 that the counter value V has exceeded 1.0, the mapping process for all rows has been completed. This means that the hair mapping process is completed, and the process proceeds to step S5 in FIG.
- step S14 if the mapping flag of the triangular polygon including the mapping position obtained in the process of step S13 is negative, it is determined that the triangular polygon is not a hair generation area, Steps S15 and S16 are skipped.
- the hair generation area By specifying the hair generation area using the mating flag in this manner, all the hair data constituting the hairstyle data read out from the hairstyle database 7 are not subjected to the mating, and the arbitrary area of the head 3D model can be specified. It is possible to make a mapping only to a person. Therefore, for example, a hairstyle such as a so-called mohawk haircut or a haircut cut into a star shape can be easily generated.
- the hairstyle database 7 Each hairstyle data is represented in the object coordinate system of the head hair, and it is possible to match the shape of the head of the person without any form of the human head, and to create a three-dimensional model of the head.
- the hairstyle database 7 is composed of a plurality of hairstyle data.
- the hairstyle data is a two-dimensional array, as described above, with the number of data u max in the horizontal direction and the number of data v mai in the vertical direction described at the top, and (u max x v ma x ) Hair data for the actual hair is described.
- Group information may be described at the beginning of the hair data.
- the hair data is approximated by a cubic parametric curve or the like, a plurality of control point data are described.
- the control point data is expressed in the hair object coordinate system (I, J, K).
- the hair data may be three-dimensional polygon data.
- a plurality of polygon vertex data is described as hair data.
- the group information described in the hair data will be described.
- the group information can be used in the hair data capturing process in step S15 of the hair style mapping process (FIG. 6) described above.
- each of the hair data of the point P 1 to P 4 and HI to H 4 I do. Furthermore, it is assumed that 1 is described as group information in the hair data HI, and 2 is described as group information in the hair data H2 to H4. That is, in this example, only the hair data H1 belongs to the group 1, and the other hair data H2 to H4 belong to the group 2.
- group information is given, in the hair data interpolation processing, interpolation is performed using only hair data belonging to the same group.
- the nearest neighbor is found among the four neighboring points of the point P a to be mapped first. And the group information is obtained.
- the nearest point is point P3 ; next, the number of neighboring points belonging to the same group as the nearest point among the four neighboring points P1 to P4 is obtained.
- the linear interpolation value between the two points is used for interpolation:
- H 3 2 d n-H 2-(1 — d n) H 3
- H 3 4 d n-H 4 + (1 — d n) H 3
- r (m P -m 3 2) / (m 3 4 -m 3 2) (7) where dn is the fractional part of np, and m 3 2 is the m address at point P 3 2 And m 3 4 is the m address at point P 3 4.
- Fig. 13B shows the state in which the hairstyle variation is represented by rotating the local coordinate system using random numbers in coordinate transformation. An example is shown.
- the local coordinate system (T, S, N) of the head is rotated around the axis N in accordance with the generated random number, and then the object coordinate system (X , Y, Z) I do.
- the direction in which each hair grows can be delicately changed without changing the position where the hair data is mapped.
- Figure 14B shows the state in which the hairstyle variation is expressed by changing the mapping position using random numbers in coordinate transformation, while the hairstyle variation shown in Figure 14A is not expressed. -An example is shown.
- the local coordinate system (T, S, ⁇ ) of the position of P h is the origin, then shifted to the P corresponding to the random numbers generated, the object of the head model ⁇ coordinate system (X, Y, Z).
- the position where the hair grows can be slightly changed without changing the hairstyle.
- the three-dimensional head reconstruction processing unit 3 deforms the three-dimensional head model into an arbitrary shape, and then can perform hairstyle data matching.
- FIG. 15B shows a state where the three-dimensional head model shown in FIG. 15A is not deformed, and is enlarged by 1.5 times in the Y-axis direction. Enlargement / reduction processing of the head 3D model is performed by affinity conversion processing. At this time, the same affinity transformation is performed for the local coordinate system (T, S, N) given to any polygon vertex of the 3D head model.
- FIG. Figure 16 shows a cylindrical coordinate system surrounding the head.
- the axis passing through the center of the cylinder is the ⁇ axis
- the distance from the ⁇ axis on a plane perpendicular to the ⁇ axis is r. It is a coordinate system where the angle from the direction is ⁇ .
- the hairstyle data was created using the system disclosed in “Kishi, Saegusa, Morishima,“ Development of Hair Design System Using Tuft Model ”, IEICE Technical Report, MVE97-102, pp.67-74, 1998”. By using it, it can be performed relatively easily.
- hairstyle data is created by editing the hair data generated on a given head model.
- the hair data output from this system is expressed in the object coordinate system of a given head model.
- the hair database 7 of the present invention records hair type data stored in a two-dimensional array by mapping this hair data into a space represented by (y, ⁇ ) in a cylindrical coordinate system.
- the following equation (8) is used to convert the (y, ⁇ ) space into a (u, V) space of a two-dimensional array as shown in FIG.
- y m is the minimum value of the Y axis of the cylindrical coordinate system on the head surface, and is the maximum value of the Y axis of the cylindrical coordinate system on the head surface.
- the transformation from the (y, ⁇ ) space to the (u, v) space may be non-linear. For example, if the transformation is made such that the step in the V direction becomes smaller as it approaches the top of the head, It is better to avoid coarse sampling. Hairstyle data composed of the hair data converted into the (u, v) space as described above is recorded in the hairstyle database 7.
- the present invention can also be applied to a case where a thread-like object other than hair is represented by CG.
- FIG. 17 illustrates a configuration example of a personal computer that realizes the image processing device 2.
- This personal computer has a built-in CPU (Central Procedure Unit) 41.
- An input / output interface 45 is connected to the CPU 41 via a bus 44.
- CPU Central Procedure Unit
- the input / output interface 45 includes an input unit 46 composed of an input device such as a keyboard and a mouse corresponding to the controller 8, an image output unit 47 that outputs a hairstyle matching image as a processing result to the display device 9, A storage unit 48 such as a hard disk drive for storing programs and various data, a communication unit 49 for communicating data via the Internet, and a recording medium such as a magnetic disk 51 to a semiconductor memory 54.
- Drive 50 is connected to read and write data.
- Knob 4 4 ⁇ (ROM (Read Only Memory) 4 2 3 ⁇ 4 1 RAM (Random Access Memory) 4 3 is connected.
- An image processing program for causing this personal computer to execute the operation as the image processing device 2 is supplied to the personal computer in a state where it is stored in the magnetic disks 51 to the semiconductor memory 54, and is read out by the drive 50. Installed on the hard disk drive built into the storage unit.
- the image processing program installed in the storage unit 48 is transferred from the storage unit 48 to the RAM 43 by a command from the CPU 41 corresponding to a command from the user input to the input unit 46. And executed.
- the processing includes, in addition to the processing performed in chronological order according to the described order, not only the processing performed in chronological order but also processing performed in parallel or individually.
- a system refers to an entire device including a plurality of devices.
- INDUSTRIAL APPLICABILITY stores hairstyle data, reads out the stored hairstyle data, and converts the hair data corresponding to one hair included in the hairstyle data into the three-dimensional shape of the head. Since the image is mapped to a predetermined position in the image, it is possible to change the hairstyle of the hair expressed by the CG one by one according to a plurality of different head shapes.
- the present invention reads out the stored hairstyle data, maps the hair data corresponding to one hair included in the read hairstyle data to a predetermined position of the three-dimensional shape image of the head, and obtains a mapping result. Since the hairstyle matching image is generated based on the CG, it is possible to change the hairstyle of the hair expressed for each hair by CG in accordance with a plurality of different head shapes.
Description
Claims
Priority Applications (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CA002376647A CA2376647A1 (en) | 2000-04-13 | 2001-04-13 | Image processor and image processing method, and recorded medium |
US10/018,033 US6842172B2 (en) | 2000-04-13 | 2001-04-13 | Image processor and image processing method, and recorded medium |
EP01921836A EP1246132A4 (en) | 2000-04-13 | 2001-04-13 | IMAGE PROCESSOR, IMAGE PROCESSING METHOD AND DATA CARRIER |
Applications Claiming Priority (6)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
JP2000112074 | 2000-04-13 | ||
JP2000112072 | 2000-04-13 | ||
JP2000-112072 | 2000-04-13 | ||
JP2001-77564 | 2000-04-13 | ||
JP2000-112074 | 2000-04-13 | ||
JP2001077564A JP2001357415A (ja) | 2000-04-13 | 2001-03-19 | 画像処理装置および方法、記録媒体、並びにプログラム |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2001080187A1 true WO2001080187A1 (fr) | 2001-10-25 |
Family
ID=27343081
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/JP2001/003202 WO2001080187A1 (fr) | 2000-04-13 | 2001-04-13 | Processeur d'images, procede de traitement d'images et support d'enregistrement |
Country Status (5)
Country | Link |
---|---|
US (1) | US6842172B2 (ja) |
EP (1) | EP1246132A4 (ja) |
JP (1) | JP2001357415A (ja) |
CA (1) | CA2376647A1 (ja) |
WO (1) | WO2001080187A1 (ja) |
Families Citing this family (36)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP3542958B2 (ja) * | 2000-07-03 | 2004-07-14 | ファイルド株式会社 | ヘアーデザインシステム及びその利用 |
FR2848660B1 (fr) * | 2002-12-13 | 2005-02-18 | Thales Sa | Procede de determination de la comptabilite biomecanique d'un equipement de tete |
US7330195B2 (en) * | 2002-12-18 | 2008-02-12 | Hewlett-Packard Development Company, L.P. | Graphic pieces for a border image |
NZ530738A (en) * | 2004-01-21 | 2006-11-30 | Stellure Ltd | Methods and systems for compositing images |
JP2005274928A (ja) * | 2004-03-24 | 2005-10-06 | Wella Ag | ヘアカラーリング用カラーシミュレーションシステム |
JP2005321986A (ja) * | 2004-05-07 | 2005-11-17 | Pioneer Electronic Corp | ヘアスタイル提案システム、ヘアスタイル提案方法、及びコンピュータプログラム |
US20110028212A1 (en) * | 2004-07-01 | 2011-02-03 | David Krien | Computerized Imaging of Sporting Trophies and Method of Providing a Replica |
JP4449723B2 (ja) * | 2004-12-08 | 2010-04-14 | ソニー株式会社 | 画像処理装置、画像処理方法、およびプログラム |
US20060177122A1 (en) * | 2005-02-07 | 2006-08-10 | Sony Computer Entertainment Inc. | Method and apparatus for particle manipulation using graphics processing |
US7418371B2 (en) * | 2005-03-30 | 2008-08-26 | Seoul National University Industry Foundation | Method and system for graphical hairstyle generation using statistical wisp model and pseudophysical approaches |
US7483763B2 (en) * | 2005-11-17 | 2009-01-27 | Centertrak, Llc | System and method for the digital specification of head shape data for use in developing custom hair pieces |
JP2007286669A (ja) | 2006-04-12 | 2007-11-01 | Sony Corp | 画像処理装置および方法、並びにプログラム |
WO2007148306A2 (en) | 2006-06-22 | 2007-12-27 | Centre National De La Recherche Scientifique | A method and a system for generating a synthesized image of at least a portion of a head of hair |
JP4986279B2 (ja) * | 2006-09-08 | 2012-07-25 | 任天堂株式会社 | ゲームプログラムおよびゲーム装置 |
EP2102785B1 (en) * | 2006-09-19 | 2016-01-27 | Sicpa Holding Sa | Apparatus and method for secure detection of an item and a method of securing access to information associated with the item |
US20080136811A1 (en) * | 2006-12-06 | 2008-06-12 | Oluwaseun Adedayo | Image displaying system |
US7714912B2 (en) * | 2007-01-24 | 2010-05-11 | International Business Machines Corporation | Intelligent mirror |
JP2009129337A (ja) * | 2007-11-27 | 2009-06-11 | Hitachi Ltd | 三次元類似形状検索装置 |
JP5808171B2 (ja) * | 2010-07-16 | 2015-11-10 | 株式会社 資生堂 | 目元画像シミュレーション装置、目元画像生成方法、及び目元画像生成プログラム |
TW201212852A (en) * | 2010-09-21 | 2012-04-01 | Zong Jing Investment Inc | Facial cosmetic machine |
TWI462030B (zh) * | 2011-06-16 | 2014-11-21 | jian wen Peng | 智慧型臉形分析與髮型搭配之系統及方法 |
JP5578186B2 (ja) * | 2012-02-16 | 2014-08-27 | カシオ計算機株式会社 | キャラクター画像の作成方法、画像処理装置及び画像処理プログラム並びに画像変換ネットワークシステム |
TWI463955B (zh) * | 2012-02-20 | 2014-12-11 | Zong Jing Investment Inc | Eye makeup device |
EP2665032A1 (en) * | 2012-05-14 | 2013-11-20 | Thomson Licensing | Methods and devices for 3d object protection using surface subdivision |
KR101521358B1 (ko) * | 2013-12-26 | 2015-05-19 | 세종대학교산학협력단 | 다중 모델 정합 방법 및 그 장치 |
JP6622421B2 (ja) * | 2016-03-01 | 2019-12-18 | コーニンクレッカ フィリップス エヌ ヴェKoninklijke Philips N.V. | 自動化されたヘアスタイル処理のためのシステム及び方法並びに毛切断デバイス |
CN106096128A (zh) * | 2016-06-08 | 2016-11-09 | 吴雨欢 | 一种发型匹配的方法及系统 |
MX2019005576A (es) | 2017-05-12 | 2019-07-04 | Procter & Gamble | Composiciones con agentes para la salud del cuero cabelludo con deposito incrementado. |
CN108389077B (zh) * | 2018-02-11 | 2022-04-05 | Oppo广东移动通信有限公司 | 电子装置、信息推荐方法及相关产品 |
MX2020012226A (es) | 2018-05-15 | 2021-01-29 | Procter & Gamble | Composiciones antioxidantes sinergicas. |
WO2019222340A1 (en) | 2018-05-17 | 2019-11-21 | The Procter & Gamble Company | Systems and methods for hair coverage analysis |
US11172873B2 (en) | 2018-05-17 | 2021-11-16 | The Procter & Gamble Company | Systems and methods for hair analysis |
US11384357B2 (en) | 2018-06-29 | 2022-07-12 | The Procter And Gamble Company | Aptamers for personal care applications |
JP2022511961A (ja) | 2018-12-20 | 2022-02-01 | ザ プロクター アンド ギャンブル カンパニー | 改善された安定性を有する頭皮ケア組成物 |
EP3956451A1 (en) | 2019-04-16 | 2022-02-23 | The Procter & Gamble Company | Aptamers for odor control applications |
CN111510769B (zh) * | 2020-05-21 | 2022-07-26 | 广州方硅信息技术有限公司 | 视频图像处理方法、装置及电子设备 |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5267154A (en) * | 1990-11-28 | 1993-11-30 | Hitachi, Ltd. | Biological image formation aiding system and biological image forming method |
US5404426A (en) * | 1991-05-21 | 1995-04-04 | Hitachi, Ltd. | Method of displaying hair style and apparatus for the same |
JP2000003455A (ja) * | 1998-06-12 | 2000-01-07 | Sony Corp | 情報処理装置および方法、並びに提供媒体 |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US4823285A (en) * | 1985-11-12 | 1989-04-18 | Blancato Vito L | Method for displaying hairstyles |
US5060171A (en) * | 1989-07-27 | 1991-10-22 | Clearpoint Research Corporation | A system and method for superimposing images |
JP2774627B2 (ja) * | 1989-12-28 | 1998-07-09 | 株式会社日立製作所 | 画像表示方法及びその装置 |
JP3084098B2 (ja) | 1991-08-30 | 2000-09-04 | 株式会社日立製作所 | コンピュータアニメーションの対話処理方式及び装置 |
JP2603445B2 (ja) * | 1994-11-10 | 1997-04-23 | インターナショナル・ビジネス・マシーンズ・コーポレイション | 髪画像適合方法及びコンピュータ・システム |
JPH0998834A (ja) * | 1995-08-03 | 1997-04-15 | Yasuda Seiko Kk | 髪形の選択システム |
US5777619A (en) * | 1996-01-02 | 1998-07-07 | Silicon Graphics, Inc. | Method for simulating hair using particle emissions |
JP3838282B2 (ja) * | 1996-06-20 | 2006-10-25 | ブラザー工業株式会社 | 絵作成装置 |
NL1007397C2 (nl) * | 1997-10-30 | 1999-05-12 | V O F Headscanning | Werkwijze en inrichting voor het met een gewijzigd uiterlijk weergeven van tenminste een deel van het menselijk lichaam. |
-
2001
- 2001-03-19 JP JP2001077564A patent/JP2001357415A/ja active Pending
- 2001-04-13 CA CA002376647A patent/CA2376647A1/en not_active Abandoned
- 2001-04-13 EP EP01921836A patent/EP1246132A4/en not_active Withdrawn
- 2001-04-13 US US10/018,033 patent/US6842172B2/en not_active Expired - Fee Related
- 2001-04-13 WO PCT/JP2001/003202 patent/WO2001080187A1/ja not_active Application Discontinuation
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US5267154A (en) * | 1990-11-28 | 1993-11-30 | Hitachi, Ltd. | Biological image formation aiding system and biological image forming method |
US5404426A (en) * | 1991-05-21 | 1995-04-04 | Hitachi, Ltd. | Method of displaying hair style and apparatus for the same |
JP2000003455A (ja) * | 1998-06-12 | 2000-01-07 | Sony Corp | 情報処理装置および方法、並びに提供媒体 |
Non-Patent Citations (2)
Title |
---|
LIEU-HEN CHEN ET AL.: "A system of 3D hair style synthesis based on the wisp model", THE VISUAL COMPUTER, SPRINGER-VERLAG, vol. 15, no. 4, 1999, pages 159 - 170, XP001035250 * |
See also references of EP1246132A4 * |
Also Published As
Publication number | Publication date |
---|---|
EP1246132A1 (en) | 2002-10-02 |
EP1246132A4 (en) | 2007-02-07 |
JP2001357415A (ja) | 2001-12-26 |
US6842172B2 (en) | 2005-01-11 |
CA2376647A1 (en) | 2001-10-25 |
US20020150287A1 (en) | 2002-10-17 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2001080187A1 (fr) | Processeur d'images, procede de traitement d'images et support d'enregistrement | |
US6256038B1 (en) | Parameterized surface fitting technique having independent control of fitting and parameterization | |
US6271856B1 (en) | Creating and modifying parameterizations of surfaces | |
US8787698B2 (en) | Methods and apparatus for directional texture generation using image warping | |
Chai et al. | Dynamic hair manipulation in images and videos | |
TWI298147B (en) | System and method for representing a general two dimensional spatial transformation | |
TWI387936B (zh) | A video conversion device, a recorded recording medium, a semiconductor integrated circuit, a fish-eye monitoring system, and an image conversion method | |
JP2013038775A (ja) | 高速反射屈折式ライトフィールドレンダリングのための光線画像モデル化 | |
JP4886645B2 (ja) | 仮想顔モデル変形装置及び仮想顔モデル変形プログラム | |
CN109801380A (zh) | 一种虚拟试衣的方法、装置、存储介质及计算机设备 | |
JPH0771936A (ja) | 画像処理装置及びその方法 | |
JP2018503470A (ja) | 統計的技術を用いて形成されたデジタル歯冠モデルに表面詳細を追加するシステム及び方法 | |
US6724383B1 (en) | System and computer-implemented method for modeling the three-dimensional shape of an object by shading of a two-dimensional image of the object | |
JP2002083318A (ja) | 画像処理装置および方法、並びに記録媒体 | |
JP2002324249A (ja) | 画像表示システム及び方法 | |
Baxter et al. | N‐way morphing for 2D animation | |
JP4229398B2 (ja) | 3次元モデリング・プログラム、3次元モデリング制御プログラム、3次元モデリング・データ伝送プログラム、記録媒体および3次元モデリング方法 | |
Laffont et al. | Interactive content-aware zooming | |
Xiao et al. | A unified method for appearance and geometry completion of point set surfaces | |
JP2009009296A (ja) | 顔画像変形方法とこの方法を実施する顔画像変形装置 | |
JPH06176107A (ja) | 距離画像処理方法及び装置 | |
CN115082640A (zh) | 基于单张图像的3d人脸模型纹理重建方法及设备 | |
JP3341549B2 (ja) | 3次元形状データ処理装置 | |
JP2001297333A (ja) | 画像処理装置および方法、並びに記録媒体 | |
Shen et al. | Completion-based texture design using deformation |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
AK | Designated states |
Kind code of ref document: A1 Designated state(s): CA US |
|
AL | Designated countries for regional patents |
Kind code of ref document: A1 Designated state(s): AT BE CH CY DE DK ES FI FR GB GR IE IT LU MC NL PT SE TR |
|
ENP | Entry into the national phase |
Ref document number: 2376647 Country of ref document: CA Ref country code: CA Ref document number: 2376647 Kind code of ref document: A Format of ref document f/p: F |
|
WWE | Wipo information: entry into national phase |
Ref document number: 2001921836 Country of ref document: EP |
|
121 | Ep: the epo has been informed by wipo that ep was designated in this application | ||
WWE | Wipo information: entry into national phase |
Ref document number: 10018033 Country of ref document: US |
|
WWP | Wipo information: published in national office |
Ref document number: 2001921836 Country of ref document: EP |
|
WWW | Wipo information: withdrawn in national office |
Ref document number: 2001921836 Country of ref document: EP |