CA2201107C - Arbitrary-geometry laser surface scanner - Google Patents
Arbitrary-geometry laser surface scanner Download PDFInfo
- Publication number
- CA2201107C CA2201107C CA002201107A CA2201107A CA2201107C CA 2201107 C CA2201107 C CA 2201107C CA 002201107 A CA002201107 A CA 002201107A CA 2201107 A CA2201107 A CA 2201107A CA 2201107 C CA2201107 C CA 2201107C
- Authority
- CA
- Canada
- Prior art keywords
- camera
- illumination means
- optical surface
- orientations
- fan beam
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Expired - Fee Related
Links
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01B—MEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
- G01B11/00—Measuring arrangements characterised by the use of optical techniques
- G01B11/24—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
- G01B11/25—Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures by projecting a pattern, e.g. one or more lines, moiré fringes on the object
- G01B11/2518—Projection by scanning of the object
- G01B11/2522—Projection by scanning of the object the position of the object changing and being recorded
Abstract
An optical non-contact surface scanning method and system is described. The three-dimensional coordinates of an illuminated profile or spot on the object are recorded and the surface is reconstructed using spatial location and orientation information derived from a spatial location system associated with the object, illumination means, and image recording means. The illumination and imaging means may be hand-held in a combined unit or in separate units.
The scanning system allows complete free and arbitrary movement of the object and all components of the scanning during scanning this allowing scanning of substantially all of the surface which may be visible at various locations and orientations around the object. The optical non-contact scanning system may find particular application in medical imaging of the human body or parts thereof. It may further find application in industrial processes whereby the passage of arbitrarily shaped objects are monitored through a production line or similar situations.
The scanning system allows complete free and arbitrary movement of the object and all components of the scanning during scanning this allowing scanning of substantially all of the surface which may be visible at various locations and orientations around the object. The optical non-contact scanning system may find particular application in medical imaging of the human body or parts thereof. It may further find application in industrial processes whereby the passage of arbitrarily shaped objects are monitored through a production line or similar situations.
Description
Wo 96/10205 PCTINZ95/00097 ARBITRARY-GEOI\~ET~RY LAS~R SURFACE SCANNER
Technical Field The invention relates to a non-contact mensuration system for sc~nning a three--limencional object in order to obtain data from which the shape of the object may be derived.
More particularly, although not exclusively, the present invention relates to a method and apparatus which does not require a fixed or predete~ illed geometrical relationship between its components per se as well as the object being sc~nn~l Any component(s) of the sc~nning system and/or the object may be moved albilldlily in space and therefore may be hand-held. Such a system is particularly suitable for obtaining measurements of irregular objects such as a human body or the like.
Background To The Invention Conventional laser surface scanners generally incorporate a rigid laser/
camera assembly. This assembly is usually fixed in space and the object being scanned is rotated or tr~n~l~te~l on a m~h~nic~l platform. AlL~llldlively7 the object is fixed and the laser/camera assembly is moved mech~nic~lly around or along it. Usually the movement takes the form of a rotation about one axis (contained within the object) or translation in one direction (along the object). In either case, the conventional sc~nnin~ process relies upon a predetermined knowledge of the geometrical relationships between the components of the sc~nning appaldlus.
Examples of such fixed-axis laser scanners include WO 94/15173, US
4,705,401, GB 2,240,623 A and US 5,193,120.
W O96/10205 PCT~NZ95/00097 --2 --Unless the object being sc~nnPd is a simple surface where every point on its surface is able to be 'seen' by the sc~nning apparatus in the course of a complete scan, a fixed-axis scarmer is unable to measure the complete surface ofan object in one scan orientation. This le~lcsellL~ a ~ignifir~nt limitation when sc~nning complex objects (for example, an entire human head). Features such as overhangs and indentations are unlikely to be visible in situations where a sc~nning device follows a fixed predetermined path.
Further, a fixed scan pattern or geometry cannot take into account variations in object detail or areas on the object' s surface which might be obscured by hl~ ellillg features protruding from the surface of the object.
Schulz ~WO 92/07233) provides, to some extent, for limited freedom of movement of a laser/camera assembly by tracking the position and orientation of a plurality of light sources fixed to the sc~nning assembly. The location of thelight sources (and therefore the sc~nning assembly) in three--limen~ional space is ~ termin~l by means of fixed photoelectronic sensors arranged in a pre~etermin~l static relationship to the fixed object. The light sources are time multiplexed with the sensors and the spatial location is derived from the three locations of the light sources. Therefore all three light sources (fixed to the sc~nning assembly) must be visible to all fixed photoelectronic sensors. There will generally be situations where at least one light source is obscured by the object itself, the operator or by tilting the assembly. Thus the Schulz scanner would appear to offer advantages over a fixed geometry scanner in only certain applications.
W O 96/10205 PCT~NZ95/00097 A further signifit~nt limitation inherent in the Schulz sc~nning system is that the object must remain stationary with respect to the reference frame defined by the photoelectronic sensors during the scan. This can present difficulties, particularly in medical applications, due to the potential for movement of the part of the subject being sc~nn~ . Accordingly, it would be desirable to be able to scan an object which may be arbitrarily located and oriented in space where suchalbiLla,"less is understood to include an object which may be moving, or to at least provide the public with a useful choice.
The present invention aLLcl~ s to overcome the limitations and disadvantages inherent in the prior art by providing a means and method of sc~nning an albiLlalily located and oriented object with the scanner components being arbitrarily located and oriented in space.
The present invention also provides a sc~nning system which may operate in ambient light and has reduced susceptibility to optical hlLclre,cllce in the process of im~ging the scanned portion of the object.
The present invention also provides for a non-optical technique for de~c- ,.,i"i"g the spatial locations and orientations of the components of the sc~nning system and object.
Disclosure Of The Invention In one aspect the invention provides an optical surface sc~nn~r including:
an illnmin~tion means adapted to emit a fan beam of light whereby the intersection between the fan beam and an object being scanned produces a profile having three-dimensional coordinates which lie in the plane of the fan beam;
a camera which images the profile in two dimensions;
a spatial location system adapted to obtain the relative positions and orientations between the object, illumination means and camera, wherein the three-~lim~n~ional coordinates of the profile are derived from the two-dimensional coordinates of the image and the relative positions and orientations of the illumin~rion means, camera and object.
In a further aspect the invention provides for an optical surface scanner inrlnt1ing:
an illnmin~tion means adapted to emit a pencil beam of light whereby the intersection between the pencil beam and an object being scanned produces a spot having three-dimensional coordinates which lie on the axis of the pencil beam;
a camera which images the spot in one or two dimensions;
a spatial location system adapted to obtain the relative positions and orientations between the object, illumination means and camera, wherein the three-~limrn~ional coordinates of the spot are derived from the one or two-dimensional coordinates of the image and the relative positions and orientations of the illllmin~tion means, camera and object.
In a further aspect the optical surface scanner inclucles:
one or more illnmin~tion means each adapted to emit a fan beam or a pencil beam of light whereby the intersection between each fan beam or pencil beam of light and the object being scanned produces a profile or spot having three-climen~ional coordinates which lie in the plane of the fan beam producing that profile or lie on the axis of the pencil producing that spot;
W O96/10205 PCT~NZ95100097 a plurality of cameras which image the profiles or spots;
means for distinguishing between the sources of illnmin~tion, a spatial location system adapted to obtain the relative positions and orientations between the object, illl-min~tion means and cameras;
wherein the three-~lim~n~ional coordinates of each profile or spot are derived from the two-dimensional or one/two dimenisional coordinates respectively of the corresponding image and the relative positions and orientations of the c~mer~, illumin~tion means and object.
Preferably the means for distinguishing between the sources of illl"~ ion uses time division multiplexing of the sources of illnmin~tion or dirr~ L wavelength sources.
Preferably the camera and illllmin~tion means are held in fixed relation.
Preferably the illllmin~tion means and camera(s) are mounted in a fixed relationship on a hand-held assembly.
Preferably the hand-held assembly includes two cameras and one illllmin~tion means wherein the illllmin~tion means is positioned between the cameras and the cameras and illll",il,~ion means are oriented so that the cameras image a profile or spot produced by the intersection of the fan beam or pencil beam respectively and the object being sc~nn~d.
Preferably the location and orientation of the hand-held assembly is deLe~ ed by means of the spatial location system.
-Preferably the spatial location system is adapted so as not to depend on m~int~ining line of sight between itS constituent components.
Preferably the spatial location system comprises a plurality of tr~ncmitrP,rs and a plurality of receivers.
Preferably the spatial location system uses electromagnetic fields.
Preferably the illumin~tion means is a laser.
Preferably the fan beam is produced by means of a cylindrical lens or similar means.
Preferably the optical surface scanner includes at least one optical band-pass filter per camera.
Preferably the band-pass filter comprises illLclrelcnce filters adapted to balance the contrast of the profile as imaged by the camera.
Preferably the one or more optical band-pass filters are located immediately in front of the camera.
The present invention further provides for a mensuration method for three-rlimencional objects comprising:
im~ging the intersection of a fan beam or pencil beam of light and an object being scanned;
mP~cll~ring the relative orientations and locations of a camera, illllmin~tion means and object;
calc~ ting three-dimensional coordinates of the intersection of the fan beam or pencil beam of light and the object being scanned;
repeating the above steps for a variety of orientations and/or locations of caInera and/or illnmin~tion means and/or object.
;
Preferably the relative position and orientation of the object, illnmin~tion means and camera are obtained by means of the spatial location system for each recorded image of the intersection of the fan beam or pencil beam and the object.
Further objects and advantages of the invention will become ~palcllL
from the following description which is given by way of example and with reference to the drawings in which:
Brief Deseli~Lion Of The D~wh~
Figure 1: illustrates a srh~m~tic diagram of an embodiment of an optical surface scanner incorporating one ill--min~tion means and one camera.
Figure 2: illustrates a scht?m~tic diagram of an embodiment of an optical surface scanner incorporating a single illllmin~rion means and two cameras embodied in a single sç~nning head.
Figure 3: illustrates a perspective view of the system in Figure 1 when used to scan a spherical object with a fan beam.
Referring to Figure 1, 100 is the object to be scanned. The major components of the sc~nning system are as follows. An illumination assembly 110 incorporates a low-power laser 101 which is directed through a cylindrical lens W O96/10205 PCTANZ9~/00097 102 thereby generating a fan beam. Other methods of producing a fan beam are envisged. For example, a point source with a slit aperture and collimator may produce the desired fan beam. An im~ging assembly 109 incorporates a two-dimensional (area array) electronic camera 103 which has an optical band-pass inte.~~ ce filter 104 covering its lens.
The spatial location system comprises a spatial location tr~n~mitter 105 and spatial location receivers 106, 107 and 108.
Spatial location electronics unit 111 and laser line extraction electronics 112 are interfaced to controlling electronics and power supply 113. 114 corresponds to a digital culll~u~
It is considered that an innovative feature enabling the construction of the truly arbitrary-geometry scanner described herein, resides in the use of spatiallocation devices 105, 106, 107 and 108 to obtain the relative position and orientalion (6 degrees of freedom) of the object 100, laser(s) 101 and camera(s)103. The spatial location system may correspond to an electronic device such as a Polhemus 'Fastrak' based on alternating electric fields or an Ascension 'Bird' based on pulsed static electric fields.
Referring to Figure 1, the dashed boxes 109 and 110 in~ te the components which are combined into the portable hand-held assemblies each associated with a spatial location receiver 107 and 108. In this case the im~ging assembly 109 consists of im~ging components (camera 103 and filter 104) and spatial location receiver 108. The illumin~tion assembly 110 consists of illllmin~tion components (laser 101 and cylindrical lens 102) and spatial location receiver 107.
WO 96/10205 PCT/~Z9~l00097 A three-dimensional representation of the system of Figure 1 is shown in - Figure 3. A cylindrical lens 302 spreads the beam emitted by the laser 301 into a fan beam 303 so that the intersection between the fan beam and the object being sc~nnt~l 300 produces a profile 304. The camera includes a lens 307 a CCD array 306 and a filter 308 in front of the camera. The image 305 of the profile 304 isobtained by the camera when viewing the object at an offset angle. The two-~lim~n.~ional coordinates of the image of the profile are recorded.
The object 300, laser 301 and camera (306 and 307) are associated with spatial location receivers 310, 311 and 312 respectively thereby allowing their orientation and location in space to be measured in real time. Thus the laser and camera may be moved albiLldlily in space so as to record desired profiles thereby building a reconstructed surface. Alternatively, the object may be movedso as to record the desired profiles.
In use, profiles 304 are recorded by directing the laser fan beam at the object while .simlllt~n~ously positioning the camera so that the profile is visible to it. The utility of the present system is illustrated by situations where features of the object might obscure either the illnmin~tion or the camera's field of view when using a fixed illnmin~ti~n/im~ging system. If such a geometry is encountered, the illllmin~tion and/or camera and/or object may be moved into an orientation or position where the profile is visible to the camera.
The spatial location system 310, 311, 312 and 313 provides the required spatial location and orientation information for the object, camera and laser whereupon the three-~iim~n~ional coordinates of the profile are (let~rmin~l fromthe t~,vo-~lim~on~ional pixel coordinates in the camera CCD array 306.
The scanner may include an hltelr~lcllce filter 308 which allows sc~nning under ambient light conditions. A further purpose of the interference filter is to balance the profile contrast as seen by the camera as discussed below.
IllLelrelcnce filters present a lower wavelength pass-band to light from off-axis and this effect is exploited. The intensity of the profile typically tiimini.~h~s with tli~t~n~e from its centre. Therefore, if a filter is chosen with a pass-band centred above the laser wavelength, the distal ends of the laser profile are subject to less attenuation by the filter. This balances the laser profile contrast as seen by the camera. Such an application of an inL~lrelcllce filter is, to the applicants' knowledge, novel.
Referring to Figure 1 the output from each two-dimensional camera 103 (such as a CCD array) is processed a~lo~.liately by line extraction electronics 112 to produce a list of pixels (two-dimensional camera coordinates) at which the intensity exceeds some threshold (constant or adaptive).
The controlling electronics unit 113 performs such tasks as synchronising the cameras, lasers and spatial location receivers, monitoring switches, supplying power, preproce~ing and interfacing this data slream to the digital input/outputport of the host digital cOlll~ul~l 114. The digital co~llL,ul~l is programmed to reconstruct the three-dimensional coordinates of the object's surface. Such a reconstruction may employ LLdllsro-lllation methods known in the art which convert each two-~limen~ional point in the profile, viewed by the camera, to a three-dimensional point in the camera frame of reference and then to a three-~limen~ional point in the object frame of reference.
wo 96/10205 PCTINZ95100097 The reconstruction process~ must der~ in~ the coordinates of a three-dimensional point, in the object's fraIne of reference, corresponding to each pixel identified by the laser line extraction electronics.
This can be accomplished in a number of ways. For example, if a simple linear model of the camera is assumed, then a ray can be traced from the pixel location through the centre of the lens whereby the intersection of the ray withthe plane of the laser will determine the three-t1imen.cional coordinate of the i11umin~ted spot on the object, the camera coordinates and the laser plane having been transformed into the object's coordinate system using the spatial location information.
~ 1tern~tively, a 'look-up table' can be used to transform from pixel coordinates to three-~limen.cional coordinates in the camera frame of reference,and then these coordinates can be transformed into the objects coordinate systemusing the spatial location hlrollllation.
The look-up table can be produced by an empirical calibration process.
Therefore allowances may be made for non-linear effects such as lens distortion as well as providing for a faster means of data conversion in the co,lll,uL~l.
It will be appreciated that the location of the object being sc~nn~l is constantly determined by means of the spatial location system. Therefore, so long as the receiver associated with the object remains in fixed relation to the object, the object is free to move during the sc~nnin~ process.
It is to be appreciated that both the number of cameras and lasers and/or the geometry of the system can be varied if desired. If, for example, any two or Wo 96/10205 PC~r/NZss/00097 more components retain a fixed relative spatial relationship throughout the scan, then the position and orientation of the combined assembly may be recorded by a single spatial location receiver (the relative locations and orientations of theassembly components being fixed and determined beforehand).
Figure 2 illustrates an embodiment in which the cameras 202 and 203 and laser 204 are mounted on a hand-held assembly 205 (dotted). In this configuration the geometrical relationship between the laser fan beam and the cameras are fixed and known. Accordingly, savings in image processing speed can be effected.
The system shown in Figure 2 is, in most respects, the same as that in Figure 1 except that additional line extraction electronics 215 are required and a single spatial location receiver 208 is required for the hand-held assembly 205.
The embodiment shown in Figure 2 is also advantageous in that the sc~nning assembly is contained in a convenient 'package' or 'wand'. However, the present invention is not restrictecl to such a construction, and there may be applications where one or more arbitrarily located cameras and/or lasers are required.
In the applicants' p~oLoLy~e (which corresponds to the example shown in Figure 2) a single laser 204 and cylindrical lens 206 is positioned between two cameras 203 and 202, the laser and cameras being fixed in relation to each otherand mounted on a portable hand-held assembly in-licate(l by the dashed box 205.
The location and orientation of 205 is ~let~rrnin~(l by the spatial location receiver 208. The location of the object 200 is ~te~ Pd by a second spatial location receiver 201 with the tr~n~mitter 209 fixed in space throughout the scan. Each camera is s~d~ d from the laser by about 250mm and the optical axis of each -Wo 96/10205 = PCTJNZ95/Ooos7 carnera is angled to cross t'ne plane of the fan beam at an enclosed angle of about 30 degrees.
In operation, the hand-held unit 205 is directed so that the fan beam intersects the object 200. Each camera 203 and 202 images the profile at a knownand f~ed offset angle. The profiles are sampled at a rate sufficient so that if the hand held unit is moved at right angles to the plane of the beam (ie; the laser fan bearn is swept across the object) the scanned area can be considered as a series of profiles, each of which is imaged and reconstructed thus forming the scanned surface.
~ ltern~tively, the spatial location tr~n~mi~ter could be attached directly tO
the object in place of a separate receiver.
In an ~ ive embodiment, the object is held in fixed relation to the tr~n.~mitter throughout the course of the sc~nnina process. Similarly, any one component of the sc~nning system itself (camera or laser) may be held stationaryt'nroughout the scan. Its position and orientation is thus an arbitrary constant and need not be recorded by the spatial location system.
Where in the foregoing description reference has been made to elem~nt~
or integers having known equivalents, then such equivalents are included as if they were individually set forth.
Although the present invention has been described by way of example and efe~ ce to particular embodiments, it is to be understood that modifications andimprovements may be made without departing from the scope of the invention as set out in the appended claims.
Technical Field The invention relates to a non-contact mensuration system for sc~nning a three--limencional object in order to obtain data from which the shape of the object may be derived.
More particularly, although not exclusively, the present invention relates to a method and apparatus which does not require a fixed or predete~ illed geometrical relationship between its components per se as well as the object being sc~nn~l Any component(s) of the sc~nning system and/or the object may be moved albilldlily in space and therefore may be hand-held. Such a system is particularly suitable for obtaining measurements of irregular objects such as a human body or the like.
Background To The Invention Conventional laser surface scanners generally incorporate a rigid laser/
camera assembly. This assembly is usually fixed in space and the object being scanned is rotated or tr~n~l~te~l on a m~h~nic~l platform. AlL~llldlively7 the object is fixed and the laser/camera assembly is moved mech~nic~lly around or along it. Usually the movement takes the form of a rotation about one axis (contained within the object) or translation in one direction (along the object). In either case, the conventional sc~nnin~ process relies upon a predetermined knowledge of the geometrical relationships between the components of the sc~nning appaldlus.
Examples of such fixed-axis laser scanners include WO 94/15173, US
4,705,401, GB 2,240,623 A and US 5,193,120.
W O96/10205 PCT~NZ95/00097 --2 --Unless the object being sc~nnPd is a simple surface where every point on its surface is able to be 'seen' by the sc~nning apparatus in the course of a complete scan, a fixed-axis scarmer is unable to measure the complete surface ofan object in one scan orientation. This le~lcsellL~ a ~ignifir~nt limitation when sc~nning complex objects (for example, an entire human head). Features such as overhangs and indentations are unlikely to be visible in situations where a sc~nning device follows a fixed predetermined path.
Further, a fixed scan pattern or geometry cannot take into account variations in object detail or areas on the object' s surface which might be obscured by hl~ ellillg features protruding from the surface of the object.
Schulz ~WO 92/07233) provides, to some extent, for limited freedom of movement of a laser/camera assembly by tracking the position and orientation of a plurality of light sources fixed to the sc~nning assembly. The location of thelight sources (and therefore the sc~nning assembly) in three--limen~ional space is ~ termin~l by means of fixed photoelectronic sensors arranged in a pre~etermin~l static relationship to the fixed object. The light sources are time multiplexed with the sensors and the spatial location is derived from the three locations of the light sources. Therefore all three light sources (fixed to the sc~nning assembly) must be visible to all fixed photoelectronic sensors. There will generally be situations where at least one light source is obscured by the object itself, the operator or by tilting the assembly. Thus the Schulz scanner would appear to offer advantages over a fixed geometry scanner in only certain applications.
W O 96/10205 PCT~NZ95/00097 A further signifit~nt limitation inherent in the Schulz sc~nning system is that the object must remain stationary with respect to the reference frame defined by the photoelectronic sensors during the scan. This can present difficulties, particularly in medical applications, due to the potential for movement of the part of the subject being sc~nn~ . Accordingly, it would be desirable to be able to scan an object which may be arbitrarily located and oriented in space where suchalbiLla,"less is understood to include an object which may be moving, or to at least provide the public with a useful choice.
The present invention aLLcl~ s to overcome the limitations and disadvantages inherent in the prior art by providing a means and method of sc~nning an albiLlalily located and oriented object with the scanner components being arbitrarily located and oriented in space.
The present invention also provides a sc~nning system which may operate in ambient light and has reduced susceptibility to optical hlLclre,cllce in the process of im~ging the scanned portion of the object.
The present invention also provides for a non-optical technique for de~c- ,.,i"i"g the spatial locations and orientations of the components of the sc~nning system and object.
Disclosure Of The Invention In one aspect the invention provides an optical surface sc~nn~r including:
an illnmin~tion means adapted to emit a fan beam of light whereby the intersection between the fan beam and an object being scanned produces a profile having three-dimensional coordinates which lie in the plane of the fan beam;
a camera which images the profile in two dimensions;
a spatial location system adapted to obtain the relative positions and orientations between the object, illumination means and camera, wherein the three-~lim~n~ional coordinates of the profile are derived from the two-dimensional coordinates of the image and the relative positions and orientations of the illumin~rion means, camera and object.
In a further aspect the invention provides for an optical surface scanner inrlnt1ing:
an illnmin~tion means adapted to emit a pencil beam of light whereby the intersection between the pencil beam and an object being scanned produces a spot having three-dimensional coordinates which lie on the axis of the pencil beam;
a camera which images the spot in one or two dimensions;
a spatial location system adapted to obtain the relative positions and orientations between the object, illumination means and camera, wherein the three-~limrn~ional coordinates of the spot are derived from the one or two-dimensional coordinates of the image and the relative positions and orientations of the illllmin~tion means, camera and object.
In a further aspect the optical surface scanner inclucles:
one or more illnmin~tion means each adapted to emit a fan beam or a pencil beam of light whereby the intersection between each fan beam or pencil beam of light and the object being scanned produces a profile or spot having three-climen~ional coordinates which lie in the plane of the fan beam producing that profile or lie on the axis of the pencil producing that spot;
W O96/10205 PCT~NZ95100097 a plurality of cameras which image the profiles or spots;
means for distinguishing between the sources of illnmin~tion, a spatial location system adapted to obtain the relative positions and orientations between the object, illl-min~tion means and cameras;
wherein the three-~lim~n~ional coordinates of each profile or spot are derived from the two-dimensional or one/two dimenisional coordinates respectively of the corresponding image and the relative positions and orientations of the c~mer~, illumin~tion means and object.
Preferably the means for distinguishing between the sources of illl"~ ion uses time division multiplexing of the sources of illnmin~tion or dirr~ L wavelength sources.
Preferably the camera and illllmin~tion means are held in fixed relation.
Preferably the illllmin~tion means and camera(s) are mounted in a fixed relationship on a hand-held assembly.
Preferably the hand-held assembly includes two cameras and one illllmin~tion means wherein the illllmin~tion means is positioned between the cameras and the cameras and illll",il,~ion means are oriented so that the cameras image a profile or spot produced by the intersection of the fan beam or pencil beam respectively and the object being sc~nn~d.
Preferably the location and orientation of the hand-held assembly is deLe~ ed by means of the spatial location system.
-Preferably the spatial location system is adapted so as not to depend on m~int~ining line of sight between itS constituent components.
Preferably the spatial location system comprises a plurality of tr~ncmitrP,rs and a plurality of receivers.
Preferably the spatial location system uses electromagnetic fields.
Preferably the illumin~tion means is a laser.
Preferably the fan beam is produced by means of a cylindrical lens or similar means.
Preferably the optical surface scanner includes at least one optical band-pass filter per camera.
Preferably the band-pass filter comprises illLclrelcnce filters adapted to balance the contrast of the profile as imaged by the camera.
Preferably the one or more optical band-pass filters are located immediately in front of the camera.
The present invention further provides for a mensuration method for three-rlimencional objects comprising:
im~ging the intersection of a fan beam or pencil beam of light and an object being scanned;
mP~cll~ring the relative orientations and locations of a camera, illllmin~tion means and object;
calc~ ting three-dimensional coordinates of the intersection of the fan beam or pencil beam of light and the object being scanned;
repeating the above steps for a variety of orientations and/or locations of caInera and/or illnmin~tion means and/or object.
;
Preferably the relative position and orientation of the object, illnmin~tion means and camera are obtained by means of the spatial location system for each recorded image of the intersection of the fan beam or pencil beam and the object.
Further objects and advantages of the invention will become ~palcllL
from the following description which is given by way of example and with reference to the drawings in which:
Brief Deseli~Lion Of The D~wh~
Figure 1: illustrates a srh~m~tic diagram of an embodiment of an optical surface scanner incorporating one ill--min~tion means and one camera.
Figure 2: illustrates a scht?m~tic diagram of an embodiment of an optical surface scanner incorporating a single illllmin~rion means and two cameras embodied in a single sç~nning head.
Figure 3: illustrates a perspective view of the system in Figure 1 when used to scan a spherical object with a fan beam.
Referring to Figure 1, 100 is the object to be scanned. The major components of the sc~nning system are as follows. An illumination assembly 110 incorporates a low-power laser 101 which is directed through a cylindrical lens W O96/10205 PCTANZ9~/00097 102 thereby generating a fan beam. Other methods of producing a fan beam are envisged. For example, a point source with a slit aperture and collimator may produce the desired fan beam. An im~ging assembly 109 incorporates a two-dimensional (area array) electronic camera 103 which has an optical band-pass inte.~~ ce filter 104 covering its lens.
The spatial location system comprises a spatial location tr~n~mitter 105 and spatial location receivers 106, 107 and 108.
Spatial location electronics unit 111 and laser line extraction electronics 112 are interfaced to controlling electronics and power supply 113. 114 corresponds to a digital culll~u~
It is considered that an innovative feature enabling the construction of the truly arbitrary-geometry scanner described herein, resides in the use of spatiallocation devices 105, 106, 107 and 108 to obtain the relative position and orientalion (6 degrees of freedom) of the object 100, laser(s) 101 and camera(s)103. The spatial location system may correspond to an electronic device such as a Polhemus 'Fastrak' based on alternating electric fields or an Ascension 'Bird' based on pulsed static electric fields.
Referring to Figure 1, the dashed boxes 109 and 110 in~ te the components which are combined into the portable hand-held assemblies each associated with a spatial location receiver 107 and 108. In this case the im~ging assembly 109 consists of im~ging components (camera 103 and filter 104) and spatial location receiver 108. The illumin~tion assembly 110 consists of illllmin~tion components (laser 101 and cylindrical lens 102) and spatial location receiver 107.
WO 96/10205 PCT/~Z9~l00097 A three-dimensional representation of the system of Figure 1 is shown in - Figure 3. A cylindrical lens 302 spreads the beam emitted by the laser 301 into a fan beam 303 so that the intersection between the fan beam and the object being sc~nnt~l 300 produces a profile 304. The camera includes a lens 307 a CCD array 306 and a filter 308 in front of the camera. The image 305 of the profile 304 isobtained by the camera when viewing the object at an offset angle. The two-~lim~n.~ional coordinates of the image of the profile are recorded.
The object 300, laser 301 and camera (306 and 307) are associated with spatial location receivers 310, 311 and 312 respectively thereby allowing their orientation and location in space to be measured in real time. Thus the laser and camera may be moved albiLldlily in space so as to record desired profiles thereby building a reconstructed surface. Alternatively, the object may be movedso as to record the desired profiles.
In use, profiles 304 are recorded by directing the laser fan beam at the object while .simlllt~n~ously positioning the camera so that the profile is visible to it. The utility of the present system is illustrated by situations where features of the object might obscure either the illnmin~tion or the camera's field of view when using a fixed illnmin~ti~n/im~ging system. If such a geometry is encountered, the illllmin~tion and/or camera and/or object may be moved into an orientation or position where the profile is visible to the camera.
The spatial location system 310, 311, 312 and 313 provides the required spatial location and orientation information for the object, camera and laser whereupon the three-~iim~n~ional coordinates of the profile are (let~rmin~l fromthe t~,vo-~lim~on~ional pixel coordinates in the camera CCD array 306.
The scanner may include an hltelr~lcllce filter 308 which allows sc~nning under ambient light conditions. A further purpose of the interference filter is to balance the profile contrast as seen by the camera as discussed below.
IllLelrelcnce filters present a lower wavelength pass-band to light from off-axis and this effect is exploited. The intensity of the profile typically tiimini.~h~s with tli~t~n~e from its centre. Therefore, if a filter is chosen with a pass-band centred above the laser wavelength, the distal ends of the laser profile are subject to less attenuation by the filter. This balances the laser profile contrast as seen by the camera. Such an application of an inL~lrelcllce filter is, to the applicants' knowledge, novel.
Referring to Figure 1 the output from each two-dimensional camera 103 (such as a CCD array) is processed a~lo~.liately by line extraction electronics 112 to produce a list of pixels (two-dimensional camera coordinates) at which the intensity exceeds some threshold (constant or adaptive).
The controlling electronics unit 113 performs such tasks as synchronising the cameras, lasers and spatial location receivers, monitoring switches, supplying power, preproce~ing and interfacing this data slream to the digital input/outputport of the host digital cOlll~ul~l 114. The digital co~llL,ul~l is programmed to reconstruct the three-dimensional coordinates of the object's surface. Such a reconstruction may employ LLdllsro-lllation methods known in the art which convert each two-~limen~ional point in the profile, viewed by the camera, to a three-dimensional point in the camera frame of reference and then to a three-~limen~ional point in the object frame of reference.
wo 96/10205 PCTINZ95100097 The reconstruction process~ must der~ in~ the coordinates of a three-dimensional point, in the object's fraIne of reference, corresponding to each pixel identified by the laser line extraction electronics.
This can be accomplished in a number of ways. For example, if a simple linear model of the camera is assumed, then a ray can be traced from the pixel location through the centre of the lens whereby the intersection of the ray withthe plane of the laser will determine the three-t1imen.cional coordinate of the i11umin~ted spot on the object, the camera coordinates and the laser plane having been transformed into the object's coordinate system using the spatial location information.
~ 1tern~tively, a 'look-up table' can be used to transform from pixel coordinates to three-~limen.cional coordinates in the camera frame of reference,and then these coordinates can be transformed into the objects coordinate systemusing the spatial location hlrollllation.
The look-up table can be produced by an empirical calibration process.
Therefore allowances may be made for non-linear effects such as lens distortion as well as providing for a faster means of data conversion in the co,lll,uL~l.
It will be appreciated that the location of the object being sc~nn~l is constantly determined by means of the spatial location system. Therefore, so long as the receiver associated with the object remains in fixed relation to the object, the object is free to move during the sc~nnin~ process.
It is to be appreciated that both the number of cameras and lasers and/or the geometry of the system can be varied if desired. If, for example, any two or Wo 96/10205 PC~r/NZss/00097 more components retain a fixed relative spatial relationship throughout the scan, then the position and orientation of the combined assembly may be recorded by a single spatial location receiver (the relative locations and orientations of theassembly components being fixed and determined beforehand).
Figure 2 illustrates an embodiment in which the cameras 202 and 203 and laser 204 are mounted on a hand-held assembly 205 (dotted). In this configuration the geometrical relationship between the laser fan beam and the cameras are fixed and known. Accordingly, savings in image processing speed can be effected.
The system shown in Figure 2 is, in most respects, the same as that in Figure 1 except that additional line extraction electronics 215 are required and a single spatial location receiver 208 is required for the hand-held assembly 205.
The embodiment shown in Figure 2 is also advantageous in that the sc~nning assembly is contained in a convenient 'package' or 'wand'. However, the present invention is not restrictecl to such a construction, and there may be applications where one or more arbitrarily located cameras and/or lasers are required.
In the applicants' p~oLoLy~e (which corresponds to the example shown in Figure 2) a single laser 204 and cylindrical lens 206 is positioned between two cameras 203 and 202, the laser and cameras being fixed in relation to each otherand mounted on a portable hand-held assembly in-licate(l by the dashed box 205.
The location and orientation of 205 is ~let~rrnin~(l by the spatial location receiver 208. The location of the object 200 is ~te~ Pd by a second spatial location receiver 201 with the tr~n~mitter 209 fixed in space throughout the scan. Each camera is s~d~ d from the laser by about 250mm and the optical axis of each -Wo 96/10205 = PCTJNZ95/Ooos7 carnera is angled to cross t'ne plane of the fan beam at an enclosed angle of about 30 degrees.
In operation, the hand-held unit 205 is directed so that the fan beam intersects the object 200. Each camera 203 and 202 images the profile at a knownand f~ed offset angle. The profiles are sampled at a rate sufficient so that if the hand held unit is moved at right angles to the plane of the beam (ie; the laser fan bearn is swept across the object) the scanned area can be considered as a series of profiles, each of which is imaged and reconstructed thus forming the scanned surface.
~ ltern~tively, the spatial location tr~n~mi~ter could be attached directly tO
the object in place of a separate receiver.
In an ~ ive embodiment, the object is held in fixed relation to the tr~n.~mitter throughout the course of the sc~nnina process. Similarly, any one component of the sc~nning system itself (camera or laser) may be held stationaryt'nroughout the scan. Its position and orientation is thus an arbitrary constant and need not be recorded by the spatial location system.
Where in the foregoing description reference has been made to elem~nt~
or integers having known equivalents, then such equivalents are included as if they were individually set forth.
Although the present invention has been described by way of example and efe~ ce to particular embodiments, it is to be understood that modifications andimprovements may be made without departing from the scope of the invention as set out in the appended claims.
Claims (19)
1. An optical surface scanner including:
an illumination means adapted to emit a fan beam of light whereby the intersection between the fan beam and an object being scanned produces a profile having three-dimensional coordinates which lie in the plane of the fan beam;
a camera which images the profile in two dimensions;
a spatial location system which senses the relative positions and orientations of elements of the scanner to determine the relative positions and orientations of the object, illumination means and camera, the three-dimensional coordinates of the profile derived from the two-dimensional coordinates of the image and the relative positions and orientations of the illumination means, camera and object.
an illumination means adapted to emit a fan beam of light whereby the intersection between the fan beam and an object being scanned produces a profile having three-dimensional coordinates which lie in the plane of the fan beam;
a camera which images the profile in two dimensions;
a spatial location system which senses the relative positions and orientations of elements of the scanner to determine the relative positions and orientations of the object, illumination means and camera, the three-dimensional coordinates of the profile derived from the two-dimensional coordinates of the image and the relative positions and orientations of the illumination means, camera and object.
2. An optical surface scanner including:
an illumination means adapted to emit a pencil beam of light whereby the intersection between the pencil beam and an object being scanned produces a spot having three-dimensional coordinates which lie on the axis of the pencil beam;
a camera which images the spot in one or two dimensions;
a spatial location system which senses the relative positions and orientations of elements of the scanner to determine the relative positions and orientations of the object, illumination means and camera, the three-dimensional coordinates of the spot derived from the one or two-dimensional coordinates of the image and the relative positions and orientation of the illumination means, camera and object.
an illumination means adapted to emit a pencil beam of light whereby the intersection between the pencil beam and an object being scanned produces a spot having three-dimensional coordinates which lie on the axis of the pencil beam;
a camera which images the spot in one or two dimensions;
a spatial location system which senses the relative positions and orientations of elements of the scanner to determine the relative positions and orientations of the object, illumination means and camera, the three-dimensional coordinates of the spot derived from the one or two-dimensional coordinates of the image and the relative positions and orientation of the illumination means, camera and object.
3. An optical surface scanner including:
one or more illumination means each emitting a fan beam of light whereby the intersection between each fan beam of light and the object being scanned produces a profile having three-dimensional coordinates which lie in the plane of the fan beam producing that profile;
a plurality of cameras which image the profiles;
means for distinguishing between illumination produced by the one or more illumination means;
a spatial location system which senses the relative positions and orientations of elements of the scanner to determine the relative positions and orientations of the object, illumination means and cameras, the three-dimensional coordinates of each profile derived from the two-dimensional coordinates or one/two dimensional coordinates respectively of the corresponding image and the relative positions and orientations of the cameras, illumination means and object.
one or more illumination means each emitting a fan beam of light whereby the intersection between each fan beam of light and the object being scanned produces a profile having three-dimensional coordinates which lie in the plane of the fan beam producing that profile;
a plurality of cameras which image the profiles;
means for distinguishing between illumination produced by the one or more illumination means;
a spatial location system which senses the relative positions and orientations of elements of the scanner to determine the relative positions and orientations of the object, illumination means and cameras, the three-dimensional coordinates of each profile derived from the two-dimensional coordinates or one/two dimensional coordinates respectively of the corresponding image and the relative positions and orientations of the cameras, illumination means and object.
4. An optical surface scanner as claimed in claim 3 wherein the means for distinguishing between the sources of illumination uses time division multiplexing of the sources of illumination or different wavelength sources.
5. An optical surface scanner as claimed in either of claims 1 or 3 wherein the fan beam is produced by means of a cylindrical lens.
6. An optical surface scanner as claimed in claim 3 wherein the cameras and illumination means are held in fixed relation.
7. An optical surface scanner as claimed in claim 6 wherein the illumination means and cameras are mounted in a fixed relationship on a hand-held assembly.
8. An optical surface scanner as claimed in claim 7 wherein the hand-held assembly includes two cameras and one illumination means, wherein the illumination means is positioned between the cameras and the cameras and illumination means are oriented so that the cameras image a profile produced by the intersection of the fan beam and the object being scanned.
9. An optical surface scanner as claimed in claim 7 wherein the location and orientation of the hand-held assembly is determined by means of the spatial location system.
10. An optical surface scanner as claimed in claim 3 wherein the spatial location system is adapted so as not to depend on maintaining line of sight between its constituent components.
11. An optical surface scanner as claimed in claim 3 wherein the spatial location system comprises a plurality of transmitters and a plurality of receivers.
12. An optical surface scanner as claimed in claim 3 wherein the spatial location system uses electromagnetic fields.
13. An optical surface scanner as claimed in claim 3 wherein the illumination means is a laser.
14. An optical surface scanner as claimed in claim 3 including at least one optical band-pass filter per camera.
15. An optical surface scanner as claimed in claim 14 wherein the band-pass filter comprises interference filters which balances the contrast of the profile as imaged by the camera.
16. An optical surface scanner as claimed in claim 14 wherein the one or more optical band-pass filters are located immediately in front of the camera.
17. A mensuration method for three-dimensional objects comprising:
imaging the intersection of a fan beam of light and an object being scanned;
measuring the relative orientations and locations of a camera, illumination means and object;
calculating three-dimensional coordinates of the intersection of the fan beam or pencil beam of light and the object being scanned;
repeating the above steps for a variety of orientations and locations of camera and illumination means and object.
imaging the intersection of a fan beam of light and an object being scanned;
measuring the relative orientations and locations of a camera, illumination means and object;
calculating three-dimensional coordinates of the intersection of the fan beam or pencil beam of light and the object being scanned;
repeating the above steps for a variety of orientations and locations of camera and illumination means and object.
18. A mensuration method for three-dimensional objects comprising:
imaging the intersection of a fan beam of light and an object being scanned;
measuring the relative orientations and locations of a camera, illumination means and object;
calculating three-dimensional coordinates of the intersection of the fan beam or pencil beam of light and the object being scanned;
repeating the above steps for a variety of orientations or locations of camera or illumination means or object.
imaging the intersection of a fan beam of light and an object being scanned;
measuring the relative orientations and locations of a camera, illumination means and object;
calculating three-dimensional coordinates of the intersection of the fan beam or pencil beam of light and the object being scanned;
repeating the above steps for a variety of orientations or locations of camera or illumination means or object.
19. A mensuration method for three-dimensional objects as claimed in claim 17 or 18 wherein the relative position and orientation of the object, illumination means and camera are obtained by means of a spatial location system for each recorded image of the intersection of the fan beam and the object.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
NZ26456494 | 1994-09-28 | ||
NZ264564 | 1994-09-28 | ||
PCT/NZ1995/000097 WO1996010205A1 (en) | 1994-09-28 | 1995-09-28 | Arbitrary-geometry laser surface scanner |
Publications (2)
Publication Number | Publication Date |
---|---|
CA2201107A1 CA2201107A1 (en) | 1996-04-04 |
CA2201107C true CA2201107C (en) | 2002-11-12 |
Family
ID=19924949
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CA002201107A Expired - Fee Related CA2201107C (en) | 1994-09-28 | 1995-09-28 | Arbitrary-geometry laser surface scanner |
Country Status (7)
Country | Link |
---|---|
US (1) | US5969822A (en) |
AU (1) | AU3621795A (en) |
CA (1) | CA2201107C (en) |
DE (1) | DE19581099B4 (en) |
GB (1) | GB2308186B (en) |
NZ (1) | NZ293713A (en) |
WO (1) | WO1996010205A1 (en) |
Families Citing this family (60)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
GB9515311D0 (en) | 1995-07-26 | 1995-09-20 | 3D Scanners Ltd | Stripe scanners and methods of scanning |
US6553138B2 (en) * | 1998-12-30 | 2003-04-22 | New York University | Method and apparatus for generating three-dimensional representations of objects |
NO313113B1 (en) * | 1999-07-13 | 2002-08-12 | Metronor Asa | System for scanning large geometry of objects |
DE50000335D1 (en) | 2000-04-05 | 2002-09-05 | Brainlab Ag | Referencing a patient in a medical navigation system using illuminated light points |
US6594516B1 (en) * | 2000-07-18 | 2003-07-15 | Koninklijke Philips Electronics, N.V. | External patient contouring |
US7625335B2 (en) | 2000-08-25 | 2009-12-01 | 3Shape Aps | Method and apparatus for three-dimensional optical scanning of interior surfaces |
DK1368986T3 (en) | 2001-03-02 | 2012-02-27 | 3Shape As | Method of modeling custom made earpieces |
FI113293B (en) * | 2001-04-19 | 2004-03-31 | Mapvision Oy | A method for indicating a point in a measuring space |
KR20030033174A (en) * | 2001-10-18 | 2003-05-01 | 이정현 | Method for Measuring 3-D Shape of Cultural Relics Having Reflective Surface |
WO2003054683A2 (en) * | 2001-12-07 | 2003-07-03 | Canesta Inc. | User interface for electronic devices |
US7881896B2 (en) | 2002-02-14 | 2011-02-01 | Faro Technologies, Inc. | Portable coordinate measurement machine with integrated line laser scanner |
US10242255B2 (en) | 2002-02-15 | 2019-03-26 | Microsoft Technology Licensing, Llc | Gesture recognition system using depth perceptive sensors |
US7340077B2 (en) * | 2002-02-15 | 2008-03-04 | Canesta, Inc. | Gesture recognition system using depth perceptive sensors |
US7151530B2 (en) | 2002-08-20 | 2006-12-19 | Canesta, Inc. | System and method for determining an input selected by a user through a virtual interface |
US7526120B2 (en) * | 2002-09-11 | 2009-04-28 | Canesta, Inc. | System and method for providing intelligent airbag deployment |
US20040066500A1 (en) * | 2002-10-02 | 2004-04-08 | Gokturk Salih Burak | Occupancy detection and measurement system and method |
US7439074B2 (en) * | 2003-09-30 | 2008-10-21 | Hoa Duc Nguyen | Method of analysis of alcohol by mass spectrometry |
US10277290B2 (en) | 2004-04-02 | 2019-04-30 | Rearden, Llc | Systems and methods to exploit areas of coherence in wireless systems |
US9826537B2 (en) | 2004-04-02 | 2017-11-21 | Rearden, Llc | System and method for managing inter-cluster handoff of clients which traverse multiple DIDO clusters |
US8654815B1 (en) | 2004-04-02 | 2014-02-18 | Rearden, Llc | System and method for distributed antenna wireless communications |
US10425134B2 (en) | 2004-04-02 | 2019-09-24 | Rearden, Llc | System and methods for planned evolution and obsolescence of multiuser spectrum |
US9819403B2 (en) | 2004-04-02 | 2017-11-14 | Rearden, Llc | System and method for managing handoff of a client between different distributed-input-distributed-output (DIDO) networks based on detected velocity of the client |
US7340316B2 (en) * | 2004-06-28 | 2008-03-04 | Hanger Orthopedic Group, Inc. | System and method for producing medical devices |
US20060055706A1 (en) * | 2004-09-15 | 2006-03-16 | Perlman Stephen G | Apparatus and method for capturing the motion of a performer |
US7375826B1 (en) * | 2004-09-23 | 2008-05-20 | The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration (Nasa) | High speed three-dimensional laser scanner with real time processing |
WO2006069491A1 (en) * | 2004-12-31 | 2006-07-06 | Intel Corporation | Remote logging mechanism |
US8009871B2 (en) | 2005-02-08 | 2011-08-30 | Microsoft Corporation | Method and system to segment depth images and to detect shapes in three-dimensionally acquired data |
GB0517992D0 (en) * | 2005-09-05 | 2005-10-12 | Sld Ltd | Laser imaging |
US8625854B2 (en) * | 2005-09-09 | 2014-01-07 | Industrial Research Limited | 3D scene scanner and a position and orientation system |
US8659668B2 (en) | 2005-10-07 | 2014-02-25 | Rearden, Llc | Apparatus and method for performing motion capture using a random pattern on capture surfaces |
KR20080064155A (en) * | 2005-10-14 | 2008-07-08 | 어플라이드 리써치 어쏘시에이츠 뉴질랜드 리미티드 | A method of monitoring a surface feature and apparatus therefor |
DE102006031142B4 (en) * | 2006-07-05 | 2012-02-16 | Prüf- und Forschungsinstitut Pirmasens e.V. | Method and device for three-dimensional measurement and detection of the complete object surface of a spherical measurement object such as a sports ball |
WO2008016645A2 (en) * | 2006-07-31 | 2008-02-07 | Onlive, Inc. | System and method for performing motion capture and image reconstruction |
US20100231692A1 (en) * | 2006-07-31 | 2010-09-16 | Onlive, Inc. | System and method for performing motion capture and image reconstruction with transparent makeup |
US20080071559A1 (en) * | 2006-09-19 | 2008-03-20 | Juha Arrasvuori | Augmented reality assisted shopping |
US20100091104A1 (en) * | 2006-09-27 | 2010-04-15 | Georgia Tech Research Corporation | Systems and methods for the measurement of surfaces |
US7256899B1 (en) * | 2006-10-04 | 2007-08-14 | Ivan Faul | Wireless methods and systems for three-dimensional non-contact shape sensing |
DE102007025520A1 (en) * | 2007-05-31 | 2008-12-04 | Khs Ag | Method for aligning containers and device for carrying out the method |
US7578178B2 (en) * | 2007-09-28 | 2009-08-25 | United Technologies Corporation | Method of inspecting turbine internal cooling features using non-contact scanners |
DE102008047816B4 (en) * | 2008-09-18 | 2011-08-25 | Steinbichler Optotechnik GmbH, 83115 | Device for determining the 3D coordinates of an object, in particular a tooth |
US8505209B2 (en) | 2009-10-27 | 2013-08-13 | N.E. Solutionz, Llc | Skin and wound assessment tool |
US9161716B2 (en) | 2009-10-27 | 2015-10-20 | N.E. Solutionz, Llc | Diagnostic imaging system for skin and affliction assessment |
US8276287B2 (en) | 2009-10-27 | 2012-10-02 | N.E. Solutionz, Llc | Skin and wound assessment tool |
WO2011063306A1 (en) | 2009-11-19 | 2011-05-26 | Modulated Imaging Inc. | Method and apparatus for analysis of turbid media via single-element detection using structured illumination |
EP2365278A1 (en) * | 2010-02-22 | 2011-09-14 | Snap-on Equipment Srl a unico socio | Scanner system |
US8467992B1 (en) * | 2010-09-15 | 2013-06-18 | The Boeing Company | Vision based location and measurement device and methods |
US8687172B2 (en) | 2011-04-13 | 2014-04-01 | Ivan Faul | Optical digitizer with improved distance measurement capability |
US9179844B2 (en) | 2011-11-28 | 2015-11-10 | Aranz Healthcare Limited | Handheld skin measuring or monitoring device |
US8892192B2 (en) | 2012-11-07 | 2014-11-18 | Modulated Imaging, Inc. | Efficient modulated imaging |
US11189917B2 (en) | 2014-04-16 | 2021-11-30 | Rearden, Llc | Systems and methods for distributing radioheads |
US10488535B2 (en) | 2013-03-12 | 2019-11-26 | Rearden, Llc | Apparatus and method for capturing still images and video using diffraction coded imaging techniques |
US9973246B2 (en) | 2013-03-12 | 2018-05-15 | Rearden, Llc | Systems and methods for exploiting inter-cell multiplexing gain in wireless cellular systems via distributed input distributed output technology |
US9923657B2 (en) | 2013-03-12 | 2018-03-20 | Rearden, Llc | Systems and methods for exploiting inter-cell multiplexing gain in wireless cellular systems via distributed input distributed output technology |
RU2767777C2 (en) | 2013-03-15 | 2022-03-21 | Риарден, Ллк | Systems and methods of radio frequency calibration using the principle of reciprocity of channels in wireless communication with distributed input - distributed output |
ITTO20130202A1 (en) * | 2013-03-15 | 2014-09-16 | Torino Politecnico | DEVICE AND THREE-DIMENSIONAL SCANNING SYSTEM, AND RELATIVE METHOD. |
US10013527B2 (en) | 2016-05-02 | 2018-07-03 | Aranz Healthcare Limited | Automatically assessing an anatomical surface feature and securely managing information related to the same |
US11116407B2 (en) | 2016-11-17 | 2021-09-14 | Aranz Healthcare Limited | Anatomical surface assessment methods, devices and systems |
WO2018185560A2 (en) | 2017-04-04 | 2018-10-11 | Aranz Healthcare Limited | Anatomical surface assessment methods, devices and systems |
CN107170035B (en) * | 2017-07-12 | 2023-08-25 | 朱培恒 | High-reflection object scanning method and system |
DE102018105132B4 (en) * | 2018-03-06 | 2023-07-13 | Haag-Streit Gmbh | triangulation device |
Family Cites Families (46)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US35816A (en) * | 1862-07-08 | Improved washing-machine | ||
US3821469A (en) * | 1972-05-15 | 1974-06-28 | Amperex Electronic Corp | Graphical data device |
US4017858A (en) * | 1973-07-30 | 1977-04-12 | Polhemus Navigation Sciences, Inc. | Apparatus for generating a nutating electromagnetic field |
US3983474A (en) * | 1975-02-21 | 1976-09-28 | Polhemus Navigation Sciences, Inc. | Tracking and determining orientation of object using coordinate transformation means, system and process |
FR2416480A1 (en) * | 1978-02-03 | 1979-08-31 | Thomson Csf | RADIANT SOURCE LOCATION DEVICE AND STEERING TRACKING SYSTEM INCLUDING SUCH A DEVICE |
US4402608A (en) * | 1980-10-02 | 1983-09-06 | Solid Photography Inc. | Room scanning system using multiple camera and projector sensors |
US4585350A (en) * | 1983-01-28 | 1986-04-29 | Pryor Timothy R | Pulsed robotic inspection |
NL8302228A (en) * | 1983-06-22 | 1985-01-16 | Optische Ind De Oude Delft Nv | MEASURING SYSTEM FOR USING A TRIANGULAR PRINCIPLE, CONTACT-FREE MEASURING A DISTANCE GIVEN BY A SURFACE CONTOUR TO AN OBJECTIVE LEVEL. |
DE3342675A1 (en) * | 1983-11-25 | 1985-06-05 | Fa. Carl Zeiss, 7920 Heidenheim | METHOD AND DEVICE FOR CONTACTLESS MEASUREMENT OF OBJECTS |
US4753528A (en) * | 1983-12-13 | 1988-06-28 | Quantime, Inc. | Laser archery distance device |
US4649504A (en) * | 1984-05-22 | 1987-03-10 | Cae Electronics, Ltd. | Optical position and orientation measurement techniques |
US4775235A (en) * | 1984-06-08 | 1988-10-04 | Robotic Vision Systems, Inc. | Optical spot scanning system for use in three-dimensional object inspection |
DE3423135A1 (en) * | 1984-06-22 | 1986-01-02 | Dornier Gmbh, 7990 Friedrichshafen | METHOD FOR READING A DISTANCE IMAGE LINE |
US4705395A (en) * | 1984-10-03 | 1987-11-10 | Diffracto Ltd. | Triangulation data integrity |
JPS6186606A (en) * | 1984-10-05 | 1986-05-02 | Hitachi Ltd | Configuration measuring method without contact |
DE3502634A1 (en) * | 1985-01-26 | 1985-06-20 | Deutsche Forschungs- und Versuchsanstalt für Luft- und Raumfahrt e.V., 5000 Köln | OPTICAL-ELECTRONIC DISTANCE METER |
US4782239A (en) * | 1985-04-05 | 1988-11-01 | Nippon Kogaku K. K. | Optical position measuring apparatus |
SE447848B (en) * | 1985-06-14 | 1986-12-15 | Anders Bengtsson | INSTRUMENTS FOR SEATING SURFACE TOPOGRAPHY |
US4743771A (en) * | 1985-06-17 | 1988-05-10 | View Engineering, Inc. | Z-axis height measurement system |
US4737032A (en) * | 1985-08-26 | 1988-04-12 | Cyberware Laboratory, Inc. | Surface mensuration sensor |
DE3677185D1 (en) * | 1985-08-12 | 1991-02-28 | Cyberware Lab Inc | FAST DIGITIZER OF A THREE-DIMENSIONAL SURFACE. |
US4705401A (en) * | 1985-08-12 | 1987-11-10 | Cyberware Laboratory Inc. | Rapid three-dimensional surface digitizer |
JPH0619243B2 (en) * | 1985-09-19 | 1994-03-16 | 株式会社トプコン | Coordinate measuring method and apparatus thereof |
US4709156A (en) * | 1985-11-27 | 1987-11-24 | Ex-Cell-O Corporation | Method and apparatus for inspecting a surface |
US4794262A (en) * | 1985-12-03 | 1988-12-27 | Yukio Sato | Method and apparatus for measuring profile of three-dimensional object |
US4760851A (en) * | 1986-03-31 | 1988-08-02 | Faro Medical Technologies Inc. | 3-dimensional digitizer for skeletal analysis |
SE469321B (en) * | 1986-04-14 | 1993-06-21 | Joenkoepings Laens Landsting | SET AND DEVICE TO MAKE A MODIFIED THREE-DIMENSIONAL IMAGE OF AN ELASTIC DEFORMABLE PURPOSE |
US4822163A (en) * | 1986-06-26 | 1989-04-18 | Robotic Vision Systems, Inc. | Tracking vision sensor |
US4767934A (en) * | 1986-07-02 | 1988-08-30 | Honeywell Inc. | Active ranging system |
US4733969A (en) * | 1986-09-08 | 1988-03-29 | Cyberoptics Corporation | Laser probe for determining distance |
US4761072A (en) * | 1986-09-30 | 1988-08-02 | Diffracto Ltd. | Electro-optical sensors for manual control |
US4764015A (en) * | 1986-12-31 | 1988-08-16 | Owens-Illinois Television Products Inc. | Method and apparatus for non-contact spatial measurement |
DE3703422A1 (en) * | 1987-02-05 | 1988-08-18 | Zeiss Carl Fa | OPTOELECTRONIC DISTANCE SENSOR |
US4745290A (en) * | 1987-03-19 | 1988-05-17 | David Frankel | Method and apparatus for use in making custom shoes |
US4836778A (en) * | 1987-05-26 | 1989-06-06 | Vexcel Corporation | Mandibular motion monitoring system |
US4829373A (en) * | 1987-08-03 | 1989-05-09 | Vexcel Corporation | Stereo mensuration apparatus |
JPH022469A (en) * | 1987-12-21 | 1990-01-08 | Univ Western Australia | Vision system |
US4982188A (en) * | 1988-09-20 | 1991-01-01 | Grumman Aerospace Corporation | System for measuring positional characteristics of an ejected object |
FR2642833B1 (en) * | 1989-02-06 | 1991-05-17 | Vision 3D | CALIBRATION METHOD OF A THREE-DIMENSIONAL SHAPE ACQUISITION SYSTEM AND ACQUISITION SYSTEM FOR IMPLEMENTING SAID METHOD |
GB9002193D0 (en) * | 1990-01-31 | 1990-03-28 | Starsmore Neil | An imaging process for detecting the shape of three dimensional objects |
US5198877A (en) * | 1990-10-15 | 1993-03-30 | Pixsys, Inc. | Method and apparatus for three-dimensional non-contact shape sensing |
US5193120A (en) * | 1991-02-27 | 1993-03-09 | Mechanical Technology Incorporated | Machine vision three dimensional profiling system |
DE4301538A1 (en) * | 1992-03-17 | 1994-07-28 | Peter Dr Ing Brueckner | Method and arrangement for contactless three-dimensional measurement, in particular for measuring denture models |
US5307072A (en) * | 1992-07-09 | 1994-04-26 | Polhemus Incorporated | Non-concentricity compensation in position and orientation measurement systems |
WO1994015173A1 (en) * | 1992-12-18 | 1994-07-07 | 3D Scanners Ltd. | Scanning sensor |
US5402582A (en) * | 1993-02-23 | 1995-04-04 | Faro Technologies Inc. | Three dimensional coordinate measuring apparatus |
-
1995
- 1995-09-28 GB GB9705754A patent/GB2308186B/en not_active Expired - Fee Related
- 1995-09-28 AU AU36217/95A patent/AU3621795A/en not_active Abandoned
- 1995-09-28 DE DE19581099T patent/DE19581099B4/en not_active Expired - Fee Related
- 1995-09-28 CA CA002201107A patent/CA2201107C/en not_active Expired - Fee Related
- 1995-09-28 NZ NZ293713A patent/NZ293713A/en not_active IP Right Cessation
- 1995-09-28 WO PCT/NZ1995/000097 patent/WO1996010205A1/en active Application Filing
- 1995-09-28 US US08/809,784 patent/US5969822A/en not_active Expired - Lifetime
Also Published As
Publication number | Publication date |
---|---|
DE19581099T1 (en) | 1997-09-18 |
GB2308186A (en) | 1997-06-18 |
CA2201107A1 (en) | 1996-04-04 |
US5969822A (en) | 1999-10-19 |
NZ293713A (en) | 1997-09-22 |
GB9705754D0 (en) | 1997-05-07 |
WO1996010205A1 (en) | 1996-04-04 |
AU3621795A (en) | 1996-04-19 |
GB2308186B (en) | 1999-06-09 |
DE19581099B4 (en) | 2006-07-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CA2201107C (en) | Arbitrary-geometry laser surface scanner | |
US7903261B2 (en) | Controlling a projected pattern | |
US5991437A (en) | Modular digital audio system having individualized functional modules | |
US20020100884A1 (en) | Digital 3-D model production method and apparatus | |
US5198877A (en) | Method and apparatus for three-dimensional non-contact shape sensing | |
AU2007334855B2 (en) | Visual aid with three-dimensional image acquisition | |
US9762883B2 (en) | Balancing colors in a scanned three-dimensional image | |
KR101605386B1 (en) | Optical measurement method and measurement system for determining 3D coordinates on a measurement object surface | |
US20040201856A1 (en) | Laser digitizer system for dental applications | |
EP0247833A2 (en) | Method and system for high-speed, 3-D imaging of an object at a vision station | |
US7009717B2 (en) | Optical probe for scanning the features of an object and methods therefor | |
EP3598174B1 (en) | Laser scanner with enhanced dymanic range imaging | |
US11035980B2 (en) | Laser scanner with projector | |
US6765606B1 (en) | Three dimension imaging by dual wavelength triangulation | |
JP2002529685A (en) | Tomographic reconstruction of electronic components from shadow sensor data. | |
GB2246044A (en) | A zoom lens for a variable depth range camera | |
CN1538461A (en) | X-ray equipment and method for generating surface image | |
JP2021076603A (en) | Photoelectric sensor and object detection method | |
US20240004076A1 (en) | Artifact filtering using artificial intelligence | |
Blais et al. | A very compact real time 3-D range sensor for mobile robot applications | |
CN116338712A (en) | Survey apparatus | |
Samuelson et al. | A device for three-dimensional registration of human movement | |
Livingstone et al. | Development of a real-time laser scanning system for object recognition, inspection, and robot control | |
EP1391690A1 (en) | Optical probe for measuring features of an object and methods therefor | |
George et al. | Optical System For Robot Guidance |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
EEER | Examination request | ||
MKLA | Lapsed |
Effective date: 20140930 |