CA2388719A1 - Rapid, automatic measurement of the eye's wave aberraion - Google Patents

Rapid, automatic measurement of the eye's wave aberraion Download PDF

Info

Publication number
CA2388719A1
CA2388719A1 CA002388719A CA2388719A CA2388719A1 CA 2388719 A1 CA2388719 A1 CA 2388719A1 CA 002388719 A CA002388719 A CA 002388719A CA 2388719 A CA2388719 A CA 2388719A CA 2388719 A1 CA2388719 A1 CA 2388719A1
Authority
CA
Canada
Prior art keywords
further characterized
search
eye
image data
spots
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
CA002388719A
Other languages
French (fr)
Inventor
David R. Williams
William J. Vaughn
Benjamin D. Singer
Heidi Hofer
Geun Young Yoon
Pablo Artal
Juan Luis Aragon
Pedro Prieto
Fernando Vargas
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
University of Rochester
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Publication of CA2388719A1 publication Critical patent/CA2388719A1/en
Abandoned legal-status Critical Current

Links

Classifications

    • AHUMAN NECESSITIES
    • A61MEDICAL OR VETERINARY SCIENCE; HYGIENE
    • A61BDIAGNOSIS; SURGERY; IDENTIFICATION
    • A61B3/00Apparatus for testing the eyes; Instruments for examining the eyes
    • A61B3/10Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions
    • A61B3/1015Objective types, i.e. instruments for examining the eyes independent of the patients' perceptions or reactions for wavefront analysis
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01JMEASUREMENT OF INTENSITY, VELOCITY, SPECTRAL CONTENT, POLARISATION, PHASE OR PULSE CHARACTERISTICS OF INFRARED, VISIBLE OR ULTRAVIOLET LIGHT; COLORIMETRY; RADIATION PYROMETRY
    • G01J9/00Measuring optical phase difference; Determining degree of coherence; Measuring optical wavelength

Abstract

A wavefront aberration of an eye is determined, e.g., in real time. The eye is illuminated, and the light reflected from the retina is converted into spots with a device such as a Hartmann-Shack detector. The displacement of each spot from where it would be in the absence of aberration allows calculation of the aberration. Each spot is located by an iterative technique in which a corresponding centroid is located in a box drawn on the image data, a smaller box is defined around the centroid, the centroid is located in the smaller box, and so on. The wavefront aberration is calculated from the centroid locations by using a matrix in which unusable data can be eliminated simply by eliminating rows of the matrix. Aberrations for different pupil sizes are handled in data taken for a single pupil size by renormalization.

Description

RAPID, .4UTOMAT1C MEASUItEMENI
OF THE EYE'S WAVE ABERRATION
s BACKGROUND OF THE INVENTION
The present invention is directed to a wavefront sensor and more particularly to a wavefront sensor for ophthalmic applications.
It is known in astronomy to detect wavefront aberrations caused by the atmosphere through the use of a Hartmann-Shack detector. Such detection is l0 disclosed, e.g., in D. M. Alloin and J. M. Mariotti, eds., Adaptave Optics for AsWonomy, Dordrccht: Kluwer Academic Publishers, 1994. More xeeently, such a technique lias been used to detect wavefront aberrations in the human eye for such purposes as intraocular surgery and contact lens fabrication. Such detection is disclosed, e.g., in Liang et al, "Objective measurement of wave aberrations of the is human eye with the user of a Hartmann-Shack wave-front sensor," Journal of the Optical Society ofArnericc~, Vol. 11, No. 7, July, 1994, pp. 1-9, the disclosure of which is hereby incorporated by reference in its entirety into the present disclosure.
That technique will be summarized with reference to Fig. 1. A beam of light from a laser diode or other light source is directed toward the pupil and is incident on 2o the retina. Because the retina is highly absorbing, a beam on the order of four orders of magnitude dimmer than the original beam is reflected by the retina and emerges from the pupil, l~pieally, the incoming and emergent light follov~~ a common optical path; the incoming Iight is brought into the common optical path with a beamsplitter.
The emergent beam is applied to a Hartmann-Shack sensor to detect the 25 aberrations. Such a detector includes an array of lenslets that break up the light into an away of spots and focus the spots onto a charge-coupled detector or other two-dimensional light detector. leach spot is located to determine its displacement D from SUBSTITUTE SHEET (RULE 26~

WO 01/28409 PCTlUS00129078 the position which it would occupy in the absence of wavefront aberrations, and the displacements of the spots allow reconstruction of the wavefront and thus detection of the aberrations tiuough known mathematical techniques.
Improvements to the technique of Liang et al are taught in J. Liang and D. R.
Williams, "Aberrations and retinal image quality of the normal human eye,"
.Ioisrnal of tlae Optical Society of America, Vol. 4, No. 11, November, 1997, pp. 2873-and in U.S. Patent No. 5,777,719 to yi'illiants et al. The disclosures of that article and of that patent are hereby incorporated by reference in its entirety into the present disclosure. Williams et al teaches techniques for detecting aberrations and for using 1o the aberrations thus detected for eye surgery and the fabrication of intraocular and contact lenses. Moreover, the techniques of those references, unlike that of the Liang et al 1994 article, lend themselves to automation. German published patent application No. DE 42 22 395 A1 teaches a further variation using poiarizing optics to control back-scatter from the lenses in the detector setup.
~s Analysis of the eye presents unique problems and issues not necessarily faced in astronomy. . For example, while wavefront sensor systems in astronomy exhibit uniform intensity across their entrance pupil, this is not the case with systems for the aye. The eye, unlike a telescope, is subject to the Stiles-Crawford effect.
That effect is a directional sensitivity of the retina, one manifestation of which is an intensity 2o variation across the pupil of the eye when light is reflected from the retina. It exists because illuminated cones radiate more light back toward the center of the pupil than toward the pupil margin. Also, unlike astronomy, stray light from other sources, such as from corneal reflection, can be introduced into the wavefront sensor from the eye, and such stray light can interfere with the measurement of spot placement.
SUBSTITUTE SHEET (RULE 26) WO O1/28~09 PCT/USU0/29078 Other problems unique to the eye have been entibuirtered, For example, a subset of the spots that should be formed in the Harlmann-Shack detector cannot be seen, either because the aberrations are unusually large (e.g., a huge aberration caused by a scar or th.e like can displace or deform the spot so much that the spot's s origin cannot be determined or the spot leaves the field of view of the detector altogether), or they are occluded by opacities in the eyes optics or by the pupil. In current wavefront sensors, the loss of any spots frustrates the computation of the wave aberration.
Another problem is that of variable pupil size, as opposed to the .fixed pupil of 1 o a telescope.
Moreover, there is the issue ofreal-time operation. Real-time wavefront sensors have been demonstrated in astronomy, but where operation is required at rates typically greater than 300 Iiz with closed loop bandwidths greater than 30I-Iz. The atmosphere is much too turbulent for real-time compensation at slower rates.
On the i5 other hand, present adaptive optics techniques for the eye operate at a very slow rate, less than 0.25 Iiz, and do not automatically compute the wave aberration, even with single exposures. Real-time operation is not possible because of the factors described above. Also, these techniques employ relatively long focal length lerrslet arrays.
Such instruments have high sensitivity to small changes in the slope of the wave 2o aberration at the expense of dynamic range and robustness. Individual spots in the wavefront sensor image often overlap, particularly near the edge of the pupil, making it difficult for automatic centroid spot computation. Such emblems can develop especially for a commercial instrument in which operator intervention should be minimized. As a result, these systems cannot measure the wave aberration in a large 25 fraction of human eyes. An optimi~,ed wavefront sensor fox the eye should therefore SUBSTITUTE SHEET (RULE 26) robustness. lndividual spots in the wavefront sensor imago often overlap, particularly near the edge of the pupil, making it di~cult for automatic centroid spot computation. Such problems can develop especially for a commercial instrument in which operator intervention should be minimized. As a result, these systems cannot measw~e the wave aberration is a largo fraction of human cyes. An optimized wavefcont sensor for the eye should thenforc properly balance sensitivity and dynamic range, operate in real-time, and be capable of use with a significant fraction of eyes.
The measurement sensitivity of a wavefront sensor for the eye is determined pr;marily by the focal length of the lenslet array. The smallest measurable wavofroiat slope is proportional to the focal length. Relatively long focal length (e.g., 97 mm) leaslet arrays used in high sensitivity wavefront sensors for measuring eye wave aberrations typically show a stroll mean standard deviation of ~speated measurements across the pupil of an artificial eye, for example, AJ487 (at 632.8 nm, the helium-neon laser wavelaegth) for a 3.4 mm pupil.
Moreover, the eye can exhibit a severe wave aberration at the edge of the pupil due to smeared or scarred areas of the eye's tissue. Thus, such wavefiront sensors exlribit more sensitivity than necessary and require an algorithm to huntllocate the migrating spots.
Another challenge to zeal-time wavefront sensing in the eye is the spadal homogeneity of the spot of light imaged on the retina. Inhomogeneity, caused, for example, by laser speckle or reflectance variations in the underlying retina, disrupts the accuracy of the spot localization. This problari is exacerbated with the short exposures required for real-time operation.
A Hartrnann sensing technique is described in Ash et.al, "Optimal Hartrnann sensing at low light levels," Opfica Contmurrlcations, Vol. 156, 1 November 1998, pages lfl-15.
However, that artick is cvncemed a3ttt the effect of atmospheric turbulence on ground-based tsiescopcs and with situations in which it may not be feasible to illuminate the object or sky OOQ687.0015713i1 AMENDED SHEET
Emafan~a~r~~ ~~.v,~.

artificially, neither of which is a consideration is the measurement of wavefront aberratio~as in a patient's< eye. Further, the article does not even hint at a technique fox overcoming the Stiles-Crawford effect, which, as noted herein, is an important consideration for the eye, nor does it overcome the herein-noted differences between astronomical observation and the detection of wavefront sbecrations in the tyc.
As a rosdt of the above-noted problcma ov~th wavetroat sensing in the eye, a robaat and real time x>uing technique fos the eye is not 1a10Wn in the art.

ooosea.oots~sz AMENDED SHEET
Empfanss.~ " ~~,~~,.

WO O1/28~09 PCT/US00J29078 properly balance sensitivity and dynamic range, operate i~,real-time, and be capable of use with a significant Exaction of eyes.
The measurement sensitivity of a wavefront sensor for the eye is determined primarily by the focal length of the lenslet array. The smallest measurable wavefront slope is proportional to the focal length. Relatively long focal length (e.g., 97 rnm) lenslet arrays used in high sensitivity wavefront sensors for measuring eye wave aberrations typically show a small mean standard deviation of repeated measurements across the pupil of an artificial eye, for example, x,/487 (at 632.8 nm, the helium-neon laser wavelength) for a 3.4 mm pupil. Moreover, the eye can exhibit a severe wave t0 aberration at the edge of the pupil due to smeared or scarred areas of the eye's tissue.
Thus, such wavefront sensors exhibit more sensitivity than necessary and require an algorithm to hunt/locate the migrating spots.
Another challenge to real-time wavefront sensing in the eye is the spatial homogeneity of the spot of light imaged on the retina. Inhomogeneity, caused, for example, by laser speckle or reflectance variations in the underlying retina, disnipts the accuracy of the spot localizatoin. This problem is exacerbated with the short exposures required for real-time operation.
As a result of the above-noted pmhlems with wavefront sensing in the eye, a robust and real-time sensing technique for the eye is not known in the art.
SUMMARY OF TIIE INVENTION
While many of the above problems have apparently been overcome in astronomy, it will be readily apparent from the foregoing that a need exists in the art for a wavefront sensor capable of handling the unque problems of the eye. It is a SUBSTITUTE SHEET (RULE 26) therefore a primary object of the invention to address such unique problems, which include nonuniform illumination of the pupil due to antenna properties of cones, ocular abnormalities, variable pupil size either among individuals or in the same individual under different levels of illumination, increasing severity of the wave aberration at the edge of the pupil, and spatial inhomogeneity of the retina, which produces centroiding errors.
To achieve the above and other objects, the present invention is directed to a system adapted to overcome such unique problems.
Errors introduced by nonuniform illumination are handled by locating the to spots through the following centxoiding technique. Once an image is taken, a set of initial boxes is set up on the image. Each initial box is centered around the location where a corresponding spot would be in the absence of wavefront aberration and has a side of length equal to what the inter-spot spacing would be in the absence of wavefront aberration. A first guess of the spot location is produced by finding the i5 intensity centroid of the portion ofthe image within each initial box. Then a smaller box is drawn, centered on that centroid. The smaller box can be clipped to lie within the initial box. A new centroid is found within that smaller box. That process is iterated until the box size reaches some criterion size, such as a width equal to the full width of half maximum of the diffraction-limited spot. Each step throws away data 2o remote from the centroid found in the previous step, since such data most likely contain noise or systematic errors rather than information useful in locating the spot.
There are two stages used in centroiding the spots. In the first stage, reference boxes are established based on centered reference positions or on the center of the image on the detector (i.e., "from scratch"). The latter technique makes less of an 25 assumption where boxes are located and decides how far oul to go as the process runs SUBSTITUTE SHEET (RULE 26) WO 1)1/28409 PCT/USOO129078 according to the hrnown size of the lenslet array and number of lenslets. The boxes in the latter technique can form an irregular array, which can be described as being constructed in rings about a center box with the location of the outer boxes being determined in a center-to-outward direction. The size of the boxes using either s technique can be adjusted by a parameter entered by the operator or stored in software, for example. 90% of the actual inter-lenslet spacing or 90°io of the area of a box having sides equal in length to the inter-lenslet spacing.
Once these boxes are detennined, a technique can be used that locates a centroid within an initial box, centers a smaller sized box on that centroid, followed 1o by locating the centroid again, followed by another smaller box centering, and so on until the diffraction limited box size is reached and the final centroid is determined.
Alternatively, a technique can be used that starts with an initial box, fords a centroid within that box, centers a smaller decremented box on that centroid, clips as described above, calculates the next centroid, centers a decremented box again, clips again, and 15 so on, all the while maintaining each centroid within the original box.
This process also terminates when the diffraction limited box size is reached and the final centroid is determined.
An additional centroid can be calculated on the entire array to hocate the center, which is especially used with longer focal length lenslet arrays.
Doing so 2d permits comparsation for spot migration, which is compensated for by the center of mass of the entire array. Iterativehy centroiding is to be contrasted with previous techniques such as simply using thresholding and simply by doing a center of mass of the entire box. The present invention better fends the center and reduces the effects of radiance caused by the dimmer spots of the centroids. Tlie technique according to the SUBSTITUTE SHEET (RULE 26~

W O 1) 1/28409 PCT/US00129078 present invention also reduces the effects of stray li~H't°>ie~aiise t'h8se effects are progressively eliminated.
The embodiment just described includes re-centering and removing pixel operations. In another embodiment according to the invention, the boxes can be shrunk first, then translated, and then clipped to a threshold value of intensity, in which only those pixel values above the threshold will be included in the calculation of the centroid. There can also be a variable threshold per box as the box size is reduced to account for data from different pixels.
The centroiding technique using shrinking boxes overcomes a difficulty found in centroiding without shrinking boxes, namely, errors when there are broad intensity gradients such as caused by the Stiles-Crawford effect.
Ocular abnormalities, such as scars, can result in spots deviated far from where they would be in the absence of wavefront aberration. Such spots can come close to, or even overlap, other spots. In fact, such spots can be displaced so far that they disappear from the field of view of the detector altogether. Other ocular abnonnalides, such as occlusions, can absorb light, so that no spot is produced at all.
To handle such abnormalities, the present invention provides a technique for wavefront reconstruction in the absence of certain data points. Part of the wavefront reconstruction involves manipulation of a matrix whose rows correspond to z0 displacements of the spots from their positions in the absence of aberration. For spots not producing usable data, the rows can simply be deleted from the matrix, or the values contained in such rows can be extrapolated from neighboring rows.
At the heart of this flexibility is the particular data structure, which is a matrix of Zernike mode number in the columns and centroid displacement number in the SUBSTITUTE SHEET (RULE 26) rows. The matt-~x is used to calculate the Gernike coefficients to determine the wave aberration of the eye.
Different criteria can be used for determining whether to omit a centroid or not, such as the standard deviation of the light within a box in the center of mass calculation, a position of a point being outside of a box or in a highly unexpected area within the box, or points being occluded totally by corneal defects. Then based on these omitted response or centroids which can be done on a frame by frame basis if desired, one calculates the Zcn-iike modes.
Variable pupil size can be handled either before or after data are taken. If the l0 variable pupil size is to be handled in the same individual, data can be taken twice. If that is not convenient or possible, data can be taken in a larger pupil radius, and then the centroids in a smaller radius can be located. A renonnalization procedure is used to provide a matrix to convert the wavefront reconstruction from the larger radius to the smaller one. The variable pupil size results in a variable number of spots used in wavefront r~onstruction, which can be handled by varying the number of Zemicke modes used in the reconstruction. The wave aberration is recalculated based on a new radius using software as opposed to changing the actual excluded centroids.
Starting with a valid number of data points, the software can determine a number of Zernike orders, and correspondingly a maximum number of modes that can be accurately calculated. For example, order 0 has mode 1, order 1 has modes 2 and 3, order 2 has modes 4, 5, and 6, etc. Generally, if one calculates particular modes within an order, it is desirable to have all the modes within that order.
Therefore, one would not calculate modes 7 and 8, but not 9 and 14 since those are all within the third order Zemike equations. A general rule of thumb for calculatilig the order is (modes + 1)(modes + 2)/2. Based on this equation, one sees that beginning with the SUBSTITUTE SHEET (RULE 26) WO Ill/28~09 PCT/US00/29078 number of available centroids, that number of centroids is divided by two, yielding the maximum calculable modes.
One can determine the highest order that can be accurately calculated based upon the number of lenslets divided by 2; yielding the number of calculable modes.
s Then, the highest complete number of modes translates to a particular order.
For example, 20 centroids yields 10 Zernike modes, which allows one to accurately compute the third order Zemike equations.
The increasing severity of the wavefront aberration at the edge of the pupil can be dealt with as just described.
Spatial inhomogeneity of the retina can be handled through the following techniques. A light source of short coherence length reduces the problem of speckle caused by interference in light of high coherence. In addition, the inhomogeneity can be averaged out by scanning the illuminating light across a short distance of the retina and de-storming the emergent light.
The above feaW rcs allow a wavefront sensor accorduig to the present invention to provide fast, robust and accurate centroiding, e.g., up to 50 Hz.
BRIEF DESCRIPTION OF THE DRAWINGS
A better understanding of the present invention can be obtained when the zG following detailed description of the preferred embodiment is considered in conjunction with the following drawings, in which:
Fig. 1 shows the optics of a Hartmann-Shack detector;
Fig. 2 is a diagram of a wavefiont sensor in accordance with a preferred embodiment of the present invention;
2s Fig. 3 is a block diagram of a computer of the wavefront sensor of Fig. 2;

SUBSTITUTE SHEET (RULE 26) WO 01/28409 PCTlUS00I29078 Fig. 4 shows plots, for three subjects, of measurements of the amount of coma, astigmatism, and spherical aberration when the subject changed his/her accommodation smoothly from one state to another;
Fig. 5 illustrates modulation transfer functions computed from measurements of the three subjects' wave aberrations;
Fig. 6 shows images obtained with and without scanning in the wavefrnnt sensor ofFig. 2;
Fig. 7 illustrates the effect of scanning in the wavefront sensor of Fig. 2;
Figs. 8A and 8B illustrate hypothetical arrays used in calculating Zenlike to coefficients;
Fig. 9 illustrates a circular region used to capture centroid data within corresponding boxes and a procedure for using the centroid data when the size of the circular area is reduced;
Figs. 10-12 illustrate a method of centroiding in accordance with an embodiment of the invention.
Fig. 13 is a flowchart diagram of a method in accordance with an embodiment of the invention;
Fig. 14 is a flowchart diagram of a module of the method of F lg. 13;
Fig. 15 is another flowchart diagram of a module of the method of Fig. 13;
2o Fig. 16 is another flowchart diagram of a module of the method of Fig. 13;
and Fig. 17 is another flowchart diagram of a module of the method of Fig. 13.
Fig. 18 shows time averaged Strehl ratio versus bandwidth of a perfect adaptive optical system for natural accommodation with far target and 5.8 mm pupil.

SUBSTITUTE SHEET (RULE 26) DETAILED DESCRIPTION
A preferred embodiment and variations thereof will now be set forth in detail with reference to the drawings, in which like reference numerals refer to Like components throughout. The detailed description will be organized thus. First, the solutions set forth in the summary will be described in greater detail in turn. Then, the hardware, software and applications of the preferred embodiment will be set forth in that order. Reference will be made throughout to Fig. 2, which shows a wavefront sensor according to the preferred embodiment, and to Fig. 3, which shows a computer for use with the wavefront sensor.
io Shrinkinu Box Centroidin~
As is known in the art, the wavefront reconstruction involves determining the displacement of each spot from its ideal location, which is the location that it would occupy in the absence of wavefront aberrations. The spots are located by centroiding, namely, finding the centroid in the image corresponding to each spot. Tlie centroid is an average location weighted by intensity and is analogous to a center of gravity. In other words, the centroid is a mathematical stand-in for the center of the spot.
Mathematically, if an area ~f of the image has an intensity distribution I(~, y), the rectangular coordinates of the centroid within that area are given by !~(x~ Y)~'dY ,~YI (x~ Y)~'dY
~x~Y~= '' - a JI(x.Y)~dY ~ jl(x.Y)~dY
A A
zo The preferred embodiment uses an iterative centroiding technique to provide robust and accurate centroid location. In one variant, the image is divided into an stray of initial boxes, each centered on the ideal location of the spot and having a linear dimension equal to the spacing between the ideal locations of the spots, much SUBSTITUTE SHEET (RULE 26) like the array shown in Figures 8A and 8B. If the quality of the optics of the eye is high enough or the wavefront sensor is made insensitive through use of a short focal length lenslet array, it is likely that each actual image spot from the detector of the camera 20 will fall within a corresponding box of this hypotlietical sduare reference array. F ig. 12 shows such an initial box 100, which in this example measures pixels by 10 pixels.
In another variant shown in Fig. 10, a different process occurs for generating inital boxes. The array of initial boxes generated may or may riot wind up being a square array. Instead, it may be an irregular array. In Fig. 10, a first box 70 is to identified about a corresponding center point 72. The center point 72 corresponds to an ideal image spot position (i.e., the that a spot would occupy in the absence of wavefront aberrations) at the center of the image of the detector camera 20.
From the box 70, all the other reference boxes are generated as fbl lows. With knowledge of the total number of lenslets in the lenslet array, and the magnitude of the lenslet spacing "C", a second box 74 is identified about its own center point 76 spaced one inter-lenslet spacing from the first box 70 along a principal direction (e.g., an "X"
direction). The X principal direction corresponds to one of the directions along which the array of ideal center points is defined and corresponds to one of the directions in which the actual lenslets 18 are laid out. Tn like fashion, another principal direction (e.g., a "Y" direction), orthogonal to the X direction, also corresponds to a direction along which the ideal center points of the lenslet array are located on the detector of the camera 20. Along the Y direction, another box 78, having its center point 80, is identified at the inter-lenslet spacing from the first box 70. Next, in order to identify another box 82 along a diagonal in between the boxes 74 and 78, a centroid,is preliminarily calculated for each of the boxes 74 and 78. As SUBSTITUTE SHEET (RULE 26) WO 01/28.109 PCT/US00129078 ~~ ~JIIyLUU 1 shown in Figure 10, centroids 84 and 86 are found thi~~r~$"po o ese oxes, both of which are located at less than the inter-lenslet spacing from the center point 72 of the box 70 in this example. It is quite possible that these preliminary centroids could be located further removed from the center 72 than the inter-lenslet spacing C, but Figure 10 is merely illustrative of the technique. Because these centroids are located less than the inter-lenslet spacing from the center 72, a reasonable estimate of where a box 82 should be placed is at a location (x1, y1) that corresponds to the X
position of the ccntroid 84 box 74 and the Y position of the cenhoid 8G of the box 78.
Thus, (xu Yi) is the position determined fbr a center 84 of the box 82, as shown in to Figure 10. The array is completed by determining successive "rings"
(actually squares) of boxes, each of which can be more or less than the inter-lenslet spacing distance from the previous ring. If the spacing in the principal directions, found in the previous ring, is more or less than the into-lcnslet distance, that spacing will be used to End the next set of four boxes in the principal directions for the next ring out from the center. The result is a reference array of boxes that could be irregular or square, and is generated from an initial center box and point with the other reference boxes generated in outer rings, although other methods could be employed. This technique is termed "from scratch."
The actual size of the boxes generated according to either of the techniques just described can be reduced by a factor under control of the operator using the software 13 or programmed into the software 13. For example, the size of a side of any or all the individual boxes may be scaled from the actual inter-lenslet array spacing C or lenslet size (e.g., the box side may only be 90% of the actual inter-lenslet spacing C).

SUBSTITUTE SHEET (RULE 26) Once the positions of the initial boxes are determined, v6lietlWr'a square array ' or an irregular array as described about, the process continues to the second stage.
Refen-ing now to Fig. 11, a technique is described in which the centroids for all spots corresponding to the image on the detector of the camera 20 are determined in s iterative fashion. In Fig. 11, one of the initial boxes (i.e., like one of the boxes in Figures 8A and 8B or one of the boxes in Fig. 10) is shown as a box 86 within which an actual spot 88 on the detector of the camera 20 has been located by the software 13. As a fast step in the iteration, within the box 86, shown to be of l OxlO
pixels in area (although other areas could suffice), a centroid 90 is calculated for the spot 88 in 1o accordance with the equations given above. The centroid 90 likely is not at the actual physical center (not shown) of the box 86. In the next step, the software 13 removes a pixel width and height from a new box 92, which becomes, for example, 9x9 in pixel area, and centers the box 92 about the just calculated centroid 90. The process is repeated within the box 92 with a new' centroid 94 being calculated for the spot 88, 15 and then re-centering with a next smaller box centered on the previously calculated eentroid and a new centroid calculated for that box and so on. For each progressively smaller box, the centroid is calculated only with respect to image data taken within that box. Once an approximate location of a centroid is determined, it is safe to assume that image data remote from that approximate location represent noise or 2o systematic error such as a ramp from the Stiles-Crawford effect rather than useful information about the location of the spot. Thus, as the boxes become progressively smaller, progressively more noise is excluded. The process terminates when a centroid 96 is calculated for a box (not shown) and the box size is decremented to produce a box 98 that is at or below the diffraction limited box size determined from 25 an optical analysis of the wavefront sensor 10. Once the diffraction limited box size i~
SUBSTITUTE SHEET RULE 26) is reached, the previously calculated centroid 96 is the final centroid determined for the spot 88 and the one whose displacement from its ideal reference position is used in calculating the wave aberration for the eye. Some other criterion for terminating iteration might be useful besides reaching diffraction-limited box size.
Fig. 12 illustrates, for the second stage above, an alternative iterative procedure for locating the centroids of the spots on the detector of the camera 20 within the reference boxes. A diffraction limited box size is again used to determine the final centroid for calculating the wave aberration of the eye. In Fig. 12, a box 100 of 10x10 pixel area (although other areas could suffice) is shown that corresponds to one of the reference boxes generated through either of the techniques described above for the first stage. The box I00 is shown to include the spot 88 discussed above. A
centroid 102 is calculated for the spot 88 within the box 100 and, as shown in Fig. 12, a new box 104 that is decrcmcnted in height and width by 1 pixel (e.g., 9x9 pixels) then is centered on the centroid 102. This particular technique then clips the box 104 to within the box 100, thereby forming a smaller box 106 for the next iteration of centroid location in order to keep the smaller box 106 and thus the new centroid found in the box 106 within the initial reference box 100. The smaller box 106 (e.g., 4x5 pixels) is formed by portions of the sides of the bax 100 and portions of the sides of the box 104, as shown in Fig. 12. Within the box 106, a new centroid 108 is calculated, which may or rnay not be at the same location of the previously calculated centroid 102. Next, a square box I 10 of decremented size (e.g., 8 x 8 pixels) is centered on the new centroid 108, and the process is repeated by clipping the new box 110 to lie within the box 100 (also to lie within the box 106) again to form a new smaller box 112 (e.g., 3x4 pixel area) in which a new centroid 114 is again calculated, which may or may not be at the same location as the centroid 108. The process SUBSTITUTE SHEET (RULE 26) continues repetitively by centering with another smal~er~b'oX~ollowed by~centroiding, and so on. Finally, a centroid 116 is calculated within a clipped box just prior to raaching the diffraction limited box size and a square box is centered on the centroid 116, which is clipped to a new box 118. If the box 118 is at or smaller than the s diffraction limited box size, then the previously calculated centroid 116 is retained and its displacement from its ideal reference on the detector of the camera 20 is used for calculating the wave aberration of the eye. In this manner, the centroids that are calculated'in the process are always maintained within the area of the iilitial box 100.
It should be understood that a variety of other centroiding techniques could be used in to which boxes are sequentially reduced in size and new centroids are calculated iteratively. These other techniques are included within the scope and spirit of the present invention.
Still another variation can be used. Since two consecutive whole numbers are never both odd or both even, two boxes whose dimensions vary by one pixel in each 15 direction cannot have the same center if they consist of whole pixels. As a consequence, asymmetries can easily appear in the spot tails. To avoid that consequence, the centroiding procedure can be adapted to work with fractional pixels.
For example, instead of reducing the box size by shaving off one row of pixels and one column of pixels, the technique using fractional pixels can shave off.one-half of a 20 row or colmun of pixels from each side. T'he contribution of each fractional pixel to the centroid is considered to be the corresponding fraction of the intensity centered in the fraction of the pixel which is used. This technique has been found to produce results ofhigh accuracy in a computationally efficient manner.
The software has a user interface on which the calculated centroids are 25 displayed to the operator. The operator can drag tl~e centroids with a mouse. When SUBSTITUTE SHEET (RULE 26) an image is acquired, the software automatically computes the centroids in the image.
The centroids have a number associated with them, for example, 1-37. Errors in this computation can occur and the user can direct them manually by dragging valid centroids to appropriate location. The software will re-compute a centroid based on the local light distribution around the location where the dragged spot was left if such a feature is enabled. If a center centroid is dragged to a new center location, the entire set of centroids will be recomputed based on the new center, if such a feature is enabled.
The centroids can be computed in three ways, any of which call be selected by to the user. Under some conditions, these alternatives may reduce the need to manually correct the centroid computation. The first alternative is from scratch; which can be a default, The second option is to use the last computed centroid location as the start estimate for finding the ccniroids on a new image. The third option uses the stored locations of the reference image that would have been obtained by a perfect eye as the starting point for locating the centroids.
Threshold clipping can occasionally aid in the accuracy of the automated centroid-fitting algorithm. Selecting the threshold, for example, by selecting a "Set Threshold" menu command, brings up a dialogue box, that allows the operator to enter a lower and an upper threshold of image intensities valid fur analysis.
These 2o values are used for all threshold clipping until another value is entered.
A current frame can be threshold clipped and threshold clipping on subsequent frames can be enabled, for example, by using a tlireshold clip. This feature may be disabled for each search box or for the entire image.
~Iandlin~ of Missing bate SUBSTITUTE SHEET (RULE 26) The wave aberration can be reconstructed when any number of wavefnont centroid spots is missing or is of insufficient quality and should be eliminated from the data collected by the camera 20. This could be for a number of reasons, for example, the operator believes the data point is an invalid data point (e.g., two spots in one box). Further, this could be automatically done such that if the data point includes too great a standard deviation from a center, or the data point is too dispersed and the standard deviation of the intensity of the pixels is too great.
Further, the centroid may simply be too dim to be recogluzed, or the staaidard deviation of the center of mass exceeds the predetermined threshold.
1o That centroids can be dynamically selected for or excluded from use differs from astronomy applications. The substantially higher Iight levels available in the ophthalmic applications allow areas of scarring, opacity, or otherwise unreliable data, to easily be excluded from the calculation. Any application of wavefront sensing for the eye, including, for example, autorefraction, design of improved contact or intraocular lenses, refractive surgery or retinal imaging with adaptive optics, will benefit from being able to compute the wave aberration when spots are missing or are of insufficient quality.
Dynamic centroid selection is facilitated using the matrix implementation.
The following simplified illustrative example is offered. A hypothetical 3x3 array 150 of reference spots and boxes 152 is shown in Fig. 8A. Each box of the array of reference spots 150 and boxes 152 arc designated by an index 1-9 from top to bottom and left to right, although another index ordering scheme could be used. For each reference spot 150 determined by the software 13, there is a centroid I54 displaced in x and y from the corresponding aperture center 15G that also is determined by the 2s software 13 within each box, as will be explained below in more detail. The software SUBSTITUTE SHEET (RULE 26) WO l11/28409 PCT/US00/29078 13 determines the x and y displacements of each centroid 154, designated dx;
and dy;, respectively. Thus, the s vector has 18 components, 9 for x and 9 for y, as shown in the matrix equation below. The G matrix is an I8 centroid (or aperture center) by 5 modes matrix (i.e., 18x5) and a is the vector of the Lenuke coefficients consisting of the 5 components corresponding to first and second order aberrations (no piston term, as explained below). Thus, s = Ga, or in other words, S[18] G[18x5] a[5]
~dx~
dxz dx3 dx4 dxs dx6 az a3 dh~ -[ G ~ a4 dy, a$
dYz a dY3 dY4 dY5 dYs to LdY9.
The preferred embodiment allows the elimination of centroid data this using the flexible matrix data structure above. Basically, a portion corresponding to the unreliable data can be eliminated from the matrix, the inverse taken, and new Zernike coefficients calculated. The G matrix is truncated with rows eliminated, depending on the number of Zernike modes to calculate and the number of valid data points one SUBSTITUTE SHEET (RULE 26) WO 01/28409 PCTlUS00/29078 wishes to examine. For every centroid missing or removed by the operator, two rows are removed from the s vector (corresponding to the x and y displacements of that ccntroid) and the G matrix (corresponding to an aperture center) to form an s' vector and a G' matrix, respectively. An example of the s' vector and the G' matrix is shown below again for a hypothetical 3x3 array 160 of lenslets, but with an index "2"
element 166 removed as indicated by the X in Fig. 8B. In other words, for every centroid removed, one dx; (i =1 to 9) component (e.g., dx2) and one dy;
component (e.g., dy2) are removed to form the 16-element s' vector. The corresponding row integral (apezture center) elements of G simply are removed also to form the 16x5 G' t o matrix given by 1 X~ Hb Y; tb aZ , ~J ' 4b ~ ~ j ax ~rW Yr"b Y - 1 Y~+b x,+b aZ~
G'~ q,b7 J f ay dx dy Ymb xr-b (rows 2 and 11 of matrix G are removed) with a still having 5 components (i.e., Zernike coefficients 2 to 6). In the above equations, b is one-half the length of the side of a box. Generally, one should include at least twice the number of eentroids as the number of Zemike modes to be calculated. Thus, for nine centroids, one would generally wish to calculate four to five Zemike modes.
s [16] G'[16 x S] a[5]

SUBSTITUTE SHEET (RULE 26) ~dx, dx, ~4 dX5 -' a3 dx9 i ~~ G a4 ~

dy, a dY3 5 a dY~ asp dYs dYs dy9 Tluning to Fig. 9, if data are to be eliminated, rows are deleted, designated as 186, and once the appropriate G' matrix and s' vector are determined, designated as 188, the a vector must be solved for the corresponding G' and s' as above. The a vector remains the same to determine the wave aberration for the same number of modes as before, but with the displacement data removed.
For a given number of lenslets, the maximum number of Zernike modes that can be used for a reasonably accurate fit is determined. For a given pupil size, which 1o depends on the individual subject bcillg tested, the number of lenslets will be determined. For a given number of lenslets, it has been found that a reasonably accurate wave aberration can be computed with about half this number of zernike modes. For example, with 37 lenslets, au accurate wave aberration representation can be obtained for approximately fifth order Zemike polynomials consisting of 21 Zernike modes. T'he above technique will work as long as the number of centroids (lenslets) removed are not too many for the number of Zernike modes to be SUBSTITUTE SHEET (RULE 26) WO Ill/28~09 PCTlUS00/29078 determined for the a vector. In other words, as the number of lenslets used decreases, it may become necessary to reduce the munber of Zernike modes computed in order to follow the above rule of thumb. If too many centroids are eliminated, then the ability to accurately determine the wave aben-ation for particular Zernike modes is diminished Another way of stating this is that i.f too many centroids are eliminated, then the G matrix becomes too "singular" and the SV17 will be inaccurate.
Following the rule of thumb that twice as mmy data points should be available as number of Zernike rows calculated, in general, one wishes to use four times the number of rows as columns in the G matrix. Therefore, the results given in the two examples above to would be questionable because the number of rows does not exceed the number of columns by a factor of four.
The G' (or G) matrix, which, as discussed above, has elements that are the derivatives of the basis functions averaged over the sub-aperture of each Ienslet 18 in the x and y directions, also becomes representative, as a metric, of the number of modes that can be calculated using the SV'D technique, in accordance with an embodiment of the invention. The number of entries in the G' matrix provides an indication of the minimum number of displaced centroids (lenslets) required for an accurate calculation of the wave aberration for particular Zernil{e modes and Zernike order. The preferred embodiment thus not only allows acquisition and analysis of 2o wave aberration data even in subjects who have severely aberrated regions or local occlusions in their pupils, but also provides a metric for accuracy of the calculated wave aberration.
The Zernike coefficients are determined using a data structure matrix, as described above. The rows of the matrix correspond to the Zernike mode number and the colLUnns correspond to the centroid numbers. The matrix multiplied by a vector of SUBSTITUTE SHEET (RULE 26) WO 1)1/28409 PCT/USUO/29078 the Zeznike coefficients is equal to a vector of the displacements of each eentroid from its ideal position. The vector of the displacements consists of the displacements in x followed by displacements in y, each subscripted x,y pair of components of the matrix having the same subscript as and corresponding to a particular lenslet.
The size of the Zernike coefficient vector is determined by the maximum Zernike mode number_ Any of the rows or columns can be eliminated in the calculation depending on the number of centroids and Zernike modes. respectively, required. Once the image data is collected and analyzed, upon elimination of a row or column, the 2emike coefficients can be recomputed on the fly. The software 13 can show where any missing data are located in this matrix. The missing data can be replaced by proxy data consisting of previously taken data or data mimiclting another one of the elements of the matrix. Other decision algorithms in the software 13 software can be included to modify the data structure matrix.
Changg in Radius of Pupil The above discussion concerns calculating the wave aberration for centroids within a pupil radius R~, in which the data are acquired. From there, if one wanted to calculate the Zemike modes for a different pupil radius, for example, to account for differences between day and night vision or to exclude large aberrations at the pupil edge, one would preferably do more than simply ignore data from the centroids lying outside the smaller radius. Instead, what is done is to select a new radius, called Rl (see Fig. 9), that captures a lesser number of centroids. The data, however, fiom the centroids calculated for the radius Ro is not discarded. Instead, it is simply that the Zernilce polynomials exterior to Rl are effectively clipped by using a transformation from the Zemike modes for 1~ ovex to the Zernike modes for Ri with renormalization, z5 designated as 184 in Fig. 9. The transformation equation is represented as {Z;}RO ~

SUBSTITUTE SHEET (RULE 26) {Z;'}~. An explanation of the rationale for its validity, and the means by which ii is implemented in accordance with an embodiment of the invention, follows. A set of Zemilke coefficients measured over a standard pupil size is renormalized to a set of coefficients measured over a smaller, concentric pupil. As was mentioned above, 5 Zemike polynomials are a set of two variate orthogonal polynomials defined over the unit-circle. To formulate the procedure used to do so, the wave aberration function is first expressed as N l f6x~r~8~=~,br?t~r~~~~ ~pl i.2 where c~R represents a measured wave aberration of a pupil o f radius R (e.g., Ro) and 1o Zr represents the ith Zemikc polynomial as a fwction of polar unit-circle coordinates.
The value 1L~ is the highest mode that is measured in the wavefront reconstruction.
Assume that the b coefficients (which can be expressed as a vector b) are calculated by known methods, but that it is desired that a set of coefficients be determined for the function N
1S ~R,~YA, B)= ~ h~2, ~rA, 9~, I r' _< 1 , ;~z which is simply e~R evaluated over a smaller, concentric pupil with radius R' (e.g., R,), where A = (R %R) < 1. 'This function can be thought of as tacking a concentric circular "slice" out of the middle of ~R
However, the equation just given is not a valid expression for the Zernike 2o representation of ~R- because such a representation must be expressed in a radial variable that ranges over the whole unit circle. Therefore, the goal is to find the coefficient vector a in the Zernike representation of c~R ~, 2.~
SUBSTITUTE SHEET (RULE 26) WO 111128.109 PCT/US00129U7ts ~' ~r~ B)= ~Ci~iUf B
!=2 in teens of the vector b and the radius ratio, .,4. In order to determine the vector c, a standard procedure is used for expressing any function defined over the uuit-circle iu terms of Zernike modes: integrate the product of the function with each Zernike polynomial over the unit-circle to find the correspazrding coefficient, In particular, for ~R ~, i~r t c'=~ j ~ ~",(rA,B)Z~(r,B)sdrdB.

The above equations allow one to derive the following:
N 1 a~ i c;=~b~- f j Z~(YA,t~Z;(r,B)rdrdB.
-z ~ 0 0 The equation just given can be expressed as the matrix equation c = TAh with T~ being a sparse mahix wliose entries only depend on A. The entries of TA can be computed using the symbolic integration tools of Matliematica available from Wolfram Research, Alternatively, they can be computed when the lenslet objects are created in the software in accordance with the new pupil size. The tatter technique allows "on the fly" calculation, since the matrices are computed "ae you go." 'fhe results of this integration shows that each c, term is a weighted sum of selected b~ terms each of which has both an order with the same parity as and is greater than or equal to the order of the c; term in question. The weights are polynomials in A whose degree is equal to the order of the corresponding bJ term. The sparseness of the 'fA is a zesult of the fact that the double integzal is non-zero only when the azimuths! angle functions of the Gernike polynomials agree.
This can happen with only one mode per order and only if the order has the same parity as the order of the original c. Perm.
It should be noted that this expression for c, is also dependent upon the number of modrs that are used in the original expression of ~R. Thearetically, all orders of ~X are involved in the expression for c, but the expression that is actually used will only involve those modes of ~R that are actually measured. Mrerefore, this method can produce a wave aberration over a smaller pupil without the need to make additions: measurements over that smaller pupil, but it can be no more accwate than the original maa~rrement.
SUBSTITUTE SHEET (RULE 26) WO 01/28409 PCTIUSUUl2vo'rs The advantage of the above method is that all of the data are still employed.
That is, one maintains the information of the wavefront inherent in the Zernike modes. However, the Zernike modes are for a smaller calculated radius instead of being calculated by just clipping the data, which could lead to less accurate results.
In this software calculation of the smaller radius R', it is necessary that the radius be smaller than the initial radius R. If it were desirable to capture a greater number of points from a larger lenslet array, for example, or to expand the radius past R.o, one would return to the beginning and select a larger Ro, thereby passing more data from the image captured by the camera 20 or recapturing more data. By to expanding Ro to a larger area, perhaps one would capture more Ienslets. Of course, the Zernike modes for a smaller radius could again be calculated as described above from the data initially captured for the larger radius.
The number of Zernike modes calculable in general is limited to the number of centroids captured. As a rule of thumb, one prefers to have at least twice as many 1 s centroids as the number of Zernike modes calculated.
In the user interface for this feature, the following options are available: a choice of pupil diameter which allows the data analysis to proceed for smaller pupil diameters than a maximum value, for example, 6 mm; a choice of Zemike mode range, which allows a choice to be made for a minimum and a maximum Zernike 2o mode to use in computing the wave aberration. 'The wave aberration is reconstructed from the selected range of Zernike coefficients and for the selected pupil size. The wave aberration can be displayed, for example, as a contour plot with values of plus or minus 0.15 microns (7J8 at helium-neon wavelength). The point spread function can be calculated from the wave aberration and is computed by the software 13 as SUBSTITUTE SHEET (RULE 26) wV U1/lts;uy P(:TllJS00/29U78 squared amplitude of the Fourier transform of the generalized pupil function of the displayed wave aberration function.
Inhomoo~eneitv and Laser Speckle As noted above, spatial inhomogeneity~ in the retina can cause inaccurate measurements. To overcome that inhomogeneity, and more specifically to average out its effects, a scatuiing and de-scanning technique is used.
To implement the scanning and de-scanning, Fig. 2 illustrates a wavefront sensor 10. In order to reduce the problem with short exposures, the wavefront sensor includes a scatminglde-scanning mirror 60 conjugate to the pupil plane 62, and a to pair of lenses 64 and 66. The mirror GO eliminates spatial noise by scanning a point source in a line on the retina, as will be appreciated by those skilled in the art. The angular extent of the scan can be less than 0.5 degrees at a temporal frequency of 400-600 Hz. The angular extent of the scan is kept smaller than the isoplanatic patch size of the eye 58. The temporal frequency is set high enough that there are many scans Is within a single frame ofthe detector of the camera 20. By having the light return from the eye 58 through the same scanning optics, it is de-scanned so that there is no movement of the image of the pupil on the lenslct array 18. Thus, spatial noise can be removed without disrupting the wave aberration measurement. Figs. 6 and 7 compare results obtained with the wavefront sensor 10 with and without scanningJde-scanning.
2o Scanning is an effective method to remove spatial noise when short exposures of the wavefront sensor 10 are required. This benefit will accrue naturally in a confocal laser ophthalinoscope equipped with adaptive optics. In that case, the same scanning beam that is used for imaging can be used for low noise wavefront sensing.
Scanning is not recommended in a commercial instrument designed only to measure 2S the static component of the wave aberration, provided it is possible to either average SUBSTITUTE SHEET (RULE 26) WO 01/28409 PCT/USUUI2v~r~a the results from a number of short exposures. or cohlect'light over a single long exposuxe. In either case, the reconunended total exposure duration is at least seconds for eye movements to eliminate spatial noise on their oven.
Laser speckle is known in various arts, including holography, and is caused by intereference among various components of a highly coherent light beam. .As discussed above, laser speckle degrades the quality of wavefront sensor spots when a coherent laser (usually either a helium-neon or diode laser) is used as the source 16.
Conventional laser diodes, when driven above lasing threshold, arc coherent sources with spectral half=uddths on the order of only a few nanometers. They therefore have very long coherence lengths so that reflections from many surfaces in the eye and in the wavefront sensor 10 can interfere with each other,, producing speckle. To reduce speckle, an SLD having a broad spectral half width (on the order of 20 nrn) and a short coherence length of 30 microns can be used. The coherence length of the SLD
is nearly independent of the diode's driver current or output power.
is The use of longer wavelength, near infrared sources has at least two advantages. First, they axe less dazzling for the patient. Second, the maximum permissible exposure is increased so more light can be used while still ensuring patient safety. Infrared light allows the pupil to dilate so that a larger portion of the wave aberration can be measured. Examples of possible Iiglit sources for the light source 16 include a 70 nm laser diode by Sharp, a 780 nm SLD by Anritsu, and an 844 nm SLD by Hamamatsu. The 780 nm Anritsu SLD provides clean, tightly focused spots with the least amount of speckle over a wide range of powers and exposure times. .The laser diodes perform similarly. The 780 nm SLD is preferable because it typically produces more compact, relatively speckle-free spots in images for a range of different exposure. times and powers than the other 840 nm SLD
or the SUBSTITUTE SHEET {RULE 26) WO 01/28:109 PCT/US00I29078 750 nm laser diode. It is also suffcialtly far into the infrared to be comfortable for the patient. A larger number of spots are visible with the 780 nm diode than with the 750 nm diode due to the increase in pupil diameter that results from the reduced brightness of the 780 nm diode.
Laser diode spot images always contain a larger amount of speckle than the SLD. It would be difficult to obtain repeatable measurements of the centroid locations using the laser diode, especially at higher powers when the coherence length is longest. When driven under losing threshold, shorter coherence lengths are possible. However, the output power obtained under threshold may be too low to 1o yield an adequate amount of light reflected from the retina. It may be possible to undemm curxent lasw diodes if Long exposure times are used in conjunction with a detector (e.g., CCD) of the camera 20 that integrates throughout the exposure.
Real-time Wavefront Sensor Hardware A real-time ophthalmic wavefront sensor 10 used to collect and analyze image 1s data from a subject's eye is illustrated in Fig. 2, in accordance with an embodiment of the invention. The wavefront sensor lU includes a computer 12, for example, a personal computer (PC), and control data acquisition and analysis software 13, as illustrated in Fig. 3. The software Z3 can be stored in. the PC 12 in memory 14, which can be volatile or non-volatile memory, or a combination of both. The software 13 is 2o executable by processor 15 of the PC 12 to control data acquisition and analysis in the wavefront sensor 10.
The wave&ont sensor 10 also includes a light source 16, for example, a gas laser, laser diode, superluminescent diode (SLD), or other source, for providing light exposures during image data collection. The wavefrnni sensor 10 further includes a 25 beam sputter 17, a lenslet array 18, a camera 20 (e.g., a CCD array detector), a paix of SUBSTITUTE SHEET (RULE 26) lenses 22 and 24, a focus corrector 26 (e.g., a Porno prism or aaother single-or double-pass slide mechanism), a scazuling/de-scanning mixzor 60, another pair of lenses 64 and 66, a pupil-aligning camera 28 and a pupil-aligning reflector 30, and a fixation target 32 and a fixation target reflector 34. A set of four eye pupil-s illuminating LEDs (not shown) is included for reflecting light from the subject's eye to aid in aligning the pupil with the wavefront sensor 10.
The PC 12 includes a PCI bus 40 controlled by PCI controller circuitry located within a memorylaccelerated graphics port (ACiP)/PCI controller 42 ("host bridge").
The host bridge 42 can be a 440f,X Integrated Circuit by Intel Corporation, also 10 known as the PCI AGP Controller (PAC). The controller 42 couples the PCI
bus 40 to the processor 15 and the memory subsystem 14. The processor x 5 can be a Pentium Pro, also by Intel Corpoxation. A variety of processors could be used for the processor 15 without detracting from the scope and spirit of the invention.
A video display 52, a mouse 54, and a keyboard 56 are also coupled to the is host bridge 42, enabling operator interaction with tile PC 12.
The wavefront sensor 10 includes additional hardware not shown to simplify the drawings, although their use Will be appreciated by those skilled in the art. For example, the wavefront sensor 10 includes hardware for tanning the light source 16 and pupil-illuminating LEDs on and off, and hardware for moving the focus corrector zo 26. For real-time operation, image acquisition is accomplished with a real-time fi-amegrabber, for example, a PCI-bus Bitflow Roadrunner (up to SO Hz). Also included for zeal-time operation is a slower PCI Framegrabber for gathering pupil images, for example an Imagenation PXL. The wavefront sensor 10 controls the pupil-illuminating LEDs via either a parallel port of the PC 12 coupled with a custom z5 control circuit, or via a PCI digital IO card, for example a Keithley MetraByte PIO-SUBSTITUTE SHEET (RULE 26) WO 0112809 PCT/US00l2ytrrts 24. The focus corrector 26 is controlled either thrnugh a serial port of the PC 12 via a serial driver, for example, from Velrnex, Inc., or through the same digital IO
card used to control the LEDs. A variety of other configurations for the hardware constituting the wavefront sensor 10 are possible, as will be appreciated to those skilled in the art.
The focal length of the lenslet array 18 is reduced relative to that of typical wavefront sensors to optimize performance and help facilitate real-time operation. As illustrated in Fig. 1, by reducing the fucal length (e.g., to 40 mm), the spot displacement (0) becomes smaller for the same wave aberration, which provides a larger dynamic range for the wavefront sensor 10. The spot diameter also becomes smaller because of the relationship between spot diameter (sd) and focal length given by sd = 2.44(f/d)~,= 2.44 a F# in which d is the lenslet diameter and F# is the F-numher of the lenslet, as will be appreciated by those skilled in the art. In Fig. 1, A is the locally averaged wavefront slope in liont of the lenslet array 18 and is related to is the spot displacement and the lenslet focal length by 8 = ~If, as will also be appreciated by those skilled in the art. The smaller spot diameter gives less error in "centroiding" (i.e., making "centroid" measure.ments). (A "centroid" is an intensity-weighted balance point of Hartmann-Shack spot image data collected from the detector of the camera 20. Centroiding is a process of locatixzg the spots imaged by 2o the lenslet array 18 on the detector of the camera 2U by calculating the positions of the corresponding centroids in the image data). Increasing lenslet diameter will achieve the same effect j usl mentioned. Larger lenslets reduce the highest spatial frequency that can be measured in the wave aberration. However, reducing focal length increases dynamic range in the wavefront sensor 10 and decreases the need for SUBSTITUTE SHEET (RULE 26) WO 01!28409 PCT/U~UUilvura operator intervention in measuring the wave aberration. Shorter focal lengths than 24 mm could also be used.
The wavefront sensor 10 should use as short an exposure as possible of the light source 16. This is especially true for a commercial instrument, because it reduces patient discomfort and allows the operator to collect the data with reduced risk of head movement during the exposure. Moreover, a wavefront sensor designed to track the eye's aberrations in zeal-time, such as would be required in a high resolution fundus camera equipped with adaptive optics, would need to use exposures on the order 30 msec. However, shortening the exposure can reduce accuracy 1U because spatial noise is increased in the spots, which reduces the accuracy of the instrument. This spatial noise can arise from several sources. Reflectance variations can occur across the retina due to, for example, the waveguide properties of cones and absorptance variations in the pigment epithelium. Moreover, when coherent light is used, laser speckle can be produced due to interference, which can exacerbate the non-uniformity of the wavefront sensor spots. 'These sources of inhomogeneity largely can be removed by using long exposures, on the order of 1-Z seconds.
The reason for this is that the eye moves enough on its own during this period to average out the non-uniformities. The light sources described above with respect to laser speckle and long wavelength, particularly the SLD's named, can overcome such 2o problems.
Moreover, laser diode spot images always contain a larger amount of speckle than the SLD. It would be difficult to obtain repeatable measurements of the centroid locations using the laser diode, especially at higher powers when the coherence length is longest. When driven under lasing threshold, a shorter coherence length (i.e., lower coherence) is possible; however, the output power obtained under threshold may be SUBSTITUTE SHEET (RULE 26) wu uiizsauv Yc:~rm auo~zvu~n too low to yield an adequate amount of light imaged from the retina. It may be possible to underrun current laser diodes if long exposure times are used in conjunction with a detector ofthe camera 20 that integrates throughout the exposure, such as a CCD.
The scanninglde-scanning mirror 60, which is conjugate to the pupil plane 62, also eliminates spatial noise, for example, when short exposures of the wavefront sensor 10 are required, by scanning a point source in a line on the retina, as will be appreciated by those skilled in the art. This benefit will accrue naturally in a confocal laser ophthalinoscope equipped with adaptive optics. In that case, the same scanning to beam that is used for imaging can be used for low noise wavefront sensing.
Scanning is not recommended in a corrunercial instrument designed only to measure the static component of the wave aberration, provided it is possible to either average the results from a number of short exposures, or collect light over a single long exposure. In c,-ither case, the recommended total exposure duration is at least 1-2 seconds for eye t5 movements to help eliminate spatial noise.
The angular extent of the scan of the mirror 60 is kept smaller than the isoplanatic patch size of an eye 58, and can be less than O.S degrees at a temporal frequency of 400-600 Hz. The temporal frequency is set high enough that there are many scans within a single frame of the detector of the camera 20. By having the z0 light return from the eye 58 through the same scanning optics, it is de-scanned so that there is no movement of the image of the pupil on the lenslet array 18. Thus, the spatial noise is xemoved without disrupting measurements made with the wavefront sensor 10.
To initiate data acquisition for determining the wavefront aberration of the eye 25 58, light from the light source 16 is directed to the eye 58. 'fhe light passes through SUBSTITUTE SHEET (RULE 26) WO 01/28409 Yl: twa~.~~.,."...
the beam sputter 17, the lens 24, the focus corrector 26, the Iens 22 and is reflected off the mirror 60 through the lens 64, the lens 66, the mirror 34, and finally into the eye to be focused onto a spot on the retina that is scanned by the mirror 60. The wavefront sensor 10 images the spot focused on the retina. The image is directed back through the system until it is reflected off the beam splitter 17 and then imaged by the lenslet array 18 onto the detector of the camera 20. The focus corrector Z6 is adjusted so that ideally the light entering at the pupil plane 62 of the eye is conjugate with the light entering the lenslet array 18. The spot on the retuia effectively becomes a point source for imaging by the array 18.
When an image is acquired by the wavefront sensor I O from the pupil of the eye 58, whetlier the wavefront sensor 10 is in real-time mode or is a static mode, the eye 58 must be aligned properly with the wavefront sensor 10. The eye 58 should be optimally centered in two dimensions while the subject is viewing the fixation target 32 during data collection. The pupil camera 28 provides an image of the subject's i s pupil for alignment purposes. The geometrical center of the pupil or the corneal reflection of the pupil-illuminating LEDs can be used to optimize pupil centering.
During image acquisition, the pupil camera 28, which later can be used to confirm subject alignment, acquires a final image of the pupil. The software 13 then turns off all the LEDs on the wavefront sensor 10, toms on the light source 16, and snaps a frame or multiple frames with the wavefront sensor camera 20. The light souxce 16 is then tamed off, which signals the end of image acquisition for the subject. In one embodiment, the view on the computer monitor 52 changes from the pupil image to the image acquired with the camera 20. An entire array of lenslet spots on the detector of the camera 20 should be visible i n a displayed image on the monitor 52.
Tn real-time mode, the wavefront sensor 10 continually acquires, processes, centxoids, SUBSTITUTE SHEET (RULE 26) WO Ol/28~09 PC1'lUauum~ m fits Zemike modes, and reconstructs the wavefront (all at up to 50 Hz) after a key press enters this mode and until another key press ends this mode.
In another scanning wavefront sensor system (not shown), instead of a Ienslet array, a single aperture is scanned across the pupil of the eye and the resulting light is imaged onto a single detector. The resulting array of spots is produced temporally rather than spatially, as in the wavefront sensor I0. Nevertheless, with the scanning instrument, centroids would be produced that could be analyzed analogously to the techniques discussed herein, as will be appreciated by those skilled in the art.
The hardware just described operates under the control of software for data to acquisition and analysis. The operation of the sensor through that software will now be described.
Data Acquisition and Analysis Software A more detailed discussion is now presented on the software 13 used for data acquisition and malysis with the wavetront sensors 10. In one embodiment, the t 5 software 13 is object-oriented code that provides a user interface, data presentation, and real time aberration acquisition. The sof~vare 13 also provides centroiding, variable lenslet number, variable Zernike mode number, missing data, and pupil centering, as will be discussed below in moxe detail. The software 13 stores the value of the ratio between pupil size of the eye 5 $ and the pupil size of the wavefront sensor 20 10, and allows a real-valued scale conversion between the two. The software 13 also can accommodate a variety of Ienslet arrays with different lenslet diameters and focal lengths.
Once centxoid image data have been collected by the wavefront sensor 10 by execution of the software 13, the wavefront sensor 10 uses the software 13 to 25 calculate the wave aberrations of the eye 58 to a given order and a given number of SUBSTITUTE SHEET (RULE 26) wu onzs~uv Yc~rmsuo~zvo~s Zemike polynomial modes, as will be appreciated by those skilled in the art.
The highest Zernike mode number for auy given Zernike polynomial order is specified by (order+1)(order+2)l2. Using the wavefront sensor 10, it has been found that the maximum number of Zemike modes that can be reconstructed accurately from a least-squares analysis of raw wave aberration data is, very roughly, equal to one-half the number of lenslets in the wavefront sensor I 0. To enbance such accuracy, the software 13 has been designed to restrict the number of Zernike modes used.
The software 13 includes a "modal" wavefronl reconstruction method to reconstruct the wavefront from the Hartrnann-Shack centroid data. The method used to by the software 13 to obtain the centroid data from the Hartmann-Shack raw spot data will be described in more detail below. With centroid data available, whether just obtained or previously obtained by the wavefront sensor 10, or loaded from an existing file in the memory 14 or from another memory location, the PC 12 is prepared to calculate the wave aberration of the eye 58. Generally, the wave aberration function is expressed as a linear combination of selected basis functions.
'fhe basis functions used here are called Zernike polynomials. The Zernike polynomials are an orthogonal set of two variate polynomials defined over the unit-circle, and allow a simple expression for the residual error in the wavefront after a certain degree of phase compensation is made.
2o To formulate the procedure used in the software 13 for detenruning the wave aberration of the eye 58, the wave aberration function, ~, is first expressed as a linear combination of Zernike polynomials, namely, N
~~r~ ~~ - ~ arZi ~r~ ~~~ ~rl ~ 1.
l=2 3fi SUBSTITUTE SHEET (RULE 26) wu uuz~soy Yc:~rm~ourzvu7s In that equation, Zirepresents the ith Zernike mode. The summation starts at two since the constant term, called piston, is of no concern and cannot be measured with slope estimates. The highest mode desired to be measured in the wavefront reconstruction is the value N. The maximum value of N in the software 13 is 66, which corresponds to Zernike order ten, although the software 13 could be modified to accommodate higher orders and modes. The actual order used depends upon the number of lenslets 18 within the pupil over which the aberrations are measured. The suh-aperhores of the lenslets 18 are typically centered on a square lattice and arranged so that as many as possible fit into the specified pupil. They are numbered fiom 1 to K
1o from left to right, top to bottom, although another numbering scheme could be used.
One example of a lenslet array- is a 37-lenslet array in which the lenslets are arranged in seven rows and seven columns, excluding the lerislets that view parts of the image falling outside the pupil. The lenslets in the array form oue 3-lenslcl row, one 5-lenslet row, three 7-lenslet rows, one 5-lenslet row and one 3-lenslet row and are numbered thus:

9 10 I1 12 I3 I4 t5 3d 31 32 33 34 The wavefront sensor 10 and the software 13 permit selection of a variable "pupil" radius by limiting lenslet points effectively to be within a certain range on the detector of the camera 20. Further, the system permits currently stored data to be normalized to a smaller pupil without discarding data, as will be described below.

SUBSTITUTE SHEET (RULE 26) WO Oi/28~09 PCTIUS00129(178 This is reflected in the flowchart ofFig. 13 also descxibedbelow."'Tlie'app~'~i~~te wavefront (and ultimately, an ablation pattern) can be calculated, for example, in determining the wave aberration for a 3mm pupil in normal lighting when the captured data is from a 6rrtm pupil in ditty lighting. In general, it is desirable to maximize the calculated pupil far imaging applications.
The general procedure by which sub-apertures of the lenslets 18 are determined to be within the pupil radius of the wavefront sensor 10 is as follows. A
priori, the size of the "pupil" radius ox radius of image that is necessary to capture all the centroids is calculated. This is done based on the inter-lenslet spacing of the to Ienslets 18 as well as the number of Ienslets 18 in the image. An advantage of this is that it eliminates other portions of the captured image from the camera 20 from having to be included in the location and calculation of centroids. As illustrated in Fig. 9, there is a circular area 170 of radius Ro, which should capture each centroid within its corresponding box 172 having aside equal to inter-lenslet spacing 174. It is from within this area that centroid calculations are performed, as will be described below. The individual centruids are calculated within that pupil radius, and from the calculate centroids as compared to a reference array of ideal centroids (i.e., centered at positions 176 in Fig.9), which can either be just a derived reference array or could actually be created using a plane wave input to the wavefront sensor 10 to take into 2o account optical aberrations of the optics. Using offset vectors (i.e., centroid displacements) from that reference array, the Zernike modes are calculated based on that initial pupil radius Ro. In Fig. 9, the centroid calculation with the pupil radius is designated as 178, the creation of fhe offset vectors from the reference array is designated as 180, and the calculation of the Zernike modes based on Ro i s designated as 182. Then., centrniding is earned out using any of the techniques described above.

SUBSTITUTE SHEET (RULE 26j Once the. eentroids are known, the wave aberration can be calculated u'~ithe Zeznike coefficients as follows. Refening again. to the equation for the wave aberration given above, and taking its derivatives with respect to the x and y yields 2K
equations. The 2K derivatives are evaluated over the array of subapertures of the s Ienslets 18. The equations can be expressed as follows (the functional arguments havc been suppressed here):
iaN
(1=_ 1,2~...~K~~
r~a rs where ~~ represents taking an average over sub-aperture 1. Although calculation of the wave aberration from knowledge of the Zernike coefficients is well-known, a matrix method is used to determine the 2emike coefficients in accordance with an is embodiment of the invention. As already noted, the two equations just set forth can be rewritten in matrix form as s-~sv~-CGYJa=Ga, where s" and s~' are the vectors of slope estimates obtained from the movement of the 2a centroids of the Ienslet spots in the x and y directions on the detector of the camera 20, respectively. In other words, s is simply the displacements dx and dy of the centroids from the positions that they would occupy in the absence of wavefront aberration. Of note, the number of 7,emike modes in the G matrix can be reduced simply by eliminating columns of the G matrix. The vector a is the array of unknown 25 Zernike coefficients, and G" and Gy are matrices of derivatives of the basis functions SUBSTITUTE SHEET (RULE 26) wo ov2s~o9 averaged over the sub-aperture of each lenslet 18, in the x and y directions.
This matrix is figured as a constant matrix based an the number of lenslets and the lenslet spacing, and is calculated to have at least the greatest cumber of Gernike modes one might be interested in determining and at least the greatest number of lenslets (x2) 5 that one might be monitoring. That G matrix can be calculated once based on the physical data. if square sub-apertures (e.g., like the box 172 of Fig. 9) are assumed and if b is one-half the sub-aperture side, then G" and G'' can be expressed as follows, where j (column) indexes the Zernike mode and i {row) indexes the aperture centers:
1 ~r+e Y,+b ~Z .
G~ -4bz ~ ~ ~ dye' ~r b Y.r_b 1 y,+b x,+b aZ
Gi'øbz ~ ~ ~dxdy~
y~-b x,-b ay These two equations can be simplified by defining the antiderivatives of the Zernike polynomials with respect to the x and y variables as IZ~ ~x, y) = ~Z; (x, y~dx and 15 IZ; ~x, y) = !Z; ~x, y)dy . The software 13 calls these functions "TntZx"
and "IntZy"
respectively. The above equations can then be expressed in terms of these functions as follows:

G~ = 4~z ~iZ~ ~x; +b, yt +b~+IZ~ ~x, -b,Ya-b)-IZ; ~x, +b, y; -b)-IZJ ~x, ~b, y; +b)~
G ~ ~ 4~2 ~jZ~ ~x; +b, y; + b~+ IZl ~x; - baY< - b~- ~; ~xr + b, Ye "b~-' IZ;
~xr -' b~Yt -+-b~~
Note that Gx depends on IZ; (x,y) and vice-versa.
ac SUBSTITUTE SHEET (RULE 26) W0 01/28:109 Because the equations set forCh until now provide an ovcrdetermined system of equations (N l~ll~, a least-squares solution is sought. A procedure called Singular Value Decomposition (SVD) provides a solution to the equation, which is optimal in tlus sense. In order to determine the Zernike coefficients, essentially a 5 least-square fit is made to the coefficients using the SVD of the matrix equation involving the measured centroid shifts (explained below) in each sub-aperture and the average of the Zernike polynomial derivatives over each sub-aperture. The G
matrix is inverted and further processing occurs to yield other matrices Iiom which a can be determined and thus the wave aberxation of the eye 58. Applying SVD to the matrix G, matrices U, V and D are obtained such that G=IJDV'. D is a diagonal mal;rix ofN
1 "singular" values whereas U is a column-orthonormal matrix with the same dimensions as G, and Y is an N 1xN 1 orthonocmal matrix. Since the inverse of an orthogonal matrix is simply its transpose and the invcrsc of a diagonal matrix is simply the diagonal matrix of its reciprocals, we have 15 a ~ G-~s = (~D'lll~s , where G-' is the inverse of the G matrix and the superscripted t represents the transpose of a matrix.1t can be shown that the wavefront error in this method is proportional to Tr(D~1), i.e., the sum of the diagonal elements of D-1.
2o The software 13 selects the lenslet geometry by asking the operator for a pupil radius (or diameter) over which the eye's wave aberration is to be measured.
The pupil radius for the eye may be different from the wavefront sensor, but ideally they are the same. From the pupil radius information and the known center-to-center distance of the Ienslets 18 as imaged on the detector of the camera 20, the program 25 pre-computes the matrix G for the largest number of lenslets that can be a1 SUBSTITUTE SHEET (RULE 26) wo ovzsao9 accommodated for that pupil radius fox the Zernike modes up to, far example, order ten. (However, the initial ozder actually used will depend upon the number of lenslets 18 used). After the initial computation, the operator is allowed to have the computer 12 (i.e., the software 13) disregard the slope estimates from selected lenslets 18 as 5 well as have the software 13 compute estimates for a larger ox smaller Zernike order, as will be described in more detail below. The number of modes can be determined for any given Zernike order fxrom the expression (ordem~-1)(order+2)I2. See Dai, Gung-Ming (1995), "Theoretical Studies and Computer Simulations of Post-Detection Atmospheric Turbulence Compensation", Ph.D. thesis, Lund University, Lund to Sweden, Mahajan, V.N. (1994), "Zernike Circle Polynomials and Optical Aberrations of Systems with Circular Pupils", I:nginc;ering & habozatoiy Notes, August 1994, pp.
S21-524, Southwell, W.H. ( 1980), Journal of the Optical Society of America, Vol. 70, p. 998, Tyson, R.K. (1991 ), Principles of Adaptive Optics, Academic Press, Inc. New York, NY, Wang, J.Y. & Silva, D.E. (1980), "Wave-front Interpretation with Zernike I5 Polynomials'", Applied Optics, VoI. I9, No. 9, pp. 1510-1518.
As a means of explaining the matrix equations above, the following simplified example is offered. A hypothetical 3x3 array 150 of reference spots and boxes 152 is shown in Fig. 8A. Each box of the array of reference spots 150 and boxes 152 are designated by an index 1-9 from top to bottom and left to right, although another 20 index ordering scheme could be used. For each reference spot 150 determined by the software 13, there is a centroid 154 displaced in x and y from the corresponding aperture center 156 that also is determined by the software 13 within each box, as will be explained below in more detail. The software 13 determines the x and y displacements of each centroid 154, designated dx; and dy;, respectively for i =1 to 9.
25 Thus, the s vector has 18 components, 9 for x and 9 for y, as shown in the matrix :l2 SUBSTITUTE SHEET (RULE 26) WO O1/28.~09 PCT/US00/29078 equation below. The G matrix is an 18 centroid (or aperture center) by 5 modes matrix (i.e., 18x5) and a is the vector of the Zernike coefficients consisting of the 5 components corresponding to first and second order aberrations (no piston teem, as explained above). The G matrix has the same form as indicated above with b equal to one-half the size of the side of the boxes.
S [ 18] G [18 x 5] a [5]
dx, dx2 dx3 ~4 ~5 dx6 a~

a~
dx9_ G ~ as dye dyZ as a6 dY3 dyy dys dY6 dYs to Once the magnitudes of the Zernike coed cients are determined, the wave aberration is found, where the only relevant terms are from the sum =~,ai Zi(xsY) i=2 1$
with the matrix elements of G for 18 (j = 1 to 18) rows and 5 (i = 2 to 6) columns.
Focusing now on particular modules of the software 13, more details are provided below. These modules relate directly to and. implement the techniques and SUBSTITUTE SHEET (RUSE 26) procedures described thus far. A flowchart diagram tlq~t"~tces in°Fig.
13 an overall method for the controlled data acquisition and analysis software 13 discussed above in more detail, in accordance with an embodiment of an invention. Control begins in a Begin Exam Process step 200 and proceeds immediately to a Set Parameters step 202.
5 In the Set Parameters step 202, the Uptics object is created and the values o~the Pupil Scale variable and the Wavefront2PupilScale variable; are set to default values based on the particular configuration of the wavefront sensor 10. 'fhe values of other variables, such as a pupil radius variable, an inter-lenslet distance variable, and a lenslet ~ocal length variable are also set based upon default values. As an alternative, 10 the variables set in the Set Parameters step 202 may be determined dynamically or asked for and entered by the operator.
Control then proceeds to a Create Lenslets step 210, which is described in more detail below with respect to Fig. 14. Next, control proceeds to a Set Pupil Radius step 215 in which a local pupil radius variable is reset to a radius value 15 determined in step 210. The value of the local pupil radius variable is employed in subsequent steps and is based on the smallest pupil radius that can fit around a set of Ienslet objects created in step 210 that correspond to a set of lenslets within the radius.
Control then proceeds to a Get Image step 220 in which an image of the retina is captured by the wavefront sensor 10 ~or processing. As an alternative, the image can 2o be retrieved from a file stored in the memory 14 of the computing system 12 (Figs. 2 and 3) or other memory.
Following the Get Image step 220, control proceeds to a Generate Reference Array step 230 in which a reference object is created. The reference object contains a set of points that represent aperture center points (like those of Figs. 8A
and 8B) of 25 each lemslet t n the set of lenslets created in step 210. The reference array is employed SUBSTITUTE SHEET (RULE 26) WO 01128409 ' in a Generate Valid Zernike Modes step 270 described in more detail below. The use of a reference array and the generation of Zernike modes has been descn'bed above.
Control proceeds to a Locate Centroids step 250 zn which a centroid is generated for each Ienslet of the lenslet array created in step 210. 'fhe Locate Centroids step 250 is described in more detail below with respect to Fig. 15.
Following the Locate Centroids step 250, control proceeds to a Select Centroids step 260 in which the operator may sei, adjust, or eliminate individual centroids generated in the Locate Centroid step 250, as discussed above.
Control then proceeds to the Generate Valid Zernike Modes step 270 in which Zemike modes are i0 generated based on the lenslets 18, the centroids, and the reference arrays, as described above. Once the Zernike modes are generated, then control proceeds to a Reconstruct 1~'avefront step 280 in which the wavefront of the image captured in. step 220 is reconstructed based on the Zerciike modes calculated in step 270. The reconstruction of the wavefront based on the Zernike modes is accomplished once the Zernike coefficients are determined using the matrix approach discussed above according to the invention. Following the Reconstruct Wavefront step 280, control proceeds to a parameter change process A, which is described in detail below in conjunction with Fig. 16, and which relates to the discussion above regarding centroid elimination and pupil radius change.
Referring again to Fig. 13, also included in the flowchart of the Begin Examination Process 200 are entry points B, C and D. Entry points B, C, and D
enable the Begin Exam Process 200 to be entered at the Set Parameters step 202, the Select Centroids step 260 or the Generate Valid Zernike Modes step 280, respectively.
Entry points B, C, and D are described in more detail below with respect to Fig. 1G.
SUBSTITUTE SHEET (RULE Z6) Taming now to Fig. 14, a flowchart is illustrated of a Create Lenslets process 300 corresponding to the Create Lenslets step 210 of the Begin Exam Process (Fig. 13). h1 the Create Lenslets process 300, control proceeds immediately to a Too Few Lenslets step 212 in which a calculation is performed to determine whether information from operator input will enable a lenslet number parameter greater than a pre-determined minimum to be generated for use in determining the wave aberration of the eye 58. If the answer is "No," the process 300 completes in a step 352 and the Begin Examination Process 200 is terminated. As an alternative, a process may be initiated to enable the operator to reselect parameters so that a sufficient number of lenslets are generated. If the answer in step 212 is "Yes," control proceeds to a Calculate Number of Lenslets step 214, which is described in more detail below with respect to Fig..l5. Following the Calculate Number of l,enslets step 214, control proceeds to a Determine Pupil Coverage step 216 in which a ratio behveen a total area of alt the lenslets created (e.g., corresponding to 37 lenslets used as the lenslets 18) in is the Create Lenslets std 210 (Fig. 13) and the area of the pupil as determined by the radius set in the Reset Pupil Radius step 2I5 (Fig. 13) is calculated. Control then proceeds to a Too Little Coverage step 217 in which the ratio calculated in step 214 is compared to both a predetermined minimum value and a predetermined maximum value. For example, the minimum could be 75%. If the ratio is less than or equal to the minimum, indicating that there are too few lenslets (centroids) to accurately reconstruct the wavefront for Zernike modes to a given Zernike order (e.g,, as described above), or if the ratio is greater than the maximum, indicating that the Generate Valid Zernike Modes step 270 and the Reconstruct Wavefront step 280 (Fig.
13) are likely to be inaccurate, then control proceeds to a Display Warning step 218.
Tn step 218, an appropriate warning message is displayed on the video screen a6 SUBSTITUTE SHEET (RULE 26) WO 01/28:f09 , (Figs. 2 and 3) to the operator. Control then proceeds to a Return step 219 in which control is returned to the Begin Examination Process 200 and proceeds to the Reset Pupil Radius step 215 (Fig. 13). As an alternative, a process may be initiated to enable the operator to adjust parameters so that the ratio is within appropriate limits.
5 If, in step 217, the ratio is gxeater than the minimum or less than or equal to the maximum, then control proceeds immediately to the Return step 219, and control proceeds as described above.
Turning now to Fig. 15, a flowchart is illustrated of a Calculate Number of Lenslets process 310 corresponding to the Calculate Number of Lenslets step 214 of 10 the Create Lenslcts process 300 (Fig. 14). Control proceeds immediately to a Create Maxgrid step 312 in which a matrix of points (grid) is generated corresponding to an array of lenslets with a predetermined maximum number of rows and columns, for example, 19x19. The maximum is a matter of convenience and reserves some memory for the largest lenslet array likely to be encountered. A subset of that is memory can be used to create the grid corresponding to the Ienslets that are actually used. In addition, a count variable and a max distance variable are both initialized to the value '0'.
Control then proceeds to a step 314 in which it is determuned whether there are points in the Maxgrid array to check. The Maxgrid is used as a maximum boundary 2o within which the locations ofboxes and box ceniers that correspond to the lenslets 18 within the pupil radius are determined in the software 13. The fast time through the process 310, the answer is typically "Yes," and control proceeds to a Get Point &
Calculate Distance From Center step 315. Process 310 processes points in the grid from left to right and top to bottom, according to design choice. Others directions and zs sequences are possible. Each point is assumed to be a center of a box with sides equal ~7 SUBSTITUTE SHEET (RULE 26) W V 1) 1111f3U9 Yl~ 1'l Ull1l1/1J11'l1f to the inter-lenslet distance and the distance from the center of the grid to a comer of the box that is farthest from the center of the grid is calculated and retained. If the calculated distance is less than or equal to the pupil radius, indicating that the entire box corresponding to a lenslet is within the pupil, then the count variable is incremented by '1' and the coordinates of that box's center are stored in a first available spot in the Maxgrid array. Control proceeds to a Distance > Max distance step 318 iu which the process 3I 0 determines whether the calculated distance is greater than the previously stored value of the max distance variable. If the value of the calculated distance is greater than the previously stored value of the max distance ~o variable, control proceeds to a Store New Max Distance step 319 in which the calculated distance is stored in the max distance variable, over-writing the previously stored value, Control then proceeds to Points To Check step 3 I4 when the new value of the max'distance variable is stored in step 319, when it is determined that the calculated distance is less than or equal to the previously stored value in step 318, or when the calculated distance is greater than the pupil radius in step 316, indicating that the box does not fit entirely within the pupil radius. In step 314, the process 310 determine whether there is another point in the Max~-id array to pxocess and, if so, loops through steps 315, 316, 317, 318 and 319, as described above. Each grid point in the 19x19 array is processed in turn to determine whether it is to be included in the final centroid array.
Referring again to step 314, if the process 310 determines that there are no more points to check, then control proceeds to step 362 in which control is returned to process 300. When control is returned to process 300, the count variable contains the value of the number of boxes, or lenslets, which fit within the pupil radius and the as SUBSTITUTE SHEET (RULE 26) wr7 oing:~o~ '' max distance variable contains the value of distance of the farthest corner of all the boxes that fit within the pupil radius. In addition, the Maxgrid array contains coordinates for each box, or lenslet, which fits within the pupil radius.
Turning now to laig. 16, a flowchart is illustrated of a Locate Centroids s Process 400 that corresponds to the Locate Centroids step 250 ofthe Begin Examination Process 200 Fig. 13). The Locate Centroids Process 400 determines the location of the centroids as described above using the method of Figs. 10-12.
Control proceeds immediately to a Lenslet To Process step 221 11 wh~cll the process determines whether there is a lenslet to process that corresponds to the lenslets created 1o in the Create Lenslet step 210 of the Begin Exam Process 200. The first time through, the answer is typically "yes" and control proceeds to a Get Lenslet step 222.
The Locate Centroids process 400 starts with the top left lenslet in the lenslet array and processes the array from left to right and top to bottom, according to design.
Other directions and sequences are possible.
15 After a lenslet is selected in step 222, the process 400 proceeds to an Initialize Boxsize & Box Center step 223. In step 223, a "search box" is defined in the image retrieved in step 220 of the process 200 using the coordinates of the center of the selected lenslet as the center of the search box and with the length of the sides equal to the inter-lenslet distance definers in step 202 of process 200. The search box 2o corresponds to the boxes shown in Figs. 10-12. The centroid fox each spot within the search box is calculated as described above. Control then proceeds to a Box Size <
Minimum step 225 in which process 400 determines whether the search box is smaller than or equal to a predefined minimum value. The minimum value may be, for example, the size of the diffraction limited spot corresponding to an idealized 25 reference plane wave impinging on and focused by the same tenslet, as will be SUBSTITUTE SHEET (RULE 26) appreciated by those skilled in the art. If the search box is greater than the predefined minimum value, control proceeds to a Find Centroid & Reset Box Center step 227.
The centroid of the search box corresponds to the center of mass of that search box. In other words, the process 400 weights each pixel in the search box by the 5 intensity of that pixel and defines the ccntroid. Other parameters may also be used instead of, or in addition to, the average intensity to determine the centroid location.
For example, minimum and maximum pixel intensity threshold values may be defined with only pixel values within the thresholds used to calculate the c;entruid.
In step 227, once the centroid of the search box is determined the center of the search box is 1 o reset to the coordinates of the centroid and control proceeds to a Determine Centroid Validity step 229. In the disclosed embodiment, in step 260 of the process 200, the operator detezmines centxoid validity. 'fhe most common reason for the operator to move a centroid manually is that the centroid location does not appear to be centered on the captured spot of light that is to be centroided. But, in an alternative 15 embodiment, centroid validity may be determined automatically in step 229.
Fox example, a centroid that is outside the borders of the initial search box defined in step 223 may be considered invalid.
After step 229, an iterative centroiding process as described above is performed. That is, control proceeds to a Reset Box Size step 231 in which both the 20 length and width of the search box are reduced by une (or more) pixel.
Control then returns to step 225 in which the process 400 determines whether the new box size is Less than or equal to the minimum box size again. As described above, a variety of iterative centroid locating processes could be used. Preferably, the optics of the wavefront sensor 10 are of sufficient qualify to guarantee that the centroids will be 25 located somewhere within each search box for each iteration. If the search box size is SUBSTITUTE SHEET (RULE 26) still greater than the minimum value, control proceeds to'syt'g'~ x'27 as~'e'~EpI~al~l~G~.
If the search box size is less than the miniumum, then control proceeds to step 223 in which the location of the centroid is stored in the cenuoid array and the validity of the centroid is stored in a corresponding location in a validity array.
5 Control then proceeds to a Centroid Valid step 235 iii which the appropriate entry in the centzoid validity array is checked. If the centroid is valid, then control proceeds to step 237 in which the centroid is drawn in a centroid image or displayed on the display 52 and stored (e.g., in the memory 14). Corrtrul then proceeds to step 221 in which the process X100 determines whether there are additional lenslets to l0 process as explained above. If, in step 235, the ccntroid is determined to be invalid, control proceeds immediately to step 221. Finally, in step 221, if the process determines that there are no more centroids to process, control'rehlrns through End Locate Centroids step 272 to step 250 of the process 200 of Fig. 13. With valid centroids and Zernike modes, the wave aberration is determined as described above.
15 After the wave aberration has been determined, the operator, in one embodiment, can either re-profile the software 13 from the program beginning (Fig.
13, step 200), change program inputs, acquire and analyze data again, while in another embodiment, program inputs can be changed on the fly without reinitialization of other parameters, meaning that a new matrix is computed.
20 The validity of the Zernike modes eau be assessed and the pupil radius can be changed aftex an initial wave aberration construction or reconstruction, as described above. Such changes vary program flow in the software 13 in reconstructyng the wave aberration. For example, the pupil of the eye 58 may initially be dilated for image capture and wave aberration constmction and then the operator decides that the 2s wave aberration should be reconstructed over only a central portion at smaller pupil s1 SUBSTITUTE SHEET (RULE 26) radius. In other words, the radius merely is truncated and all the valid data collected previously can be used again. However, if the pupil radius is increased, then new information is needed that was not gathered previously. A new nuraber of lenslets in the lenslet array 18 may be needed to cover the larger pupil radius as well as a 5 corresponding new number of. centroids and reference points used for wave aberration reconstruction.
Whether to change radius can be included for automatic execution in the software 13 in certain embodunents, depending on conditions of the measurements and the desired result. In other embodiments, the operator decides at decision points i0 or the operatar decides to reexecute the software 13 from the beginning. In reducing pupil radius, the operator may or may not want to use the same Zernike polynomial order or number of Zemike modes. The operator may want to Lower one or bath of these. If the same order and modes are desired, the wave aberration is merely re-normalized to the reduced radius size using a new set of Zemike coefficients.
is Turning now to Fig. 17, a flowchart is illustrated of the parameter change process A first introduced above in conjunction with Fig. I3 that allows parameters of the wavefront sensor 10 and data acquisition slid analysis to be changed, as discussed above. Control proceeds immediately to a Change Parameters step 29U. if the operator does not want to change parameters, the process A and the Begin 2o Examination Process 220 (Fig. 13) are completed in a Dane step 291. If the operator chooses to modify parameters in step 29U, then control proceeds to step 292 in which it is determined whether one desired modification is an increase in the pupil radius, a modification in the inter-lenslet distance, or a modification in the lenslet focal length.
Other modifications can be made in certain embodiments, as will be apparent to those 25 skilled in the art. In that case, control proceeds to step 292 in which the lenslet abject SUBSTITUTE SHEET (RULE 26) WO O1/28t09 PCT/US00129078 created in step 210, the reference object created in step 230, and the Zernike modes created in step 2?0 can be destroyed to free up memory. Control then proceeds to the B entry point of the Begin Examination Process 200 or the Set Parameters step 202.
It should be understood that if any of the conditions that initiate this control sequence are true, then other parameters, such as the number of Zernike modes, may also be modified without changing the sequence.
Referring again to step 292, if the answer is "No," then control proceeds to step 294 in which it is determined whether the desired modification is a decrease in the pupil radius or a modification of the centroid data. If the answer is "Yes," then control proceeds to a Delete Zernike step 298 in which the Zemike modes created in step 270 are destroyed. Control then proceeds to the C entry point of the Begin Examination process 20U or the Select Centroids step 260. Ifthe conditions that initiate this control sequence are tn~e, then the number of Zernike modes also may be changed without changing the sequence.
Referring again to step 294, if the answer is "No," then conh~ol proceeds to an t~lter Zernike Modes step 296 in which the number of valid Zernike modes is altered, and then proceeds to the D entry point of the Begin Examination Process 200 or the Generate Valid Zernike Modes step 280. At this point, the description of Fig.
13 is complete for the software 13.
Applications and Experimental Data The present ilivention has applications in such areas as retinal imaging, fabrication of lenses such as contact lenses and intraocular lenses, and laser refractive surgery. To show the utility of the present invention, some experimental results will be set forth.

SUBSTITUTE SHEET (RULE 26) WO 01/28.109 PCT/US00I29078 One useful application of the real-time wavefront sensor is to measure the stability of the wave aberration of the eye. Tl>is is important because the effectiveness of a custom contact lens used to correct higher order aberrations depends on the stability of the eye's wave aberration. If the wave aberration changes either over time or due to changes in viewing conditions, these changes will limit the ability of any static optical element to provide sharp visiorL Experiments were performed that address this question and the results are provided in the following sections.
Fig. 4 shows, for three subjects, SC, I3H, and PA, a measurement of the amount of coma, astigmatism, and spherical aberration when the subject changed to his/her accommodation smoothly from one state to another. Pupil size was 4.7 mm.
The accommodative state is expressed as Zernike defocus in microns, where 1 micron is equivalent to 1.26 diopters of defocus. These data show that there is considerable individual variability between subj eels in exactly how the aberrations change.
Nonetheless, it is clear in each subject that there are substantial, systematic changes in the aberrations that are dependent on accommodative state. The implication of this is that a static optical element or refined surgical procedure that was designed for one viewing distance would be less effective for other distances.
Fig. 5 illustrates this point with the modulation transfer functions (MTFs) computed from measurements of three subjects' wave aberrations. The Curve 1 for 2o each subject shows the M'1:'F that would be obtained if an ideal static correcting element were placed in front of the eye when it is focused at infinity. The pupil size for this calculation was 4.7 mm. This MTF shows the effects of diffraction alone.
The Curve 2 shows the M'TF when all the higher order aberrations (that is, those other than defocus) have been perfectly corrected had the subject been accommodating at infinity, when in fact the subject was accommodating at two dioptcrs. It 1S
nllpOrtant SUBSTITUTE SHEET (RULE 26) WO 01/28.109 PCT/USOO129078 to keep in nvnd that defocus is not responsible for the fact that this curve lies way below the MTF when all aberrations are corrected, shown here. In fact, the amount of defocus has been optimized for this curve under the assumption that the subject's accommodation was perfect. Fox comparison purposes, also shown is the subject's MTF (Curve 3) when none of the higher order aberrations have been corrected for the same accommodative state. fVhat this shows is that a static contact lens custom designed to correct perfectly all the higher order aberrations at one accommodative state does little or no good at correcting higher orders when the acconnnodative slate is changed by two diopters: Though the changes of the wave aberration with to accommodation limits the conditions under which a supercorrecting lens would provide a benefit in young people who can still accommodate, it does not imply that such a lens would have no use. This limitation is not particularly severe because it does not apply to presbyopes and, even in younger people, there would be value in designing the lens to correct for di stance vision.
Rapid temporal fluctuations in the wave aberration might reduce the value of a static correction element. Though it has been larown for some time that the eye's optics are not stable on short time scales, apart from the microfluctuations of accommodation, this instability has never been characterized. Short-term instability of the eye's aberrations will affect other types of wavefront correction besides a static 2o element. For example, a previous wavefront sensor can be coupled to a deformable mirror (see the aforementioned U.S. patent 5,777,719, to correct most of the eye's aberrations, and can provide unprecedented retinal image quality, but its performance is fundamentally limited because the correction it applies is essentially static. Greater gains in optical performance could he achieved with a dynamic, real-time system SUBSTITUTE SHEET (RULE 26) WO O1/28.~09 PCT/US00129078 capable of tracking the fluctuations in the eye's aberrations. Designing suc'~i a.°system requires knowledge of the wave aberration dynamics.
To measure these dynamics, in accordance with an embodiment of the invention, the real-time wavetiont sensor 10 was used. The wavefront sensor IO
can s also be coupled to a defonnable mirror, like that in U.S. patent 5,777,719, to provide superb retinal image quality, as will be appreciated by those skilled in the art.
The bandwidth a perfect adaptive optics system would need to be able to keep up with these temporal fluctuations in the higher-order wave aberrations was investigated. How fast it would need to go so that to really achieve diffraction limited 14 performance was addressed. To do this, the time averaged Strehl ratio, defined as the time averaged ratio of the peak height of the actual point spread function to the height of a diffraction limited point spread function that would be obtained after correction by a perfect adaptive optics system of various bandwidths was computed. The Strehl ratio is a metric often used to assess image quality. 'fhe curves for two subjects are 15 shown in Fig. 18 for a 5.8 nun pupil and natural accommodation, one curve being the average. A Strehl ratio of 0.8 or above is considered to be diffraction limited. Also shown is the Strehl ratio for the best refracted case. Going from the best correction of defocus and astigmatism to the best static correction of all the monochromatic aberrations (which corresponds to zero bandwidth) goes alinost all the way to 20 diffraction limited performance. This is an important result because it implies that the error introduced by temporal dynamics of the eye's aberrations are small compared with the errors that can be corrected with a static correcting element. In other wards, there is probably no need to worry very much that the dynamics of the wave aberration will obscure attempts to achieve better visual performance with a static-SUBSTITUTE SHEET (RULE 26~

WO O1/28.t09 PCT/US00/29078 correcting element, such as a customized contact lens, or with refined surgical methods.
In order to go the rest of the way from perfect static correction to the diffraction limit, an adaptive optics (AO) system with a bandwidth of less than 1 tiz is s needed. This does not mean that the eye does not exhibit fluctuations beyond a frequency of 1 Hz. It has been observed that significant frequency components exist at least out of 6 Hz or so. This simply means that for this pupil size and this condition the power carried by the fluctuations in frequencies above 1 Hz is just too small to significantly impact image quality, which is advantageous. A closed-loop AO
correction requires a sampling rate at least 10-20 times the bandwidkh of the fluctuations that are to be corrected, implying that an AO system operating at Hz would be able to track all the significant fluctuations in the eye's optics for the 518 mm pupil size. This is encouraging because the technology to do this is available and the present invention achieves the first real-time AO correction for the eye In summary, the first measurements of the temporal properties of the eye's wave aberration have been demonstrated. These temporal ffuciuations are small compared with the static higher order aberrations of the eye, implying that they will not interfere significantly with the benefits of a static custonvzed optical element.
While a preferred embodiment and variations theleon have been set forth in detail above, those skilled in the art will appreciate that other embodiments can be realized within the scope of the invention. For example, while the hardware has been disclosed in terms of an Intel-based YC, another type of computex could be used, as could a dedicated device. Also, the software can be prepared according to any suitable prograrruning techniques, such as object-oriented programming.
Therefore, the invention should be construed as limited only by the appended claims.

SUBSTITUTE SHEET (RULE 26)

Claims

What is claimed is:

1. A method of measuring a wavefront aberration of an eye, the method comprising:
(a) taking image data from the eye, the image data comprising a plurality of spots (88) having positions determined by the wavefront aberration;
(b) defining a plurality of search boxes (86) in the image data;
(c) determining the positions (96) of the upon; and (d) calculating the wavefront aberration in accordance with the positions of the spots;
characterized in that step (c) comprises:
(i) locating s centroid (90, 94, 96) its each of the search boxes (86);
(ii) replacing each of the search boxes (86) with a search box (92, 98) of reduced physical size;
(iii locating the centroid (94, 96) in each of the search boxes (92, 98) of reduced physical size; and (iv) repeating steps (i) and (iii) until each of the search boxes of reduced physical size (92, 98) reaches a minimum size.

2. The method of claim 1, further characterized in that:
step (a) is performed with a Hartmann-Shack detector (20) having a plurality of lenslets (18);
and the plurality of lenslets (18) form a plurality of spots (88).

3. The method of claim 2, further chanced is that:
step (d) is performed using a matrix having rows corresponding to the lenslets (18); and step (d) comprises:
(i) determining which lenslets (18) do not provide data that are usable in accurately determining the wavefront aberration; and (ii) eliminating (186) the rows corresponding to the lenslets which do not provide data that are usable in accurately determining the wavefront aberration.

4. The method of claim 1, further characterized in that the minimum size is a diffraction limited size.

5. The method of claim 1, further characterized to that step (c)(ii) comprises reducing a size of each search box to be replaced (86, 92) by one pixel in each direction.

6. The method of claim 1, further characterized io that steps (c)(i) and (c)(iii) are performed only in accordance with pixels in the image data whose intensities lie between a lower threshold and an upper threshold.

7. The method of claim 6, further characterized in that the method prompting an operator to select the lower threshold and the upper threshold.

8. The method of claim 1, further characterized in that each of the plurality of search boxes (86) defined in step (b) is centered on a position (156) which one of the spots (88) would occupy in an absence of the wavefront aberration.

9. The method of claim 8, further characterized is that each of the plurality of search (86) boxes defined in step (b) has a dimension equal to a spacing between the positions (156) which the spots would occupy in the absence of the wavefront aberration.

10. The method of claim 8, further characterized in that each of the plurality of search boxes (86) defined to step (b) has a dimension which is scaled down from a spacing (x1, y1) between the positions (156) which the spots would occupy in the absence of the wavefront aberration by a factor less than one.

11. The method of claim 10, further characterized in that step (b) comprises prompting an operator for the factor.

12. The method of claim 1, further characterized in that step (c)(ii) comprises reducing a size of each search box to be replaced (86, 92) by a fraction of a pixel along each side.

13. The method of claim 12, further characterized is that the fraction of a pixel is one-half of a pixel.

14. The method of claim 12, further characterized in that, for each search box of reduced physical size (92, 98) which inchtdes at least one fractional pixel, stop (c)(iii) comprises assigning to each of said at least one fractional pixel a contribution to the centroid which is equal to a corresponding fraction of as intensity detected in the fractional pixel centered in the fractional pixel.

15. The method of claim 1, further characterized in that step (c) further comprises:
(v) allowing an operator to drag one of the centroids to a new location; and (vi) recalculating the centroid which has been dragged to dragged to the new location, using a light distribution around the new location.

16. The method of claim 1, further characterized in that step (c) further comprises:

(v) allowing an operator to drag a central one of the centroids to a new location; and (vi) recalculating all of the centroids around the new location.

17. The method of claim 1, further characterized in that step (c)(iii) is performed without reference to a location of a previously located centroid.

18. The method of claim 1, further characterized in that step (c)(iii) is performed using a location of a previously located centroid as a start centroid.

19. The method of claim 1, further characterized in that step (c)(iii) is performed using, as a start estimate, a location which a spot would occupy in an absence of the wavefront aberration.

20. The method of claim 3, further characterized in that step (d)(i) comprises allowing as operator to select the lenslets which do not provide usable data.

21. The method of claim 3, further characterized in that step (d)(i) comprises determining automatically which lenslets do not provide usable data.

22. The method of claim 21, further characterized in that step (d)(i) comprises automatically which lenslets do not provide usable data in accordance with standard deviations of intensity around the corresponding centroids.

23. The method of claim 21, further characterized in that step (d)(i) comprises determining automatically which lenslets do not provide usable data in accordance with overall intensity levels of the corresponding centroids.

24. The method of claim 3, further characterized in that:
the matrix has two rows corresponding to each of the lenslets; and step (d)(ii) comprises eliminating (186) both of the rows corresponding to each of the lenslets which do not provide usable data.

25. The method of claim 24, further characterized in that the two rows corresponding to each of the lenslets are a row corresponding to an x coordinate and a row corresponding to a y coordinate.

26. The method of claim 1, further characterized in that step (d) comprises calculating a number of Zemike modes by using a number of centroids which is at least twice the number of Zemike modes.

21. The method of claim 1, further characterized in that step (d) comprises calculating the wavefront aberration for a first pupil radius R0 of the eye and also for a second pupil radius R1 of the eye, wherein R1 < R0.

28. The method of claim 27, further characterized in that the wavefront aberration is calculated for R1 by renormalizing the wavefront aberration calculated for R0.

29. The method of claim 27, further characterized in that step (d) further comprises prompting as operator for R1.

30. The method of claim 29, further characterized in that step (d) further comprises prompting an operator for a minimum Zernike mode and a maximum Zernike mode for use in calculating the wavefront aberration for R1.

31. The method of claim 3, further characterized in that step (d) is performed using singular value decomposition.

32. A system (10) for measuring a wavefront aberration of an eye, the system comprising:
image data taking means (20) for taking image data from the eye, the image data comprising a plurality of spots (88) having positions determined by the wavefront aberration; and computation means (12), receiving the image data, for:
(a) defining a plurality of search boxes (86) in the image data;
(b) determining the positions (96) of the spots (88); and (c) calculating the wavefront aberration in accordance with the positions (96) of the spots (88);
characterized in that the computer means (12) performs step (b) by:
(i) locating a centroid (90) in each of the search boxes (86);
(ii) replacing each of the search boxes (86) with a search box (92, 98) of reduced physical size;
(iii) locating the centroid (94, 96) in each of the search boxes (92, 98) of reduced physical size; and (iv) repeating steps (ii) and (iii) until each of the search boxes of reduced physical size (98) reaches a minimum size.

33. The system of claim 32, further characterized in that:
the image data taking means comprises a Hartman-Shack detector (20) having a plurality of lenslets (18); and the plurality of lenslets (18) form the plurality of spots.

34. The system of claim 33, further characterized in that:

the computer means (12) performs step (c) using a matrix having rows corresponding to the lenslets; and step (c) comprises:
(i) determining which lenslets (18) do not provide data that are usable in accurately determining the wavefront aberration; and (u) eliminating (186) the rows corresponding to the lenslets (18) which do not provide data usable in accurately determining the wavefront aberration.

33. The system of claim 32, further characterized in that the minimum size a a diffraction limited size.

36. The system of claim 32, further characterized in that step (b)(ii) comprises reducing a size of each search box (86, 92) to be replaced by one pixel in each direction.

37. The system of claim 32, characterized in that steps (b)(i) and (b)(iii) are performed only in accordance with pixels in the image data whose intensities lie between a lower threshold and an upper threshold.

38. The system of claim 37, further characterized in that it comprises interface means (54, 56) for prompting an operator to select the lower threshold and the upper threshold.

39. The system of claim 32, further characterized in that each of the plurality of search boxes (86) defined in step (a) is centered on a position (156) which one of the spots (88) would occupy in an absence of the wavefront aberration 40. The system of claim 39, further characterized in that each of the plurality of search boxes (86) defined in step (a) has a dimension equal to a spacing between the positions (156) which the spots (88) would occupy in the absence of the wavefront aberration.

41. The system of claim 39, further characterized in that each of the plurality of search (86) boxes defined in step (a) has a dimension which is scaled down from a spacing (x1, y1) between the positions which the spots (88) would occupy in the absence of the wavefront aberration by a factor less than one.

42. The system of claim 41, further characterized in that it comprises interface means (54, 56) for prompting an operator for the factor.

43. The system of claim 32, further characterized in that step (b)(ii) comprises reducing a size of each search box (86, 92) to be replaced by a fraction of a pixel along each side.

44. The system of claim 43, further characterized in that the fraction of a pixel is one-half of a pixel.

45. The system of claim 43, further characterized in that, for each search box of reduced physical size (92, 98) which includes at least one fractional pixel, step (b)(iii) comprises assigning to each of said at least one fractional pixel a contribution to the centroid which is equal to a corresponding fraction of an intensity detected in the fractional pixel centered in the fractional pixel.

46. The system of claim 32, further characterized in that is comprises interface means (54, 56) for allowing an operator to drag one of each centroids to a new location, and in that step (b) further comprises recalculating the centroid which has been dragged to the new location, using a light distribution around the new location.

47. The system of claim 32, further characterized in that it comprises interface means (54, 56) for allowing an operator to drag a central one of the centroids to a new location, and in that step (b) further comprises recalculating all of the centroids around the new location.

48. The system of claim 32, further characterized in that step (b)(iii) is performed without reference to a location of a previously located centroid.

49. The system of claim 32, further characterized in that step (b)(iii) is performed using a location of a previously located centroid as a start estimate.

50. The system of claim 32, further characterized in that step (b)(iii) is performed using, as a start estimate, a location which a spot would occupy in an absence of the wavefront aberation.

51. The system of Claim 34, further characterized in that it comprises interface means (54, 56) for allowing as operator to select the lenslets which do not provide usable data.

52. The system of claim 34, further characterized in that step (c)(i) comprises determining automatically which lenslets do not provide usable data.

53. The system of claim 52, further characterized in that step (c)(i) comprises determining automatically which lenslets do not provide usable data in accordance with standard deviations of intensity around the corresponding centroids.

54. The system of claim 52, further characterized in that step (c)(i) comprises determining automatically which lenslets do not provide usable data in accordance with overall intensity levels of the corresponding centroids.

55. The system of claim 34, further characterized in that:

the matrix has two rows corresponding to each of the lenslets; and step (c(=)(XII) comprises eliminating (186) both of the rows corresponding to each of the lenslets which do not provide usable data.

56. The system of claim 55, further characterized in that the two rows corresponding to each of the lenslets are a row corresponding to an x coordinate and a row corresponding to ay coordinate.

57. The system of clean 32, further characterized in that step (c) comprises calculating a number of Zernike modes by using a number of centroids which is at least twice the number of Zernike modes.

38. The system of claim 32, further characterized in that step (e) comprises calculating the wavefront aberration for a first pupil radius R0 of the eye and also for a second pupil radius R1 of the eye, wherein R1 < R0.

59. The system of Claim 38, further characterized in that the wavefront aberration is calculated for R1 by renormalizing the wavefront aberration calculated for R0.

60. The system of claim 38, further characterized in that it comprises interface means (54, 56) for prompting as operator for R1.

61. The system of claim 60, further characterized in that it comprises interface means (54, 56) for prompting an operator for a minimum Zernike mode and a maximum Zernike mode for use in calculating the wavefront aberration for R1.

62. The system of claim 34, further characterized is that step (d) is performed using singular value decomposition.

63. A method of measuring a wavefront aberration of an eye, the method comprising:
(a) taking image data from the eye, the image data comprising a plurality of spots (88) having positions determined by the wavefront aberration;
(b) defining a plurality of search boxes (100) in the image data;
(c) determining the positions (116) of the spots (88); and (d) calculating the wavefront aberration in accordance with the positions (116) of the spots (88);
characterized in that step (e) comprises:
(i) locating a centroid (102, 114, 116) in each of the scotch boxes (100);

(ii) replacing each of the search boxes (100) with a search box (106, 112,118) of reduced physical size, the search box of reduced physical size (106, 112, 118) being clipped to fit entirely within the search box (100) which is replaced;
(iii) locating the centroid (102, 114, 116) in each of the search boxes (106, 112, 118) of reduced physical size; and (iv) repeating steps (ii) and (iii) until each of the search boxes (106, 112, 118) of reduced physical size reaches a minimum size (118).

64. The method of claim 63, characterized in that step (c)(ii) comprises forming the search box (106, 112, 118) of reduced physical size by:

(A) reducing a size of the search box (100, 106, 112) to be replaced by one pixel in each direction to form a pixel-reduced search box (104, 110);
(B) centering the pixel-reduced search box (104, 110) on the centroid (102) located in step (c)(i); and (C) clipping the pixel-reduced search box (104, 110) to the boundaries of the search box to be replaced (100, 106, 112) to fit entirely within the search box to be replaced (104, 106, 112).

65. The method of claim 63, further characterized in that the minimum size in step (c)(iv) is a diffraction limited size.

66. The method of claim 63, further characterized in that steps (c)(i) and (c)(iii) are performed only in accordance with pixels in the image data whose intensities lie between a lower threshold and an upper threshold.

67. The method of claim 66, further characterized in that it comprises prompting an operator to select the lower threshold and tho upper threshold.

68. A method of measuring a wavefront aberration of an eye, the method comprising:
(a) taking image data from the eye, the image data comprising a plurality of spots (88) having positions determined by the wavefront aberration;
(b) defining a plurality of search boxes (74, 74, 80, 82) in the image data;
(c) determining the positions (96) of the spots (88);
(d) calculating the wavefront aberration in accordance with the positions (96) of the spots (88) characterized io that step (b) comprises:
(e) defining a central search box (70);
(ii) defining a first set of search boxes (74, 78, 82) relative to the central search box (70);
(iii) locating centroids (84, 86) in the first set of search boxes, and (iv) defining successive sets of search boxes (74, 78, 82), each successive set defined in accordance with locations of centroids (84, 86) in a previous set;
and is that step (c) comprises:
(i) locating a centroid (90) in each of the search boxes;
(ii) replacing each of the search boxes (86) with a search box (92, 98) of reduced physical size;
(iii) locating the centroid (94, 96) in each of the search boxes (92, 98) of reduced physical size; and (iv) repeating steps (11) and (iii) until each of the search boxes (92, 98) of reduced physical size reaches a minimum size.

69. The method of claim 68, further characterized in that the central search box (14) is defined about a center point (72) which is a position which one of the spots (88) at a center of the image data would occupy its as absence of the wavefront aberration.

70. The method of claim 68, further characterized in that step (a) comprises forming the spots (88) by a plurality of lenslets (18) having a lenslet spacing C.

71. The method of claim 70, further characterized in that at least some of the plurality of search boxes have linear dimensions scaled from C by a factor of less than one.

72. The method of claim 71, further characterized in that it comprises prompting an operator for the factor.

73. The method of claim 70, further characterized is that the central search box (70) is defined about a center point (72) which is a position which one of the spots (88) at a center of the image would occupy in an absence of the wavefront aberration.

74. The method of claim 73, further characterized in that:
the first set of search boxes (74, 78, 82) defined in sup (b)(ii) have centers spaced from the center point (72) of the central search box (70) by C; and each successive set of search boxes (74, 78, d2) defined is step (b)(iv) have concern spaced from the centroids (84, 86) is the previous set by C.

75. The method of claim 68, further characterized in that, is step (c)(ii), the search box of reduced physical aim (106) is clipped to fit entirely within the search box which is replaced (100).

76. A method of measuring a wavefront aberration of an eye, the method comprising:
(a) taking image data from the eye, the image data comprising a plurality of spots (88) having positions determined by the wavefront aberration;
(b) defining a plurality of search boxes (86) in the image data;
(c) determining the positions (96) of the spots (88); and (d) calculating the wavefront aberration in accordance with the positions (96) of the spots (88);
characterized is that step (s) comprises correcting (16, 60) for a spatial inhomogeneity is a retina of the eye and in that step (c) comprises:
(i) locating a centroid (90) is each of the search boxes (86);
(ii) replacing each of the search boxes (E6) with a search box (92, 98) of reduced physical size:
(iii) locating the centroid (94, 96) in each of the search boxes (92, 98) of reduced physical size; and (iv) repeating steps (ii) and (iii) until each of the search boxes (92, 98) of reduced physical size reaches a minimum size.

77. The method of claim 76, further characterized is that step (a) comprises illuminating (16) the eye with near infrared light.

78. The method of claim 76, further characterized is that said step of correcting for the spatial inhomogeneity of the retina of the eye comprises illuminating (16) the eye with light of a coherence length selected such that the spatial inhomogeneity does nor appear is the image data.

79. The method of claim 78, further characterized is that the coherence length is 30 microns.

80. The method of claim 78, further characterized is that said step of correcting for the spatial inhomogeneity of the retina of the eye comprises:

(i) providing an optical scanning device (60);
(ii) illuminating the eye with light which has been scanned by the optical scanning device (60); and (iii) causing the light emerging from the eye to be scanned again by the optical scanning device (60) and subsequently made incident on a detector (20) such that the light is stationary as the light is made incident on the detector (20).

81. The method of claim 80, further characterized in that:
the detector (20) comprises a camera (20) having a frame time; and the optical scanning device (60) performs a plurality of scans within a single one of said frame time.

82. The method of claim 81, further characterized in that the optical scanning device (60) has a scanning frequency of 400-600 Hz.

83. The method of claim 81, further characterized in that the optical scanning device (60) scans with an angular agent of lass than 0.5 degrees.

84. The method of claim 80, further characterized in that the optical scanning device comprises a single scanning/de-scanning mirror (60) which scans both the light which illuminates the eye and the light emerging from the eye.

85. A method of measuring a wavefront aberration of an eye, the method comprising:
(a) taking image data from the eye, the image data comprising a plurality of spots (88) having positions determined by the wavefront aberration;
(b) defining a plurality of search boxes (100) in the image data;
(c) determining the positions (116) of the spots (88); and (d) calculating the wavefront aberration in accordance with the positions (116) of the spots (88);
characterized in that step (a) comprises taking image data from a plurality of locations in the eye, the image data comprising a plurality of spots (88) corresponding to the plurality of locations and having positions determined by the wavefront aberration, the data from the plurality of locations in the eye being taken by scanning light into the era such that the plurality of spots are separated in time, and in that step (c) comprises:
(i) locating a centroid (102) in each of the search boxes (100);

(ii) replacing each of the search boxes (100) with a search box (106,112, 118) of reduced physical size, the search box (106, 112, 118) of reduced physical size being clipped to fit entirely within the search box (100) which is replaced;
(iii) locating the centroid (102, 114, 116) in each of the search boxes (106, 112, 118) of reduced physical size; and (iv) repeating steps (ii) and (iii) until each of the search boxes (106, 112,118) of reduced physical size reaches a minimum size (118).

86. A device for measuring a wavefront aberration of an eye, the device comprising:
image data taking means (20) for taking image data from the eye, the image data comprising a plurality of spots (88) having positions determined by the wavefront aberration;
and data processing moans (12), receiving the image data, for:
(a) defining a plurality of search boxes (100) in the image data;
(b) determining the positions (116) of the spots (88); and (e) calculating the wavefront aberration in accordance with the positions (116) of the spots (88):
characterized in that the data processing means (12) performs step (b) by:
(i) locating a centroid (102) in each of the search boxes (100);
(ii) replacing each of the search boxes (100) with a search box (106, 112, 118) of reduced physical size, the search box (106, 112, 118) of reduced physical size being clipped to fit entirely within the search box (100) which is replaced;
(iii) locating the centroid (114, 116) in each of the search boxes (106, 112, 118) of reduced physical size; and (iv) repeating steps (ii) and (iii) until each of the search boxes (106, 112, 118) of reduced physical size reaches a minimum size (118).

87. The device of claim 86, further characterized is that the data processing means performs step (b)(ii) by forming the search box ( 106, 112, 118) of reduced physical size by:
(A) reducing a size of the search box (100) to be replaced by one pixel in each direction to form a pixel-reduced search box (104).

(B) centering the pixel-reduced search box (104) on the centroid (102) located in step (c) (~); and (C) clipping the pixel-reduced search box (104) to the boundaries of the search box to be replaced (100) to fit entirely within the search box (100) to be replaced.

88. The device of claim 86, further characterized in that the minimum size in step (b)(iv) is a diffraction limited size.

89. The device of claim 86, further characterized in that steps (b)(i) and (b)(iii) are performed only in accordance with pixels in the image data whose intensities lie between a lower threshold said an upper threshold.

90. The device of claim 89, further characterized in that the data processing means (12) comprises interface means (54, 56) for prompting an operator to select the lower threshold and the upper threshold.

91. A device for measuring a wavefront aberration of an eye, the device comprising:
image data taking means (20) for taking image data from the eye, the image data comprising a plurality of spots (88) having position determined by the wavefront aberration;
end data processing means (12), receiving the image data, for:
(a) defining a plurality of search boxes (70, 74, 78, 82) in the image data;
(b) determining the position (116) of the spots (88); and (c) calculating the wavefront aberration in accordance with the positions (116) of the spots (88);
characterized in that the data processing means (12) performs slap (a) by:
(i) defining a central search box (70);
(ii) defining a first set of search boxes (74, 78, 82) relative to the central search box (70);
(iii) locating centroids (84, 86) in the first set of search boxes (74, 78, 82); and (iv) defining successive sets of search boxes (74, 78, 82), each successive set defined in accordance with locations of centroids (84, 86) is a previous set;
and in that the data processing means (12) performs step (b) by:
(i) locating a centroid (90) in each of the search boxes (74, 78, 82, 86);

(ii) replacing each of the search boxes (86) with a search box (92, 98) of reduced physical size;
(iii) locating the centroid (94, 96) in each of the search boxes (92, 98) of reduced physical; and (iv) repeating steps (ii) and (iii) until each of the search boxes (92, 98) of reduced physical size reaches a minimum size.

92. The device of claim 91, further characterized in that the central search box (70) b defined about a center point (72) which is a position which one of the spots (88) at a center of the image data would occupy in an absence of the wavefront aberration.

93. The device of claim 91, further characterized in that the image data taking means (10) comprises a plurality of lenslets (18) having a lenslet spacing C for forming the spots.

94. The device of claim 93, further characterized in that at least some of the plurality of search boxes (70, 74, 78, 82) have linear dimensions scaled from C by a factor of less than one.

95. The device of claim 94, further characterized in that the data processing means (12) comprises interface means (54, 56) for prompting an operator for the factor.

96. The device of claim 93, further characterized in that the central search box (70) is defined about a center point (72) which is a position which one of the spots (88) at a center of the image would occupy in an absence of the wavefront aberration.

97. The device of claim 96, further characterized in that the fast set of search boxes (74, 78, 82) defined ist step (a)(ii) have centers spaced from the center point (72) of the central search box (70) by C; and each successive set of search boxy (74, 78, 82) defined in step (b)(iv) have centers spaced from the centroids (84, 86) in the previous set by C.

98. The device of claim 91, further characterized in that, in step (b)(ii), the search box (106) of reduced physical size is clipped to fit entirely within the search box (100) which is replaced.

99. A device for measuring a wavefront aberration of an eye, the device comprising:
image data taking means (20) for taking image data from the eye, the image data comprising a plurality of spots (88) having position (96) determined by the wavefront aberration; and data processing means (12), receiving the image data, for:

(a) defining a plurality of search boxes (86) in the image data;
(b) determining the positions (96) of the spots (88); and (c) calculating the wavefront aberration in accordance with the positions (96) of the spots (88);
characterized in that the image data taking means (20) comprises means (16, 60) for correcting for a spatial inhomogeneity in a retina of the eye and in that step (b) comprises:
(i) locating a centroid (90) in each of the search boxes (86);
(ii) replacing each of the starch boxes (86) with a search box (92, 98) of reduced physical size;
(iii) locating the centroid (94, 96) in each of the search boxes (92, 98) of reduced physical size; and (iv) replacing sips (ii) and (iii) until each of the search boxes (92, 98) of reduced physical size reaches a minimum size.

100. The device of claim 99, further characterized in that the image data taking means (20) comprises means (16) for illuminating the eye with near infrared light.

101. The device of claim 99, further characterized in that the image data taking means (20) comprises means (16) for illuminating the eye with light of a coherence length selected such that the spatial inhomogeneity does not appear in the image data.

102. The device of claim 101, further characterized in that the coherence length is 30 microns.

103. The device of claim 101, further characterized in that the image data taking means (20) comprises:
a detector (20); and an optical scanning device (60) for illuminating the eye with light which has been scanned by the optical scanning device (60) and scanning again the light emerging from the eye, the light emerging from the eye and scanned again being subsequently made incident on the detector (20) such that the light is stationary as the light is made incident on the detector (60).

104. The device of claim 103, further characterized in that:
the detector (20) comprises a camera (20) having a frame time; and the optical scanning device (60) performs a plurality of scans within a single one of said frame time.

105. The device of claim 104, further characterized is that the optical scanning device (60) has a scanning frequency of 400-400 Hz.

146. The device of claim 104, further characterized in that the optical scanning device (60) scans with an angular extent of less than 0.5 degrees.

107. The device of claim 103, further characterized in that the optical scanning device (60) comprises a single scanning/de-scanning mirror (60) which scans both the light which illuminates the eye and the light emerging from the eye.

108. A device for measuring a wavefront aberration of an eye, the device comprising:
image data taking means (20) for taking image data from a plurality of locations in the eye, the image data comprising a plurality of spots (88) corresponding to the plurality of locations and having positions (116) determined by the wavefront aberration;
and data processing means (12), receiving the image data, for:
(a) defining a plurality of search boxes (100) in the image data;
(b) determining the positions (116) of the spots (88); and (c) calculating the wavefront aberration in accordance with the position (116) of the spots (118);
characterized in that the image data taking mans (20) comprises a scanner, the data from the plurality of locations in the eye being taken by scanning light into the eye with the scanner such that the plurality of spots are separated in time, and in that step (b) comprises:
(i) locating a centroid (102) in each of the search boxes (100;
(ii) replacing each of the search boxes (100) with a search box (106, 112, 118) of reduced physical sue, the search box (106, 112, 118) of reduced physical size being clipped to fit entirely within the search box (100) which is replaced;
(iii) locating the centroid (114, 116) in each of the search boxes (106,112, 118) of reduced physical size; and (iv) repeating steps (ii) and (iii) until each of the search boxes (106, 112, 118) of reduced physical size reaches a minimum size (118).
CA002388719A 1999-10-21 2000-10-20 Rapid, automatic measurement of the eye's wave aberraion Abandoned CA2388719A1 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US09/421,892 1999-10-21
US09/421,892 US6199986B1 (en) 1999-10-21 1999-10-21 Rapid, automatic measurement of the eye's wave aberration
PCT/US2000/029078 WO2001028409A1 (en) 1999-10-21 2000-10-20 Rapid, automatic measurement of the eye's wave aberration

Publications (1)

Publication Number Publication Date
CA2388719A1 true CA2388719A1 (en) 2001-04-26

Family

ID=23672505

Family Applications (1)

Application Number Title Priority Date Filing Date
CA002388719A Abandoned CA2388719A1 (en) 1999-10-21 2000-10-20 Rapid, automatic measurement of the eye's wave aberraion

Country Status (7)

Country Link
US (2) US6199986B1 (en)
EP (1) EP1221889A1 (en)
JP (1) JP4660047B2 (en)
AU (1) AU1339101A (en)
BR (1) BR0015231A (en)
CA (1) CA2388719A1 (en)
WO (1) WO2001028409A1 (en)

Families Citing this family (178)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
UA67870C2 (en) 2002-10-04 2004-07-15 Сергій Васильович Молебний Method for measuring wave aberrations of eyes
US7303281B2 (en) * 1998-10-07 2007-12-04 Tracey Technologies, Llc Method and device for determining refractive components and visual function of the eye for vision correction
DE19904753C1 (en) * 1999-02-05 2000-09-07 Wavelight Laser Technologie Gm Device for photorefractive corneal surgery of the eye for correcting high-order visual defects
US6619799B1 (en) * 1999-07-02 2003-09-16 E-Vision, Llc Optical lens system with electro-active lens having alterably different focal lengths
DE19938203A1 (en) 1999-08-11 2001-02-15 Aesculap Meditec Gmbh Method and device for correcting visual defects in the human eye
DE19950792A1 (en) * 1999-10-21 2001-04-26 Technolas Gmbh Ophthalmic wavefront aberration diagnostic tool, has camera that aids in focusing aerial image from lenslet array, on wavefront sensor
US6477273B1 (en) 1999-10-21 2002-11-05 3M Innovative Properties Company Centroid integration
US6633669B1 (en) 1999-10-21 2003-10-14 3M Innovative Properties Company Autogrid analysis
US6563947B1 (en) * 1999-12-10 2003-05-13 20/10 Perfect Vision Optische Geraete Gmbh Application specified integrated circuit for use in wavefront detection
WO2001050084A1 (en) * 1999-12-30 2001-07-12 Ade Corporation Specimen topography reconstruction
US7455407B2 (en) * 2000-02-11 2008-11-25 Amo Wavefront Sciences, Llc System and method of measuring and mapping three dimensional structures
US6550917B1 (en) 2000-02-11 2003-04-22 Wavefront Sciences, Inc. Dynamic range extension techniques for a wavefront sensor including use in ophthalmic measurement
US6659613B2 (en) * 2000-03-27 2003-12-09 Board Of Regents, The University Of Texas System Methods and systems for measuring local scattering and aberration properties of optical media
JP4491663B2 (en) * 2000-03-28 2010-06-30 株式会社トプコン Ophthalmic optical characteristic measuring device
US6565209B2 (en) * 2000-04-25 2003-05-20 Alcon Universal Ltd. Range-extending system and spatial filter for enhancing Hartmann-Shack images and associated methods
US6460997B1 (en) 2000-05-08 2002-10-08 Alcon Universal Ltd. Apparatus and method for objective measurements of optical systems using wavefront analysis
CA2409663A1 (en) * 2000-05-09 2001-11-15 Memphis Eye & Cataract Associates Ambulatory Surgery Center (Dba Meca La Laser And Surgery Center) Method and system for control of high resolution high speed digital micromirror device for laser refractive eye surgery
US8020995B2 (en) * 2001-05-23 2011-09-20 Amo Groningen Bv Methods of obtaining ophthalmic lenses providing the eye with reduced aberrations
US6685317B2 (en) 2000-06-13 2004-02-03 Massie Research Laboratories, Inc. Digital eye camera
US6486943B1 (en) * 2000-09-19 2002-11-26 The Schepens Eye Research Institute, Inc. Methods and apparatus for measurement and correction of optical aberration
CA2422855C (en) 2000-09-21 2012-01-03 Visx, Inc. Enhanced wavefront ablation system
US6616279B1 (en) * 2000-10-02 2003-09-09 Johnson & Johnson Vision Care, Inc. Method and apparatus for measuring wavefront aberrations
DE60121976D1 (en) * 2000-10-10 2006-09-14 Univ Rochester DETERMINATION OF THE OKULAR REFECTION BY WAVE FRONT BRAKING DATA
JP4598261B2 (en) * 2000-11-16 2010-12-15 株式会社トプコン Ophthalmic optical characteristic measuring device
US7011410B2 (en) * 2000-11-22 2006-03-14 Eyetect, L.L.C. Method and apparatus for monitoring eye tremor
MXPA03005113A (en) * 2000-12-08 2004-01-29 Visx Inc Direct wavefront-based corneal ablation treatment program.
US6547391B2 (en) 2000-12-08 2003-04-15 Johnson & Johnson Vision Care, Inc. Ocular aberration correction taking into account fluctuations due to biophysical rhythms
US6331059B1 (en) * 2001-01-22 2001-12-18 Kestrel Corporation High resolution, multispectral, wide field of view retinal imager
UA59488C2 (en) 2001-10-03 2003-09-15 Василь Васильович Молебний Method for measuring wave aberrations of eye and device for its realization (variants)
EP1367935B1 (en) 2001-03-15 2008-09-17 AMO WaveFront Sciences, LLC Tomographic wavefront analysis system and method of mapping an optical system
EP1379158B1 (en) * 2001-04-16 2015-01-21 Tracey Technologies, Llc Method for determining clinical refraction of eye from objective source
MXPA03009496A (en) 2001-04-18 2004-02-12 Bausch & Lomb Objective manifest refraction.
JP2004534964A (en) * 2001-04-27 2004-11-18 ノバルティス アクチエンゲゼルシャフト Automatic lens design and manufacturing system
US6582079B2 (en) * 2001-06-05 2003-06-24 Metrologic Instruments, Inc. Modular adaptive optical subsystem for integration with a fundus camera body and CCD camera unit and improved fundus camera employing same
US6961476B2 (en) * 2001-07-27 2005-11-01 3M Innovative Properties Company Autothresholding of noisy images
AU2002318905A1 (en) * 2001-07-30 2003-02-17 Lasersight Technologies, Inc. Determining a target optical zone for abalation based on the stiles-crawford effect
EP1427328B1 (en) * 2001-08-30 2014-07-02 University Of Rochester Adaptive optics in a scanning lase ophtalmoscope
US6631991B2 (en) * 2001-08-31 2003-10-14 Adaptive Optics Associates, Inc. Ophthalmic instrument having hartmann wavefront sensor deriving location of spots with spot fitting techniques
US6575572B2 (en) 2001-09-21 2003-06-10 Carl Zeiss Ophthalmic Systems, Inc. Method and apparatus for measuring optical aberrations of an eye
US6724479B2 (en) 2001-09-28 2004-04-20 Infineon Technologies Ag Method for overlay metrology of low contrast features
US6712466B2 (en) * 2001-10-25 2004-03-30 Ophthonix, Inc. Eyeglass manufacturing method using variable index layer
JP3869345B2 (en) * 2001-11-09 2007-01-17 株式会社トプコン Eye characteristics measuring device
JP3978024B2 (en) * 2001-12-03 2007-09-19 株式会社ニデック Ophthalmic device and corneal surgery device
US7034949B2 (en) * 2001-12-10 2006-04-25 Ophthonix, Inc. Systems and methods for wavefront measurement
US6781681B2 (en) 2001-12-10 2004-08-24 Ophthonix, Inc. System and method for wavefront measurement
US6666857B2 (en) * 2002-01-29 2003-12-23 Robert F. Smith Integrated wavefront-directed topography-controlled photoablation
US6952435B2 (en) * 2002-02-11 2005-10-04 Ming Lai Speckle free laser probe beam
CA2475896C (en) * 2002-02-11 2011-08-23 Visx, Inc. Determining relative positional and rotational offsets
US20050174535A1 (en) * 2003-02-13 2005-08-11 Lai Shui T. Apparatus and method for determining subjective responses using objective characterization of vision based on wavefront sensing
US6761454B2 (en) 2002-02-13 2004-07-13 Ophthonix, Inc. Apparatus and method for determining objective refraction using wavefront sensing
US7083609B2 (en) * 2002-06-13 2006-08-01 Visx, Incorporated Corneal topography-based target warping
US7322694B2 (en) * 2002-09-06 2008-01-29 Synergeyes, Inc. Hybrid contact lens system and method
US7163292B2 (en) * 2002-09-06 2007-01-16 Synergeyes, Inc. Hybrid contact lens system and method
US7097301B2 (en) * 2002-09-06 2006-08-29 Synergeyes, Inc. Hybrid contact lens system and method
US7104648B2 (en) * 2002-09-06 2006-09-12 Synergeyes, Inc. Hybrid contact lens system and method
SE0203564D0 (en) 2002-11-29 2002-11-29 Pharmacia Groningen Bv Multifocal opthalmic lens
US7896916B2 (en) 2002-11-29 2011-03-01 Amo Groningen B.V. Multifocal ophthalmic lens
US6736510B1 (en) 2003-02-04 2004-05-18 Ware Tec Vision Systems, Inc. Ophthalmic talbot-moire wavefront sensor
US7414712B2 (en) 2003-02-13 2008-08-19 University Of Rochester Large dynamic range Shack-Hartmann wavefront sensor
US20040227932A1 (en) * 2003-02-13 2004-11-18 Geunyoung Yoon Large dynamic range shack-hartmann wavefront sensor
US6988801B2 (en) * 2003-03-25 2006-01-24 University Of Rochester Compact portable wavefront sensor
US7556378B1 (en) 2003-04-10 2009-07-07 Tsontcho Ianchulev Intraoperative estimation of intraocular lens power
US7057806B2 (en) * 2003-05-09 2006-06-06 3M Innovative Properties Company Scanning laser microscope with wavefront sensor
US20040263779A1 (en) * 2003-06-12 2004-12-30 Visx, Inc. Hartmann-Shack wavefront measurement
US7175278B2 (en) * 2003-06-20 2007-02-13 Visx, Inc. Wavefront reconstruction using fourier transformation and direct integration
US7168807B2 (en) 2003-06-20 2007-01-30 Visx, Incorporated Iterative fourier reconstruction for laser surgery and other optical applications
US7018039B2 (en) * 2003-11-14 2006-03-28 Synergeyes,Inc. Contact lens
US7481536B2 (en) * 2004-02-19 2009-01-27 Amo Manufacturing Usa, Llc Methods and systems for differentiating left and right eye images
ES2665536T3 (en) 2004-04-20 2018-04-26 Alcon Research, Ltd. Integrated surgical microscope and wavefront sensor
CN100430031C (en) * 2004-04-26 2008-11-05 中国科学院光电技术研究所 Vision simulation system with stereopsis higher-order aberration correction function for double eyes
US7487145B1 (en) 2004-06-22 2009-02-03 Google Inc. Method and system for autocompletion using ranked results
US7836044B2 (en) 2004-06-22 2010-11-16 Google Inc. Anticipated query generation and processing in a search engine
FR2872590B1 (en) * 2004-07-02 2006-10-27 Essilor Int METHOD FOR PRODUCING AN OPHTHALMIC GLASS AND OPTICAL COMPONENT SUITABLE FOR CARRYING OUT SAID METHOD
EP1997422B1 (en) * 2004-09-01 2013-11-13 WaveLight GmbH Quick wave front measurement
US7922326B2 (en) 2005-10-25 2011-04-12 Abbott Medical Optics Inc. Ophthalmic lens with multiple phase plates
US9801709B2 (en) 2004-11-02 2017-10-31 E-Vision Smart Optics, Inc. Electro-active intraocular lenses
US8915588B2 (en) 2004-11-02 2014-12-23 E-Vision Smart Optics, Inc. Eyewear including a heads up display
US8778022B2 (en) 2004-11-02 2014-07-15 E-Vision Smart Optics Inc. Electro-active intraocular lenses
US7499940B1 (en) 2004-11-11 2009-03-03 Google Inc. Method and system for URL autocompletion using ranked results
US20060106769A1 (en) 2004-11-12 2006-05-18 Gibbs Kevin A Method and system for autocompletion for languages having ideographs and phonetic characters
US7594727B2 (en) * 2004-12-03 2009-09-29 Searete Llc Vision modification with reflected image
US7350919B2 (en) * 2004-12-03 2008-04-01 Searete Llc Vision modification with reflected image
US7344244B2 (en) * 2004-12-03 2008-03-18 Searete, Llc Adjustable lens system with neural-based control
US7334892B2 (en) * 2004-12-03 2008-02-26 Searete Llc Method and system for vision enhancement
US7931373B2 (en) * 2004-12-03 2011-04-26 The Invention Science Fund I, Llc Vision modification with reflected image
US7390088B2 (en) * 2004-12-03 2008-06-24 Searete Llc Adjustable lens system with neural-based control
US8244342B2 (en) * 2004-12-03 2012-08-14 The Invention Science Fund I, Llc Method and system for adaptive vision modification
US9155483B2 (en) * 2004-12-03 2015-10-13 The Invention Science Fund I, Llc Vision modification with reflected image
US7334894B2 (en) * 2004-12-03 2008-02-26 Searete, Llc Temporal vision modification
US7656569B2 (en) * 2004-12-03 2010-02-02 Searete Llc Vision modification with reflected image
US7486988B2 (en) * 2004-12-03 2009-02-03 Searete Llc Method and system for adaptive vision modification
US7470027B2 (en) * 2004-12-03 2008-12-30 Searete Llc Temporal vision modification
US8104892B2 (en) * 2004-12-03 2012-01-31 The Invention Science Fund I, Llc Vision modification with reflected image
US20060126018A1 (en) * 2004-12-10 2006-06-15 Junzhong Liang Methods and apparatus for wavefront sensing of human eyes
FR2879757B1 (en) * 2004-12-17 2007-07-13 Essilor Int METHOD FOR PRODUCING A TRANSPARENT OPTICAL ELEMENT, OPTICAL COMPONENT INVOLVED IN THIS METHOD AND OPTICAL ELEMENT THUS OBTAINED
US7401922B2 (en) * 2005-04-13 2008-07-22 Synergeyes, Inc. Method and apparatus for reducing or eliminating the progression of myopia
US7543936B2 (en) * 2005-05-06 2009-06-09 Synergeyes, Inc. Hybrid contact lens system and method of fitting
US7268937B1 (en) 2005-05-27 2007-09-11 United States Of America As Represented By The Secretary Of The Air Force Holographic wavefront sensor
WO2007002487A2 (en) 2005-06-24 2007-01-04 Boston Foundation For Sight Scleral contact lens with grooves and method of making lens
FR2888951B1 (en) * 2005-07-20 2008-02-08 Essilor Int RANDOMIZED PIXELLIZED OPTICAL COMPONENT, METHOD FOR MANUFACTURING THE SAME, AND USE THEREOF IN THE MANUFACTURE OF A TRANSPARENT OPTICAL ELEMENT
FR2888948B1 (en) 2005-07-20 2007-10-12 Essilor Int PIXELLIZED TRANSPARENT OPTIC COMPONENT COMPRISING AN ABSORBENT COATING, METHOD FOR PRODUCING THE SAME AND USE THEREOF IN AN OPTICAL ELEMENT
FR2888947B1 (en) * 2005-07-20 2007-10-12 Essilor Int OPTICAL CELL COMPONENT
US7331674B2 (en) * 2005-09-02 2008-02-19 Visx, Incorporated Calculating Zernike coefficients from Fourier coefficients
US7377637B2 (en) * 2005-10-11 2008-05-27 Synergeyes, Inc. Hybrid contact lens system and method of fitting
DE102005054691A1 (en) * 2005-11-16 2007-05-24 Bausch & Lomb Inc. Method and device for determining the visual acuity of an eye
US8010523B2 (en) * 2005-12-30 2011-08-30 Google Inc. Dynamic search box for web browser
US9101292B2 (en) 2006-01-20 2015-08-11 Clarity Medical Systems, Inc. Apparatus and method for operating a real time large dipoter range sequential wavefront sensor
US8777413B2 (en) 2006-01-20 2014-07-15 Clarity Medical Systems, Inc. Ophthalmic wavefront sensor operating in parallel sampling and lock-in detection mode
US8820929B2 (en) 2006-01-20 2014-09-02 Clarity Medical Systems, Inc. Real-time measurement/display/record/playback of wavefront data for use in vision correction procedures
US7445335B2 (en) * 2006-01-20 2008-11-04 Clarity Medical Systems, Inc. Sequential wavefront sensor
US8356900B2 (en) 2006-01-20 2013-01-22 Clarity Medical Systems, Inc. Large diopter range real time sequential wavefront sensor
US8100530B2 (en) 2006-01-20 2012-01-24 Clarity Medical Systems, Inc. Optimizing vision correction procedures
US9113819B2 (en) 2006-01-20 2015-08-25 Clarity Medical Systems, Inc. Apparatus and method for operating a real time large diopter range sequential wavefront sensor
EP1999443B1 (en) * 2006-03-14 2017-12-27 AMO Manufacturing USA, LLC Spatial frequency wavefront sensor system and method
US7780294B2 (en) * 2006-03-23 2010-08-24 Amo Manufacturing Usa, Llc. Systems and methods for wavefront reconstruction for aperture with arbitrary shape
JP4794341B2 (en) * 2006-03-30 2011-10-19 株式会社トプコン Ophthalmic optical characteristic measuring device
US7758189B2 (en) 2006-04-24 2010-07-20 Physical Sciences, Inc. Stabilized retinal imaging with adaptive optics
EP2012653B1 (en) * 2006-05-01 2012-12-12 Physical Sciences, Inc. Hybrid spectral domain optical coherence tomography line scanning laser ophthalmoscope
US7537339B2 (en) * 2006-05-25 2009-05-26 Synergeyes, Inc. Hybrid contact lens system and method of fitting
EP2040697A2 (en) * 2006-07-03 2009-04-01 Ranbaxy Laboratories Limited Polymorphic form of duloxetine hydrochloride
US20080074611A1 (en) * 2006-09-22 2008-03-27 Meyers William E Hybrid contact lens with improved resistance to flexure and method for designing the same
US20080084541A1 (en) * 2006-10-06 2008-04-10 Ming Lai Ophthalmic system and method
US7942527B2 (en) * 2006-10-18 2011-05-17 Lawrence Livermore National Security, Llc Compact adaptive optic-optical coherence tomography system
FR2907559B1 (en) * 2006-10-19 2009-02-13 Essilor Int ELECRO-COMMANDABLE OPTICAL COMPONENT COMPRISING A SET OF CELLS
JP4988305B2 (en) * 2006-10-31 2012-08-01 株式会社ニデック Ophthalmic measuring device
FR2910642B1 (en) * 2006-12-26 2009-03-06 Essilor Int TRANSPARENT OPTICAL COMPONENT WITH TWO CELL ARRAYS
FR2911404B1 (en) * 2007-01-17 2009-04-10 Essilor Int TRANSPARENT OPTICAL COMPONENT WITH CELLS FILLED WITH OPTICAL MATERIAL
AR064985A1 (en) 2007-01-22 2009-05-06 E Vision Llc FLEXIBLE ELECTROACTIVE LENS
US8215770B2 (en) 2007-02-23 2012-07-10 E-A Ophthalmics Ophthalmic dynamic aperture
EP2150169B1 (en) 2007-05-17 2016-04-06 AMO Development, LLC Customized laser epithelial ablation systems
US7828432B2 (en) * 2007-05-25 2010-11-09 Synergeyes, Inc. Hybrid contact lenses prepared with expansion controlled polymeric materials
US8740381B2 (en) * 2007-06-27 2014-06-03 Bausch & Lomb Incorporated Method and apparatus for extrapolating diagnostic data
US8333474B2 (en) 2007-10-19 2012-12-18 Wavetec Vision Systems, Inc. Optical instrument alignment system
US7594729B2 (en) 2007-10-31 2009-09-29 Wf Systems, Llc Wavefront sensor
US7874674B2 (en) * 2007-12-12 2011-01-25 Allred Lloyd G Aberrometer having reduced noise
DE102008014294A1 (en) * 2008-03-14 2009-09-17 Bausch & Lomb Inc. Fast algorithm for wavefront data stream
CA2718521A1 (en) 2008-03-18 2009-09-24 Pixeloptics, Inc. Advanced electro-active optic device
US9335563B2 (en) 2012-08-31 2016-05-10 Amo Groningen B.V. Multi-ring lens, systems and methods for extended depth of focus
EP2306889A4 (en) * 2008-07-10 2012-11-28 Univ Indiana Res & Tech Corp Ophthalmic apparatuses, systems and methods
US8312032B2 (en) 2008-07-10 2012-11-13 Google Inc. Dictionary suggestions for partial user entries
US8254724B2 (en) * 2008-11-06 2012-08-28 Bausch & Lomb Incorporated Method and apparatus for making and processing aberration measurements
WO2010054268A2 (en) 2008-11-06 2010-05-14 Wavetec Vision Systems, Inc. Optical angular measurement system for ophthalmic applications and method for positioning of a toric intraocular lens with increased accuracy
US7695135B1 (en) 2008-11-11 2010-04-13 Boston Foundation For Sight Scleral lens with scalloped channels or circumferential fenestrated channels
US7980698B2 (en) 2008-11-19 2011-07-19 Bausch & Lomb Incorporated Power-adjusted aberrometer
US8201943B2 (en) * 2009-01-15 2012-06-19 Physical Sciences, Inc. Adaptive optics line scanning ophthalmoscope
CN102483752A (en) 2009-06-03 2012-05-30 谷歌公司 Autocompletion for partially entered query
US8876290B2 (en) 2009-07-06 2014-11-04 Wavetec Vision Systems, Inc. Objective quality metric for ocular wavefront measurements
CN104367299B (en) 2009-07-14 2017-09-15 波技术视觉系统公司 Ophthalmic surgery measurement system
ES2524618T3 (en) * 2009-07-14 2014-12-10 Wavetec Vision Systems, Inc. Determination of the effective position of the lens of an intraocular lens using afractive refractive power
WO2011091253A2 (en) * 2010-01-21 2011-07-28 Physical Sciences, Inc. Multi-functional adaptive optics retinal imaging
WO2011139605A2 (en) * 2010-04-27 2011-11-10 Haddad Daniel S Dynamic real time active pupil centroid compensation
DK2592995T3 (en) 2010-07-16 2017-02-13 Zeiss Carl Vision Inc PUPIL DEPENDENT WAVE FRONT FRACTION
US9662004B2 (en) * 2011-04-29 2017-05-30 Taiwan Biophotonic Corporation Apparatus for non-invasive glucose monitoring
RU2502212C2 (en) * 2011-11-25 2013-12-20 Виктор Ильич Бетчер Method of forming image on display of digital device
WO2013082466A1 (en) 2011-11-30 2013-06-06 Amo Development, Llc. System and method for ophthalmic surface measurements based on sequential estimates
CA3167661A1 (en) 2012-01-06 2013-07-11 E-Vision Smart Optics, Inc. Eyewear docking station and electronic module
CN102607718B (en) * 2012-03-19 2013-11-13 中国科学院光电技术研究所 Hartmann wavefront sensor employing different-time exposure
CN102589720B (en) * 2012-03-19 2014-01-01 中国科学院光电技术研究所 Hartmann wavefront measuring instrument applicable to inhomogeneous light illumination
TWI588560B (en) 2012-04-05 2017-06-21 布萊恩荷登視覺協會 Lenses, devices, methods and systems for refractive error
EP2846679A1 (en) 2012-04-30 2015-03-18 Clarity Medical Systems, Inc. Ophthalmic wavefront sensor operating in parallel sampling and lock-in detection mode
US9554889B2 (en) 2012-05-07 2017-01-31 Boston Foundation For Sight Customized wavefront-guided methods, systems, and devices to correct higher-order aberrations
US9072462B2 (en) 2012-09-27 2015-07-07 Wavetec Vision Systems, Inc. Geometric optical power measurement device
US9201250B2 (en) 2012-10-17 2015-12-01 Brien Holden Vision Institute Lenses, devices, methods and systems for refractive error
CN104768499B (en) 2012-10-17 2017-06-23 华柏恩视觉研究中心 For ametropic eyeglass, device, method and system
JP6224908B2 (en) 2013-04-17 2017-11-01 キヤノン株式会社 Imaging device
CN103385691B (en) * 2013-08-02 2015-07-15 中国科学院光电技术研究所 Binocular adaptive optical vision simulator and simulation method
AU2017218681B2 (en) 2016-02-09 2021-09-23 Amo Groningen B.V. Progressive power intraocular lens, and methods of use and manufacture
EP3496682A1 (en) 2016-08-10 2019-06-19 AMO Development, LLC Epithelial ablation systems and methods
FR3062476B1 (en) * 2017-01-27 2020-12-25 Imagine Optic METHOD OF EVALUATING THE QUALITY OF MEASUREMENT OF A WAVEFRONT AND SYSTEMS IMPLEMENTING SUCH A METHOD
EP3595584A1 (en) 2017-03-17 2020-01-22 AMO Groningen B.V. Diffractive intraocular lenses for extended range of vision
JP7158827B2 (en) * 2017-04-21 2022-10-24 キヤノン株式会社 Ophthalmic imaging device and its control method
US20180341686A1 (en) * 2017-05-26 2018-11-29 Nanfang Hu System and method for data search based on top-to-bottom similarity analysis
US11523897B2 (en) 2017-06-23 2022-12-13 Amo Groningen B.V. Intraocular lenses for presbyopia treatment
CA3067116A1 (en) 2017-06-28 2019-01-03 Amo Groningen B.V. Diffractive lenses and related intraocular lenses for presbyopia treatment
AU2018292030B2 (en) 2017-06-28 2024-02-08 Amo Groningen B.V. Extended range and related intraocular lenses for presbyopia treatment
US11327210B2 (en) 2017-06-30 2022-05-10 Amo Groningen B.V. Non-repeating echelettes and related intraocular lenses for presbyopia treatment
CN109645956B (en) * 2018-12-25 2021-08-06 重庆远视科技有限公司 Eye diopter measuring device
AU2020416055A1 (en) 2019-12-30 2022-08-25 Amo Groningen B.V. Lenses having diffractive profiles with irregular width for vision treatment

Family Cites Families (60)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPS5838839A (en) 1981-08-31 1983-03-07 Tokyo Optical Co Ltd Method and apparatus for measuring refractive index
DE3150124C2 (en) 1981-12-18 1985-01-31 Fa. Carl Zeiss, 7920 Heidenheim Device for examining the anterior segments of the eye
DE3422144A1 (en) 1984-06-14 1985-12-19 Josef Prof. Dr. 6900 Heidelberg Bille DEVICE FOR DISPLAYING AREA AREAS OF THE HUMAN EYE
US4669466A (en) 1985-01-16 1987-06-02 Lri L.P. Method and apparatus for analysis and correction of abnormal refractive errors of the eye
WO1987005205A1 (en) 1986-03-08 1987-09-11 G. Rodenstock Instrumente Gmbh Laser installation for examining and treating the eye
US5177511A (en) 1986-11-08 1993-01-05 G. Rodenstock Instruments Gmbh Apparatus for producing images of an object and in particular for observing the rear portions of the eye
DE3733549A1 (en) 1987-10-03 1989-04-20 Messerschmitt Boelkow Blohm FIBER OPTICAL SENSOR
US4764930A (en) 1988-01-27 1988-08-16 Intelligent Surgical Lasers Multiwavelength laser source
US4848340A (en) 1988-02-10 1989-07-18 Intelligent Surgical Lasers Eyetracker and method of use
US4881808A (en) 1988-02-10 1989-11-21 Intelligent Surgical Lasers Imaging system for surgical lasers
DE3821973A1 (en) 1988-06-29 1990-02-08 Rodenstock Instr DEVICE FOR PRODUCING AN IMAGE OF AN OBJECT (III)
US4991953A (en) 1989-02-09 1991-02-12 Eye Research Institute Of Retina Foundation Scanning laser vitreous camera
US5159361A (en) 1989-03-09 1992-10-27 Par Technology Corporation Method and apparatus for obtaining the topography of an object
JP2931325B2 (en) 1989-06-29 1999-08-09 興和株式会社 Ophthalmic measurement device
DE4002029A1 (en) 1990-01-24 1991-07-25 Peter Hoefer METHOD FOR THE PRODUCTION OF CONTACT LENSES AND CONTACT LENS PRODUCTION SYSTEM
JP2854657B2 (en) 1990-03-14 1999-02-03 興和株式会社 Ophthalmic measurement device
US5062702A (en) 1990-03-16 1991-11-05 Intelligent Surgical Lasers, Inc. Device for mapping corneal topography
WO1992001417A1 (en) 1990-07-19 1992-02-06 Horwitz Larry S Vision measurement and correction
US5258791A (en) 1990-07-24 1993-11-02 General Electric Company Spatially resolved objective autorefractometer
AT395523B (en) 1990-07-24 1993-01-25 Spellitz Fritz OPTICAL MEASURING SYSTEM FOR THE HUMAN CORNUS
US5722427A (en) 1993-05-10 1998-03-03 Eyesys Technologies, Inc. Method of refractive surgery
US5139022A (en) 1990-10-26 1992-08-18 Philip Lempert Method and apparatus for imaging and analysis of ocular tissue
US5214456A (en) 1991-10-09 1993-05-25 Computed Anatomy Incorporated Mapping of corneal topography with display of pupil perimeter
JP3112108B2 (en) 1991-11-26 2000-11-27 株式会社トプコン Corneal thickness measuring device
US5229889A (en) 1991-12-10 1993-07-20 Hughes Aircraft Company Simple adaptive optical system
US5233174A (en) 1992-03-11 1993-08-03 Hughes Danbury Optical Systems, Inc. Wavefront sensor having a lenslet array as a null corrector
US5473392A (en) 1992-05-01 1995-12-05 Summit Technology, Inc. Method and system for topographic measurement
US5841511A (en) 1992-06-02 1998-11-24 Eyesys Technologies, Inc. Method of corneal analysis using a checkered placido apparatus
DE4222395A1 (en) 1992-07-08 1994-01-13 Amtech Ges Fuer Angewandte Mic Optical system for measuring refraction of eye with light source - has variable imaging optics adjustable in its power of refraction focussing light emanating from light source on retina of eye
US5307097A (en) 1992-11-05 1994-04-26 Kera-Metrics, Inc. Corneal topography system including single-direction shearing of holograph grating in orthogonal directions
US5418714A (en) 1993-04-08 1995-05-23 Eyesys Laboratories, Inc. Method and apparatus for variable block size interpolative coding of images
US5293871A (en) 1993-05-05 1994-03-15 Cornell Research Foundation Inc. System for ultrasonically determining corneal layer thicknesses and shape
US5452031A (en) 1993-05-05 1995-09-19 Boston Eye Technology, Inc. Contact lens and a method for manufacturing contact lens
JP3305410B2 (en) 1993-05-18 2002-07-22 株式会社トプコン Ophthalmic equipment
DE69414913T2 (en) 1993-06-10 1999-07-01 Konan Common Co Method and device for obtaining images of the cornea
AU716040B2 (en) 1993-06-24 2000-02-17 Bausch & Lomb Incorporated Ophthalmic pachymeter and method of making ophthalmic determinations
US5474548A (en) 1993-07-14 1995-12-12 Knopp; Carl F. Method of establishing a unique machine independent reference frame for the eye
US5502518A (en) 1993-09-09 1996-03-26 Scient Optics Inc Asymmetric aspheric contact lens
DE69402281T2 (en) 1993-09-17 1997-09-04 Essilor Int Method and device for the absolute measurement of the geometric or optical structure of an optical component
US5475452A (en) 1994-02-24 1995-12-12 Keravision, Inc. Device and method for mapping objects
US5980513A (en) 1994-04-25 1999-11-09 Autonomous Technologies Corp. Laser beam delivery and eye tracking system
US5442412A (en) 1994-04-25 1995-08-15 Autonomous Technologies Corp. Patient responsive eye fixation target method and system
US5861955A (en) 1994-04-25 1999-01-19 Medjet Inc. Topographical cornea mapping for corneal vision correction
US5632742A (en) 1994-04-25 1997-05-27 Autonomous Technologies Corp. Eye movement sensing method and system
US5493391A (en) 1994-07-11 1996-02-20 Sandia Corporation One dimensional wavefront distortion sensor comprising a lens array system
US5491524A (en) 1994-10-05 1996-02-13 Carl Zeiss, Inc. Optical coherence tomography corneal mapping apparatus
JPH10507953A (en) 1994-10-28 1998-08-04 アイシーズ、テクノロジーズ、インコーポレーテッド Corneal analyzer for compound camera
US5528791A (en) 1995-06-23 1996-06-25 New Knight Inc. Wringer floor mop with pivoting head
US5629765A (en) 1995-12-15 1997-05-13 Adaptive Optics Associates, Inc. Wavefront measuring system with integral geometric reference (IGR)
US5784146A (en) 1995-12-28 1998-07-21 Nidek Co., Ltd Ophthalmic measurement apparatus
US5822035A (en) 1996-08-30 1998-10-13 Heidelberg Engineering Optische Messysteme Gmbh Ellipsometer
US5735283A (en) 1996-10-09 1998-04-07 Snook; Richard Kieth Surgical keratometer system for measuring surface topography of a cornea during surgery
US5777719A (en) * 1996-12-23 1998-07-07 University Of Rochester Method and apparatus for improving vision and the resolution of retinal images
US5740803A (en) 1997-03-07 1998-04-21 Autonomous Technologies Corporation Locating the center of the entrance pupil of an eye after pupil dilation
EP1032809B1 (en) * 1997-11-21 2007-01-10 Alcon Inc. Objective measurement and correction of optical systems using wavefront analysis
US6007204A (en) * 1998-06-03 1999-12-28 Welch Allyn, Inc. Compact ocular measuring system
UA46833C2 (en) * 1998-10-07 2002-06-17 Інститут Біомедичної Техніки Та Технологій Академії Технологічних Наук України EYE ABERATIONAL REFRACTION MEASURING INSTRUMENT
FR2788597B1 (en) * 1999-01-15 2001-02-23 Imagine Optic Sarl HIGH DYNAMIC WAVEFRONT ANALYSIS METHOD AND DEVICE
US6086204A (en) * 1999-09-20 2000-07-11 Magnante; Peter C. Methods and devices to design and fabricate surfaces on contact lenses and on corneal tissue that correct the eye's optical aberrations
DE10014334C2 (en) * 2000-03-24 2002-03-21 Zeiss Carl Device and method for spatially resolved refractive power determination

Also Published As

Publication number Publication date
JP2003511182A (en) 2003-03-25
US6199986B1 (en) 2001-03-13
WO2001028409A1 (en) 2001-04-26
US6299311B1 (en) 2001-10-09
WO2001028409B1 (en) 2001-10-18
BR0015231A (en) 2002-07-23
AU1339101A (en) 2001-04-30
JP4660047B2 (en) 2011-03-30
EP1221889A1 (en) 2002-07-17
WO2001028409A9 (en) 2002-05-10

Similar Documents

Publication Publication Date Title
CA2388719A1 (en) Rapid, automatic measurement of the eye&#39;s wave aberraion
US6827444B2 (en) Rapid, automatic measurement of the eye&#39;s wave aberration
CA2275762C (en) Apparatus for improving vision and resolution of retinal images
AU781722B2 (en) Spatial filter for enhancing Hartmann-Shack images and associated methods
AU2007226795B2 (en) Spatial frequency wavefront sensor system and method
US9931033B2 (en) System and method for controlling a fundus imaging apparatus
KR101453327B1 (en) Fundus imaging method, fundus imaging apparatus, and storage medium
US20160198952A1 (en) Imaging Device
JP2023009530A (en) Image processing method, image processing device, and program
US8714741B2 (en) Method for selecting images
US6648473B2 (en) High-resolution retina imaging and eye aberration diagnostics using stochastic parallel perturbation gradient descent optimization adaptive optics
US20220261074A1 (en) Methods and systems for forming images of eye features using a non-imaging, scanning-mems-based eye-tracking system
CN107730545B (en) Optical imaging method and system for dynamically eliminating ghost image
US7226166B2 (en) Optimizing the properties of electromagnetic energy in a medium using stochastic parallel perturbation gradient descent optimization adaptive optics
US20040165147A1 (en) Determining iris biometric and spatial orientation of an iris in accordance with same
JP2017189617A (en) Imaging apparatus control method, computer-readable medium, and controller for controlling imaging apparatus
JP2013215340A (en) Imaging method and imaging apparatus
JP7182855B2 (en) Method for controlling optical imaging device, storage medium storing same, controller, and optical imaging device
JP6116227B2 (en) Aberration measuring apparatus and method

Legal Events

Date Code Title Description
EEER Examination request
FZDE Dead