WO2003034167A2 - Method to optimize test data - Google Patents

Method to optimize test data Download PDF

Info

Publication number
WO2003034167A2
WO2003034167A2 PCT/US2002/033084 US0233084W WO03034167A2 WO 2003034167 A2 WO2003034167 A2 WO 2003034167A2 US 0233084 W US0233084 W US 0233084W WO 03034167 A2 WO03034167 A2 WO 03034167A2
Authority
WO
WIPO (PCT)
Prior art keywords
data
maneuver
test
flight
file
Prior art date
Application number
PCT/US2002/033084
Other languages
French (fr)
Other versions
WO2003034167A3 (en
Inventor
Robert W. Vanderwiel
Original Assignee
Lockheed Martin Corporation
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Lockheed Martin Corporation filed Critical Lockheed Martin Corporation
Priority to EP02776224A priority Critical patent/EP1436679B1/en
Priority to DE60208415T priority patent/DE60208415T2/en
Publication of WO2003034167A2 publication Critical patent/WO2003034167A2/en
Publication of WO2003034167A3 publication Critical patent/WO2003034167A3/en

Links

Classifications

    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05BCONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
    • G05B21/00Systems involving sampling of the variable controlled
    • G05B21/02Systems involving sampling of the variable controlled electric
    • BPERFORMING OPERATIONS; TRANSPORTING
    • B64AIRCRAFT; AVIATION; COSMONAUTICS
    • B64FGROUND OR AIRCRAFT-CARRIER-DECK INSTALLATIONS SPECIALLY ADAPTED FOR USE IN CONNECTION WITH AIRCRAFT; DESIGNING, MANUFACTURING, ASSEMBLING, CLEANING, MAINTAINING OR REPAIRING AIRCRAFT, NOT OTHERWISE PROVIDED FOR; HANDLING, TRANSPORTING, TESTING OR INSPECTING AIRCRAFT COMPONENTS, NOT OTHERWISE PROVIDED FOR
    • B64F5/00Designing, manufacturing, assembling, cleaning, maintaining or repairing aircraft, not otherwise provided for; Handling, transporting, testing or inspecting aircraft components, not otherwise provided for
    • B64F5/60Testing or inspecting aircraft components or systems
    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05BCONTROL OR REGULATING SYSTEMS IN GENERAL; FUNCTIONAL ELEMENTS OF SUCH SYSTEMS; MONITORING OR TESTING ARRANGEMENTS FOR SUCH SYSTEMS OR ELEMENTS
    • G05B23/00Testing or monitoring of control systems or parts thereof
    • G05B23/02Electric testing or monitoring
    • G05B23/0205Electric testing or monitoring by means of a monitoring system capable of detecting and responding to faults
    • G05B23/0218Electric testing or monitoring by means of a monitoring system capable of detecting and responding to faults characterised by the fault detection method dealing with either existing or incipient faults
    • G05B23/0256Electric testing or monitoring by means of a monitoring system capable of detecting and responding to faults characterised by the fault detection method dealing with either existing or incipient faults injecting test signals and analyzing monitored process response, e.g. injecting the test signal while interrupting the normal operation of the monitored system; superimposing the test signal onto a control signal during normal operation of the monitored system

Definitions

  • the present invention relates generally to the field of system tests and more specifically to determining whether a testing maneuver involving multi-state variables produces new data. Description of the Related Art
  • Aircraft flight tests are expensive and time-consuming aspects of certifying aircraft for use. Typically, a number of flight tests will be conducted for an aircraft, wherein each flight test will consist of a number of prearranged maneuvers. During the flight test, instruments will record various data relating to the aircraft during the tests.
  • a method for optimizing data gathering comprising the steps of designing a first system test comprising two or more maneuvers, compiling one or more data variables for each of the two or more maneuvers from the first system test, pooling each data variable from each maneuvers into a first set of data, removing a first maneuver's data variables from the pooled first set of data, comparing the first maneuver's data variables against the first set of data with the first maneuver's data variables removed, and determining whether to perform the first maneuver based on whether the maneuver provided new data.
  • a method for optimizing data gathering comprising the steps of performing a first system test, compiling a first set of test data from the first system test, designing a second system test, compiling a second set of test data from the second system test, comparing the first data set to the second data set, and determining whether to perform the second test based on the comparing step.
  • a system for optimizing data gathering comprising of a first set of data gathered from a first system test, a second set of data from a second system test, and an algorithm for comparing the first set of data to the second set of data, wherein the algorithm provides an optimization result.
  • An apparatus for optimizing data gathering comprising of a computer, a first data set from a first system test, a second set of data from a second system test and an algorithm running on the computer, wherein the algorithm compares the first set of data to the second set of data and provides a result.
  • Figure 1C shows a pattern recognition comparison between the training file vs. the test file;
  • Figure 2A shows a non-normal distribution of training file data arranged in three blocks;
  • Figure 2B shows a test file of random data covering the range 0 ⁇ X ⁇ 10 and 0 ⁇ Y ⁇ 1;
  • Figure 2C shows the present invention highlighting those test file data points outside of the training file;
  • Figure 2D shows the present invention's electronic screen display of the test file output which provided the data necessary to create Figure 2C;
  • Figure 3A shows a tri-variate training file versus a test file results, the example using one pair of variables with DM at 35;
  • Figure 3B shows the data from Figure 3 A with the DM at 15
  • Figure 3C shows the data from Figure 3A with DM at 7
  • Figure 4 shows the present invention's ability to detect new information in a test file when using aircraft data distributions
  • Figure 5 shows the interactions between one of two variable combinations
  • Figure 6 shows the interactions between the two variable combinations as in Figure 5 with a different flight used as the test file
  • Figures 7A shows an example of one pair of variables, vertical load and lateral acceleration
  • Figure 7B shows an example of one pair of variables, lateral acceleration and roll acceleration
  • Figure 7C shows an example of one pair of variables, lateral acceleration and angle of attack
  • Figure 8 illustrates a data subset analysis figure depicting a comparison of one flight vs. a training file of the previous flights taken
  • Figure 9 displays in chart format the amount of useful information in each flight test when compared to a training file composed of all previous flights taken
  • Figure 10 illustrates the percent of useful information in each flight test (y-axis) when compared to a training file of all previous flights taken (x-axis);
  • Figure 11 illustrates the close relationship between the amount of new information in independent (y-axis) when compared to percent of useful information in independent variables (x-axis);
  • Figure 12A illustrates maneuver candidates for elimination;
  • Figure 12B illustrates different maneuver candidates for elimination
  • Figure 12C illustrates different maneuver candidates for elimination
  • Figure 13 shows a typical combination of variables, which shows that all of the data in the test file is indeed included in the training file;
  • Figure 14 illustrates at least one useful data point using the statistical parameters Flight A Run B maneuver when compared to the training file database using the same combination of variables;
  • Figure 15 illustrates an example of data gathered from three hypothetical flights
  • Figure 16 illustrates the percentage of accumulated maneuvers selected as candidates for elimination (y-axis) as flight test is in progress (x-axis);
  • Figure 17A illustrates a flight test with all data (vertical load vs. lateral acceleration);
  • Figure 17B illustrates a flight test after 41.1% data reduction (vertical load vs. lateral acceleration);
  • Figure 17C illustrates a flight test with all data (vertical load vs. angle of attack);
  • Figure 17D illustrates a flight test after 41.1% data reduction (vertical load vs. angle of attack);
  • Figure 17E illustrates a flight test with all data (lateral acceleration vs. roll acceleration);
  • Figure 17F illustrates a flight test after 41.1% data reduction (lateral acceleration vs. roll acceleration);
  • Figure 17G illustrates a flight test with all data (lateral acceleration vs. angle of attack);
  • Figure 17H illustrates a flight test after 41.1% data reduction (lateral acceleration vs. angle of attack)
  • Figure 18 illustrates in flow chart form, the methodology of the present invention
  • FIG 19 illustrates step 5 of Figure 18 in greater detail
  • Figure 20 illustrates step 7 of Figure 18 in greater detail
  • FIG. 21 illustrates step 9 of Figure 18 in greater detail.
  • the present invention comprises of an advanced pattern recognition algorithm coupled with fault detection model based on a set of statistical hypothesis tests.
  • One such advanced pattern recognition algorithm developed by Argonne National Laboratory and titled the Multivariate State Estimation Technique (MSET) has demonstrated the capability of identifying the amount of useful information in a test file composed of multivariate data when compared to the information contained in a training file data base.
  • MSET Multivariate State Estimation Technique
  • the present invention using algorithms such as MSET (other suitable algorithms can be used) is capable of analyzing situations where thousands of different variables are utilized. These attributes provide the present invention with the capability of enhancing any data gathering process including flight test.
  • the MSET algorithm was created under contract with the United States Government.
  • Utilization of the present invention to determine the amount of useful information in a test file when compared to the information in a training file involves a training phase, a system phase and a testing phase.
  • the training phase utilizes the training file to establish normal data boundaries and "trains itself to recognize normal process behavior and correlations between variables or sensors.
  • an estimate of each data point in the test file is calculated by comparing the test file data point values with comparable data points in the training file.
  • the difference between the test file data point value and the estimate (the estimation error) is then analyzed in the testing phase by a statistically based hypothesis test to determine which test file data points contain useful information values (i.e., which test data points are significantly and meaningfully different from the data points already accumulated).
  • the major criteria for successful application of the present invention involves a small degree of mutual correlation among the sensor readings that are gathered.
  • Major decisions encountered when applying the present invention involve determining the parameter settings (the mean test disturbance magnitude, the variance test disturbance magnitude, the false alarm probability, or the missed alarm probability, for example). Different parameter settings increase or decrease the sensitivity of the algorithm.
  • a pattern recognition comparison between the training file vs. the test file (shown in Figure 1C) highlights the data 110 found in the test file not included in the training file.
  • Test file data 110 is "useful" data since it fills in the gaps of the training file.
  • Test file data 115 may be considered redundant since it is already included in the training file.
  • the amount of "useful" data contained in this test file when compared to the training file is 2/5 or 40%.
  • Figure 2A shows a non-normal distribution of Training File data arranged in three blocks 205.
  • a Test File of random data 210 covers the range 0 ⁇ X ⁇ 10 and 0 ⁇ Y ⁇ 1.
  • the present invention may not have been expected to perform well under these circumstances, because the correlation between the two variables in the training file is almost non-existent.
  • Figure 2C illustrates that the present invention is able to highlight those test file data points 220 outside of the training file.
  • Figure 2D shows the present invention's electronic screen display of the test file output which provided the data necessary to create Figure 2C.
  • DM disturbance magnitude
  • Alpha probability of wrongful rejection
  • beta probability of wrongful acceptance
  • the disturbance magnitude for the mean test specifies the number of standard deviations that the mean of the alternative hypothesis distribution is shifted relative to the null hypothesis distribution.
  • Figure 4 illustrates the present invention's ability to detect new information in a test file when using aircraft data distributions.
  • Figure 4 shows a comparison of a theoretical distribution vs. a distribution of aircraft data.
  • a training file composed of data variables (engine power Level Ev. fan rotor speed) taken from an aircraft flight is compared to a test file composed of the same two variables arranged in a diagonal line (moving randomly off center).
  • the portion of the data 410 line formed from dashes
  • test file 420 solid dots
  • Comparisons between distributions of aircraft data demonstrate the invention's capability of determining the amount of useful in formation obtained in one distribution when compared to another. Each data point may have not two but hundreds of variables associated with it.
  • One of the advantages in using a pattern recognition algorithm is the capability of analyzing data points associated with large numbers of variables in multi-dimensional space. For ease of illustration, comparison examples used herein are limited to graphs of two significant variables.
  • Figure 5 shows comparison made between two aircraft flights in which data pertaining to the aircraft engine was accumulated. Data distributions of three variables (engine power level angle, fan rotor speed Nl and high pressure rotor speed N2) were utilized to illustrate the amount and location of useful data gathered in Flight XXA (test file) when compared to the engine data gathered in Flight XXB (training file).
  • Figure 5 represents the interactions between one of the two variable combinations.
  • the amount of new or useful information 530 (dashes) in the test file 520 (hollow dots) and 530 when compared to the training file 510 (solid dots) was estimated
  • Flight XXA had been flown first instead of Flight XXB and the flight engineers were interested in determining the amount of new or useful information gathered in Flight XXB (pertaining to the same three variables)
  • Flight XXA data would be placed in the training file and Flight XXB data would be placed in the test file.
  • Figure 6 represents the interactions between the same two variable combinations when Flight XXB was used as the test file.
  • the amount of new or useful information 630 (dashes) in the test file Flight XXB when compared to the training file Flight XXA (both solid and hollow dots) was estimated to be approximately 16%.
  • FIG. 8 illustrates a data- subset analysis figure depicting a comparison of one flight vs. a training file of the previous flights taken.
  • Subsets 810, 815 and 820 represent the training file data, wherein Subset 830 represents the percent of information from the test file that is outside the training file.
  • Figure 9 provides another illustration of the amount of useful information from a previously recorded flight test, in each flight test when compared to a training file composed of all previous flights taken As a test program progresses and as more data is added to the training file, the amount of new data in each flight test gradually decreases.
  • Figure 10 illustrates the percent of useful information in each flight test (Y axis) when compared to a training file of all previous flights taken (X axis). As the entire envelope of performance is mapped into the training file, it becomes much more difficult for subsequent test flights to add any new information. When a flight test program reaches the point at which subsequent test flights are not able to add a significant amount of new information, consideration should be given to end the flight test program.
  • the present invention can be used with a flight simulation to reduce flight test costs.
  • the strategy for reducing flight test cost involves reducing the number of specific maneuvers, which contribute no new or useful information. If each maneuver is compared to a training file of all the other remaining test maneuvers combined, the resulting estimate of new information in each maneuver provides an indication of the individual maneuver's data gathering cost effectiveness. Maneuvers which display little or no significant amount of useful information or contribution to the program are eliminated from the training file before the next comparison is made until all non-contributing maneuvers are eliminated.
  • Store configuration QA data consisted of 6 of 16 flight test program flights, 65 of 163 maneuvers, and 11,968 of 47,711 observation points.
  • Observational data consisted of 28 independent variables (airspeed, aircraft gross weight, angle of attack, etc.) and 33 dependent variables (vertical tail shear, center fuselage end moment, etc.). Since only independent variable data is gathered in flight simulation, only independent variable data was used in the analysis. If simulation is used to gather data, the amount of data pertaining to dependent variables can be estimated due to the correlation between dependent and independent variables.
  • Figure 11 depicts a graph illustrating the close relationship between the amount of new information in independent (y-axis) when compared to percent of useful information in independent variables (x-axis).
  • New independent variable information gathered in simulation means that new dependent variable information would be gathered in flight test.
  • Figure 12A illustrates maneuver candidates for elimination.
  • Figure 12B illustrates different maneuver candidates 1205 for elimination. Data sets were considered to provide an insignificant amount of new data if the amount of new data in the data set was less than or equal to 0.5%. The percentage of 0.5% was selected since 0.5% represented less than one data point.
  • Flight A Run 12 was one of the maneuvers identified as containing no useful data (i.e., zero new observations) when compared to a database of the other maneuvers.
  • Figure 13 depicts a typical combination of variables which shows that all of the data in the test file 1320 is indeed included in the training file 1310. Other two variable combinations provide similar results.
  • Figure 15 provides a simple example of data gathered (in alphabetical sequence) from three hypothetical flights. If we were to apply our pattern recognition data gathering optimization analysis to this situation it would most likely find that flights A and B contain significant amounts of redundant information and are candidates for elimination. However, flights A and B were most likely needed to accomplish flight C so no reduction in flight test program cost's could have been obtained by using this method and only performing flight C.
  • the present invention's pattern recognition flight test program cost reduction techniques would probably not provide much useful information if flight test programs could be planned using a small, manageable number of variables or sensor inputs.
  • the number of independent and dependent variables being recorded in flight-test sensors and flight recorders grows larger and larger with subsequent flight test programs. Determining if an aircraft is operating in the
  • Figure 15 A, B, or C range becomes a difficult task when 100 variables resulting in 4950 combinations of two variables are interacting.
  • the need for using pattern recognition data gathering optimization analysis to assist the flight test planner in gathering flight test data in the most cost effective manner increases.
  • 41.5% of the maneuvers showed no useful information.
  • Most of these maneuvers were located at the beginning and at the end of the flight test program.
  • Figure 16 illustrates the percentage of accumulated maneuvers selected as candidates for elimination (Y axis) as the QA store configuration flight test program progressed (X axis).
  • the accumulated percentage of maneuvers possessing redundant information did not increase as the program progressed but fluctuated around 40 percent throughout the test program.
  • 27 out of 65 or 41.54% of the maneuvers were deemed to provide an insignificant amount of new data and considered strong candidates for elimination providing a safety risk was not incurred (see Table 12A and B).
  • the present invention allows this information to be gathered using flight simulation before the flight test program begins and if all of the 27 non-contributing maneuvers could be removed (i.e., never flown), the amount of data in the flight test program would be reduced by 41.1% and the number of maneuvers by 41.54%.
  • each removed data point has 28 independent variables associated with it.
  • Figures 17A through 17H represent several store configuration 29 A flight test two variable before/after comparisons when all of the data was included and when the data from the 27 non-contributing maneuvers was removed. Significant points (points on the edge perimeters) remain even though 41.1 % of the data points have been removed.
  • step 1 a request is submitted for a system's test, in this case, a flight test for a specific aircraft. Once submitted, the flight test data gathering goals and requirements are defined in step 2.
  • the flight test which is composed of numerous flights and maneuvers designed to meet the data gather goals and requirements, is proposed to the flight test engineering, analysis and design group in step 3. Using their experience, the flight test engineers will design the flight test.
  • step 4 the proposed flight, as designed by the flight test engineering group is executed, with the various flights and maneuvers, using a simulator and the flight data is gather for analysis.
  • step 5 the flight simulation data is evaluated using a data gathering optimization statistical algorithm to determine which maneuvers provide little information and should be considered as candidates for removal from the flight test. Additionally, the flights and maneuvers will be prioritized in relationship to the data gather effectiveness.
  • step 6 non-data gathering maneuvers from the flight test are identified as possible candidates for removal from the flight test. These are presented to the flight test engineering, analysis and design group for evaluation.
  • step 7 the flight test engineering group determines which maneuver candidates may be eliminated and re-design the flight test using the prioritized data gathering information determined in step 5, so that the flight test is conducted in the most efficient data gathering fashion.
  • step 8 the actual flight test is conducted with the reduced number of flights and maneuvers to gather the actual flight test data.
  • step 9 the actual flight test data gathered in step 8 is analyzed while the test is in progress to determine the actual amount of significant data being gathered, using the data gathering optimization techniques in step 5. After the completion of the flight test, the completed flight test data is also analyzed to determine the accuracy of earlier estimations based on the flight simulation data.
  • FIG 19 illustrates step 5 of Figure 18 in greater detail.
  • each flight simulation data point which comprises of a vector composed of numerous independent variables, such as velocity, altitude, and a flight time marker indicating when the data was gathered during the flight simulation's various flights and maneuvers is collected during the simulation of the flight.
  • the flight simulation data for each flight test maneuver is converted into a database spreadsheet file with the flight test flight number, maneuver number, and time marker placed in separate columns for identification purposes.
  • a separate row in the spreadsheet can be assigned for each data vector gathered and a separate column in the spreadsheet can be assigned for each independent variable.
  • step 5.3 the spreadsheet information for each maneuver is combined into one large database file.
  • a typical large database spreadsheet file might contain over 50,000 rows and 50 columns.
  • step 5.4 the importance of each independent variable associated with each vector is determined.
  • a determination is also made for the number of independent variable alarms criteria for identifying when a vector associated with a test file is considered to be outside the data region associated with a training file.
  • step 5.5 a determination is made for the amount of significant data criteria used in identifying those maneuvers which are non-contributing data gatherers when compared to the other maneuvers of the test flight.
  • the significant data criteria will comprise the percent of data in a test file, which is considered to be outside the data region associated with a training file.
  • step 5.6 one maneuver's worth of data information is removed from the main database file.
  • This set of data is the test file.
  • the remaining portion of the database information file is the training file.
  • the test and training file is converted into text files for analysis and evaluation by a subsequent advanced pattern recognition algorithm such as the Multivariate State Estimation Technique (MSET) developed by Argonne National Laboratory.
  • MSET Multivariate State Estimation Technique
  • step 5.7 the basic description of determining which vectors within the test file are in variance with normal or outside the region of vectors and thus represent new data contained in the training file is determined. This is accomplished by setting up a system model that provides analytically derived values of all monitored sensors is integrated with a statistically-based hypothesis test that compares the analytically estimated signal value with the test file measured values to detect the development or presence of new data.
  • Three essential models of the flight for the algorithm include a training algorithm for the collection of representative data from sensors during normal system operations, an empirically based model for system state estimation, and a statistically based fault detection system.
  • the training model is used to produce a training set or set points whose data encompasses all expected normal operating states of the system, ideally.
  • the system-modeling module (the test file) is used to estimate the values of all signals that are present in the process that is being monitored.
  • the fault detection model is used to detect disturbances through an examination of the difference between the estimated and measured signal values.
  • An algorithm (MSET for example) is used to detect incipient faults between current and normal behavior. However, instead of detecting incipient faults, the algorithm will be estimating which vectors in the test file exhibit abnormal behavior or could be considered new or useful data points when compared with the vector data stored in the training file.
  • a test file is created. This is accomplished by using the data associated with one maneuver from the training file database.
  • the data associated with the test file is removed in step 5.8.1 from the training file for the purpose of analyzing the maneuver.
  • step 5.8.2 the test file is deliberately seeded with certain data known to be outside the training file.
  • step 5.8.3 a comparison of the test file to the training file is made using the advanced pattern recognition algorithm.
  • the pattern recognition algorithm parameters such as wrongful rejection probability, wrongful acceptance probability, sequential probability ratio test sensitivity, is adjusted to ensure that the algorithm is operating at optimal performance when detecting the data known to be outside the training file.
  • step 5.9 once the algorithm parameters, the number of independent variable alarm criteria, and the non-data gathering criteria for maneuvers have been established, the test file, seeded with false data for determining optimal the algorithm parameter values, is returned back into the training master database in its original state, i.e., with the seeded data removed. At this point the algorithm is ready to begin the complete analysis of each maneuver of the simulated flight data or system test.
  • Step 5.10 runs the analysis of the flight tests maneuvers.
  • step 5.10.1 the data associated with a specific maneuver is removed from the training file and placed in a test file.
  • the training file is normalized or fit to a normal distribution. If the maneuver's test file is found to be a non-contributor of data, that maneuvers data is removed from the training file and the process of ranking the maneuvers in terms of data gathering capabilities starts all over again.
  • step 5.10.2 the test file is compared using the algorithm to the training file associated with the data of all remaining maneuvers combined to determine the variance of the maneuver from normal.
  • step 5.10.3 the data associated with the maneuver is returned to the training file and the next maneuver is removed and placed into the test file. This process is repeated until all maneuvers have had their data placed in the test file and compared to the training file comprising of data from all the other non-test file maneuvers.
  • the resulting estimate of new information from each maneuver provides an indication of the individual maneuver's data gathering cost effectiveness, A non-contributing data-gathering maneuver contains predominantly redundant information when compared to the data from the rest of the maneuvers in the training file.
  • a table is prepared comprising of the flight test maneuvers and the percent of useful information contained within each maneuver.
  • the percent of useful information is calculated by determining the number of vectors, which meet the number of independent variable alarm criteria, divided by the total number of vectors in the maneuver.
  • the flight test engineers will then select the maneuver with the smallest percent contribution of useful information. If this value is less than the previously agreed upon value for the criteria for identifying a non-contributing data gathering maneuver, then the flight test engineer may remove the maneuver's data from the master training file.
  • step 5.12 a record is maintained of those maneuver vector values associated with the variable alarms for the non-contributing data gathering maneuver's data, which has been removed from the training file for future analysis.
  • the vector value may provide c ⁇ tical information and justify maintaining the maneuver in the flight test even though only one significant vector value was gathered.
  • step 5 13 the process of analyzing the simulated data should be repeated (from step 5.9 through 5 11). Subsequently, each time a non-contributing data-gathering maneuver is removed, the analysis of the training file against a test file for each maneuver as described above, should be repeated to ensure that all non-cont ⁇ butmg data is accurately analyzed and identified. Finally, in step 5 14, a listing of all non-contributing data gathering maneuvers and maneuver vector values associated with variable alarms are prepared for presentation to the flight test engineers This list will show the maneuver candidates for flight test elimination.
  • Figure 20 illustrates step 7 of Figure 18 m greater detail.
  • the flight test engineering, analysis and design groups receive a listing of non-data gathe ⁇ ng maneuvers identified as possible candidates for removal from the flight test.
  • the flight test engineers determine which non-data gathering maneuvers may be eliminated from the flight test m step 7 2. Criteria for elimination may include the lack of data gathered, the importance of the minimal data gathered, or flight test safety aspects if the maneuver is eliminated, as well as others.
  • step 7 a final framing file database composed of only those maneuvers selected for the final test flight is created in step 7.3.
  • a table listing the percent of useful information contained within each final flight test maneuver using steps 5.9 through 5.10 is prepared in step 7.4.
  • the list of maneuvers is prioritized showing which maneuvers gather the most useful data and the data gathe ⁇ ng cost effectiveness of each maneuver
  • the flight test engineers redesign the reduced size flight test to accommodate the elimination of the non-data gathering maneuvers chosen for removal.
  • the prioritized list of maneuvers indicating each maneuver's data gathe ⁇ ng effectiveness can also be utilized in the flight test redesign to ensure that the flight test redesign is conducted in the most efficient data gathering fashion.
  • FIG 21 illustrates step 9 of Figure 18 m greater detail.
  • step 9.1 actual flight test data with maneuver markers and time markers is translated into various database files and test files for subsequent processing.
  • Each subsequent maneuver with a framing file composed of the previous maneuvers is compared to obtain an estimate of how much new and useful data is being gathered du ⁇ ng the flight test in step 9.2.
  • As the entire envelop of performance is mapped into the framing file it becomes difficult for subsequent maneuvers to add any new information and the amount of new information being added should decrease.
  • the percentage of useful dependent and independent information in each maneuver is determined, when compared to a training file composed of data from all flight test maneuvers in step 9.3.
  • the independent variable useful information values obtained for each maneuver in actual flight test to those obtained in flight simulation to determine the accuracy of the flight.
  • step 9.4 graphs of various variable combinations depicting the training file data and test file data are prepared to confirm that the test file data identified as useful is visually outside the training file region.
  • the correlation coefficient between independent variable new information obtained versus dependent variable new information is determined to obtain enhanced future estimates.
  • step 9.6 the test files of flight simulation data is compared to the training file of actual flight test data using the parameters set in step 5. Those regions of the simulation that need adjustment to better represent actual aircraft performance is highlighted, and the flight simulation is enhanced by adjusting the flight simulation parameters.
  • Figure 4 shows a comparison of a theoretical distribution vs. a distribution of aircraft data.
  • a fraining file composed of data variables (fan rotor speed vs. EPLA engine power level angle) taken from an aircraft flight is compared to a test file composed of the same two variables arranged in a diagonal line (moving randomly off center). As the test file line intersects the training file data, the portion of the data from the diagonal line test file (highlighted in pink) outside of the training file becomes apparent.
  • Comparisons between distributions of aircraft data demonstrate the invention's capability of determining the amount of useful in formation obtained in one distribution when compared to another.
  • Each data point may have not two but hundreds of variables associated with it.
  • One of the advantages in using a pattern recognition algorithm is the capability of analyzing data points associated with large numbers of variables in multi-dimensional space. For ease of illustration, comparison examples will be limited to graphs of two significant variables.

Abstract

A method for optimizing data gathering, is provided wherein the method comprising the steps of designing a first system test comprising two or more maneuvers, compiling a one or more data variable for each of the two or more maneuvers from the first system test, pooling each data variable from each maneuvers into a first set of data, removing a first maneuver's data variables from the pooled first set of data, comparing the first maneuver's data variables against the first set of data with the first maneuver's data variables removed, and determining whether to perform the first maneuver based on whether the maneuver provided new data.

Description

METHOD TO OPTIMIZE TEST DATA
BACKGROUND OF THE INVENTION Technical Field The present invention relates generally to the field of system tests and more specifically to determining whether a testing maneuver involving multi-state variables produces new data. Description of the Related Art
Aircraft flight tests are expensive and time-consuming aspects of certifying aircraft for use. Typically, a number of flight tests will be conducted for an aircraft, wherein each flight test will consist of a number of prearranged maneuvers. During the flight test, instruments will record various data relating to the aircraft during the tests.
While the flight tests do not repeat maneuvers from previous flight tests or maneuvers, often some maneuvers will not add additional information regarding the aircraft from the data, which has been previously recorded during other maneuvers. When no new data is added by a maneuver, from a pure new data acquisition model, maneuvers could be candidates for modification or removal from the flight test in order to maintain the efficiency, both in terms of time and costs of the flight-test.
BRIEF SUMMARY OF THE INVENTION
A method for optimizing data gathering, is provided wherein the method comprising the steps of designing a first system test comprising two or more maneuvers, compiling one or more data variables for each of the two or more maneuvers from the first system test, pooling each data variable from each maneuvers into a first set of data, removing a first maneuver's data variables from the pooled first set of data, comparing the first maneuver's data variables against the first set of data with the first maneuver's data variables removed, and determining whether to perform the first maneuver based on whether the maneuver provided new data.
A method for optimizing data gathering is provided wherein the method comprising the steps of performing a first system test, compiling a first set of test data from the first system test, designing a second system test, compiling a second set of test data from the second system test, comparing the first data set to the second data set, and determining whether to perform the second test based on the comparing step.
A system for optimizing data gathering comprising of a first set of data gathered from a first system test, a second set of data from a second system test, and an algorithm for comparing the first set of data to the second set of data, wherein the algorithm provides an optimization result.
An apparatus for optimizing data gathering comprising of a computer, a first data set from a first system test, a second set of data from a second system test and an algorithm running on the computer, wherein the algorithm compares the first set of data to the second set of data and provides a result. The foregoing has outlined some of the more pertinent objects and features of the present invention. These objects should be construed to be merely illustrative of some of the more prominent features and applications of the invention. Many other beneficial results can be attained by applying the disclosed invention in a different manner or modifying the invention as will be described. Accordingly, other objects and a fuller understanding of the invention may be had by referring to the following Detailed Description of the Preferred Embodiment.
BRIEF DESCRIPTION OF THE DRAWINGS
For a more complete understanding of the present invention and the advantages thereof, reference should be made to the following Detailed Description taken in connection with the accompanying drawings in which:
Figure 1 A shows a training file of the equation X=Y between 0<X<1,2<X<3, and 4<X<5; Figure IB shows the same equation X=Y, range 0<X<5, but without the training file data gaps; Figure 1C shows a pattern recognition comparison between the training file vs. the test file; Figure 2A shows a non-normal distribution of training file data arranged in three blocks;
Figure 2B shows a test file of random data covering the range 0<X<10 and 0<Y<1; Figure 2C shows the present invention highlighting those test file data points outside of the training file;
Figure 2D shows the present invention's electronic screen display of the test file output which provided the data necessary to create Figure 2C;
Figure 3A shows a tri-variate training file versus a test file results, the example using one pair of variables with DM at 35;
Figure 3B shows the data from Figure 3 A with the DM at 15; Figure 3C shows the data from Figure 3A with DM at 7; Figure 4 shows the present invention's ability to detect new information in a test file when using aircraft data distributions;
Figure 5 shows the interactions between one of two variable combinations; Figure 6 shows the interactions between the two variable combinations as in Figure 5 with a different flight used as the test file; Figures 7A shows an example of one pair of variables, vertical load and lateral acceleration;
Figure 7B shows an example of one pair of variables, lateral acceleration and roll acceleration; Figure 7C shows an example of one pair of variables, lateral acceleration and angle of attack; Figure 8 illustrates a data subset analysis figure depicting a comparison of one flight vs. a training file of the previous flights taken; Figure 9 displays in chart format the amount of useful information in each flight test when compared to a training file composed of all previous flights taken; Figure 10 illustrates the percent of useful information in each flight test (y-axis) when compared to a training file of all previous flights taken (x-axis);
Figure 11 illustrates the close relationship between the amount of new information in independent (y-axis) when compared to percent of useful information in independent variables (x-axis); Figure 12A illustrates maneuver candidates for elimination;
Figure 12B illustrates different maneuver candidates for elimination;
Figure 12C illustrates different maneuver candidates for elimination
Figure 13 shows a typical combination of variables, which shows that all of the data in the test file is indeed included in the training file; Figure 14 illustrates at least one useful data point using the statistical parameters Flight A Run B maneuver when compared to the training file database using the same combination of variables;
Figure 15 illustrates an example of data gathered from three hypothetical flights;
Figure 16 illustrates the percentage of accumulated maneuvers selected as candidates for elimination (y-axis) as flight test is in progress (x-axis); Figure 17A illustrates a flight test with all data (vertical load vs. lateral acceleration);
Figure 17B illustrates a flight test after 41.1% data reduction (vertical load vs. lateral acceleration);
Figure 17C illustrates a flight test with all data (vertical load vs. angle of attack);
Figure 17D illustrates a flight test after 41.1% data reduction (vertical load vs. angle of attack); Figure 17E illustrates a flight test with all data (lateral acceleration vs. roll acceleration);
Figure 17F illustrates a flight test after 41.1% data reduction (lateral acceleration vs. roll acceleration);
Figure 17G illustrates a flight test with all data (lateral acceleration vs. angle of attack);
Figure 17H illustrates a flight test after 41.1% data reduction (lateral acceleration vs. angle of attack)
Figure 18 illustrates in flow chart form, the methodology of the present invention;
Figure 19 illustrates step 5 of Figure 18 in greater detail;
Figure 20 illustrates step 7 of Figure 18 in greater detail; and
Figure 21 illustrates step 9 of Figure 18 in greater detail.
DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
The present invention comprises of an advanced pattern recognition algorithm coupled with fault detection model based on a set of statistical hypothesis tests. One such advanced pattern recognition algorithm, developed by Argonne National Laboratory and titled the Multivariate State Estimation Technique (MSET), has demonstrated the capability of identifying the amount of useful information in a test file composed of multivariate data when compared to the information contained in a training file data base. The present invention using algorithms such as MSET (other suitable algorithms can be used) is capable of analyzing situations where thousands of different variables are utilized. These attributes provide the present invention with the capability of enhancing any data gathering process including flight test. The MSET algorithm was created under contract with the United States Government.
Utilization of the present invention to determine the amount of useful information in a test file when compared to the information in a training file involves a training phase, a system phase and a testing phase. In the training phase, the present invention utilizes the training file to establish normal data boundaries and "trains itself to recognize normal process behavior and correlations between variables or sensors.
In the system phase, an estimate of each data point in the test file is calculated by comparing the test file data point values with comparable data points in the training file. The difference between the test file data point value and the estimate (the estimation error) is then analyzed in the testing phase by a statistically based hypothesis test to determine which test file data points contain useful information values (i.e., which test data points are significantly and meaningfully different from the data points already accumulated).
The major criteria for successful application of the present invention involves a small degree of mutual correlation among the sensor readings that are gathered. Major decisions encountered when applying the present invention involve determining the parameter settings (the mean test disturbance magnitude, the variance test disturbance magnitude, the false alarm probability, or the missed alarm probability, for example). Different parameter settings increase or decrease the sensitivity of the algorithm.
The following figures and descriptions illustrate the present invention's ability to detect new information in a test file when compared to existing information in a training file when using theoretical distributions.
Figure 1A shows an example of a training file of the equation X=Y between 0<X<1,2<X<3, and 4<X<5. The variables in this distribution have a high degree (correlation coefficient r=1.0) of correlation. The test file shown in Figure IB contains the same equation X=Y, range 0<X<5, but without the training file data gaps. A pattern recognition comparison between the training file vs. the test file (shown in Figure 1C) highlights the data 110 found in the test file not included in the training file. One might argue that the test file data 110 is "useful" data since it fills in the gaps of the training file. Test file data 115 may be considered redundant since it is already included in the training file. Hence the amount of "useful" data contained in this test file when compared to the training file is 2/5 or 40%. Figure 2A shows a non-normal distribution of Training File data arranged in three blocks 205.
Correlation between X and Y is almost zero (r=-.0124). A Test File of random data 210 (see Figure 2B) covers the range 0<X<10 and 0<Y<1. The present invention may not have been expected to perform well under these circumstances, because the correlation between the two variables in the training file is almost non-existent. Yet Figure 2C illustrates that the present invention is able to highlight those test file data points 220 outside of the training file. Figure 2D shows the present invention's electronic screen display of the test file output which provided the data necessary to create Figure 2C.
To illustrate the importance of the present invention is parameter called the disturbance magnitude (DM), a training file and a test file were both generated at random from the same trivariate (three variable) distribution with a correlation coefficient between XY of r=.791 and between XZ of r=- .367. Alpha (probability of wrongful rejection) and beta (probability of wrongful acceptance) remained fixed at 0.1 throughout the test. The disturbance magnitude for the mean test specifies the number of standard deviations that the mean of the alternative hypothesis distribution is shifted relative to the null hypothesis distribution.
Because both files illustrated in Figure 3A were generated from the same distribution, the present invention (with a DM=35) found no useful data in the test file as illustrated in Figure 3 A. In Figure 3 A, training file data 310 (solid dots) and test file data 320 (hollow dots) subsume all new data, • therefore no new data is provided. However, as we increase the sensitivity of the algorithm (e.g., reduce the DM down to 15 and then down to 7) the number of data points identified in the test file as being useful increases (see Figure 3B and 3C). In Figure 3B, new data 330 (dashes) are present in the graph, while in Figure 3C, more new data 330 is present. Hence, when using pattern recognition algorithms of this type, it should be noted that a delicate balance must be achieved when setting parameters to accurately identify useful information.
Figure 4 illustrates the present invention's ability to detect new information in a test file when using aircraft data distributions. Figure 4 shows a comparison of a theoretical distribution vs. a distribution of aircraft data. A training file composed of data variables (engine power Level Ev. fan rotor speed) taken from an aircraft flight is compared to a test file composed of the same two variables arranged in a diagonal line (moving randomly off center). As the test file line intersects the training file data, the portion of the data 410 (line formed from dashes) from the diagonal line test file 420 (solid dots) outside of the training file becomes apparent.
Comparisons between distributions of aircraft data demonstrate the invention's capability of determining the amount of useful in formation obtained in one distribution when compared to another. Each data point may have not two but hundreds of variables associated with it. One of the advantages in using a pattern recognition algorithm is the capability of analyzing data points associated with large numbers of variables in multi-dimensional space. For ease of illustration, comparison examples used herein are limited to graphs of two significant variables. Figure 5 shows comparison made between two aircraft flights in which data pertaining to the aircraft engine was accumulated. Data distributions of three variables (engine power level angle, fan rotor speed Nl and high pressure rotor speed N2) were utilized to illustrate the amount and location of useful data gathered in Flight XXA (test file) when compared to the engine data gathered in Flight XXB (training file). Figure 5 represents the interactions between one of the two variable combinations. The amount of new or useful information 530 (dashes) in the test file 520 (hollow dots) and 530 when compared to the training file 510 (solid dots) was estimated to be approximately 19%.
If Flight XXA had been flown first instead of Flight XXB and the flight engineers were interested in determining the amount of new or useful information gathered in Flight XXB (pertaining to the same three variables), Flight XXA data would be placed in the training file and Flight XXB data would be placed in the test file. Figure 6 represents the interactions between the same two variable combinations when Flight XXB was used as the test file. The amount of new or useful information 630 (dashes) in the test file Flight XXB when compared to the training file Flight XXA (both solid and hollow dots) was estimated to be approximately 16%.
To demonstrate the ability of using this approach to identify data outside of a training file in a higher (five dimension) multivariate situation; a training file of one flight maneuver (during a flight test) was compared to a test file of another maneuver using data containing five variables associated with each data point. Run G of a particular flight, maneuver 9G wind up turn (planned mach 1.1 at 15000 feet) was used as the training file and a second flight, Run M, maneuver 5.86 g 360 roll (L) (planned mach 1.2 at 5000 feet) was used as the test file. Training and test file variables included vertical load (N2), lateral acceleration (Ny), roller acceleration, true airspeed, and angle of attack. Analysis indicated that approximately 53 per cent of the observations in the test file were outside the framing file. Figures 7 A through 7C represents the interactions between three of the variable combinations. In each of these examples, significant new data 730 (dashes) outside the training file 710 (solid dots) is identified.
To further illustrate the present invention, a flight test was conducted to gather aircraft structural data. Available flight test data was composed of 16 flights comprised of 163 maneuvers and 47,711 observations. Observational data consisted of 28 independent variables (airspeed, aircraft gross weight, angle of attack, etc.) and 33 dependent variables (vertical tail shear, center fuselage end moment, etc.). If each subsequent flight is compared with a training file composed of the previous flights, an indication is obtained of how much new and useful data is in each flight as the test is in progress. Figure 8 illustrates a data- subset analysis figure depicting a comparison of one flight vs. a training file of the previous flights taken. Subsets 810, 815 and 820 represent the training file data, wherein Subset 830 represents the percent of information from the test file that is outside the training file.
Separately, Figure 9 provides another illustration of the amount of useful information from a previously recorded flight test, in each flight test when compared to a training file composed of all previous flights taken As a test program progresses and as more data is added to the training file, the amount of new data in each flight test gradually decreases. Figure 10 illustrates the percent of useful information in each flight test (Y axis) when compared to a training file of all previous flights taken (X axis). As the entire envelope of performance is mapped into the training file, it becomes much more difficult for subsequent test flights to add any new information. When a flight test program reaches the point at which subsequent test flights are not able to add a significant amount of new information, consideration should be given to end the flight test program.
The present invention can be used with a flight simulation to reduce flight test costs. The strategy for reducing flight test cost involves reducing the number of specific maneuvers, which contribute no new or useful information. If each maneuver is compared to a training file of all the other remaining test maneuvers combined, the resulting estimate of new information in each maneuver provides an indication of the individual maneuver's data gathering cost effectiveness. Maneuvers which display little or no significant amount of useful information or contribution to the program are eliminated from the training file before the next comparison is made until all non-contributing maneuvers are eliminated.
To illustrate this strategy, data associated with store configuration QA included within the program was analyzed to obtain a list of store configuration QA maneuver elimination candidates. Store configuration QA data consisted of 6 of 16 flight test program flights, 65 of 163 maneuvers, and 11,968 of 47,711 observation points. Observational data consisted of 28 independent variables (airspeed, aircraft gross weight, angle of attack, etc.) and 33 dependent variables (vertical tail shear, center fuselage end moment, etc.). Since only independent variable data is gathered in flight simulation, only independent variable data was used in the analysis. If simulation is used to gather data, the amount of data pertaining to dependent variables can be estimated due to the correlation between dependent and independent variables. Figure 11 depicts a graph illustrating the close relationship between the amount of new information in independent (y-axis) when compared to percent of useful information in independent variables (x-axis). New independent variable information gathered in simulation means that new dependent variable information would be gathered in flight test.
When the present invention was applied to store configuration QA flight test data (simulation data was not available), 27 out of 65 or 41.54% of the maneuvers were deemed to provide an insignificant amount of new data and considered strong candidates for elimination providing a safety risk was not incurred (see Figures 12A and B). Figure 12A illustrates maneuver candidates for elimination. Figure 12B illustrates different maneuver candidates 1205 for elimination. Data sets were considered to provide an insignificant amount of new data if the amount of new data in the data set was less than or equal to 0.5%. The percentage of 0.5% was selected since 0.5% represented less than one data point.
Flight A Run 12 was one of the maneuvers identified as containing no useful data (i.e., zero new observations) when compared to a database of the other maneuvers. Figure 13 depicts a typical combination of variables which shows that all of the data in the test file 1320 is indeed included in the training file 1310. Other two variable combinations provide similar results.
Using the same statistical parameters Flight A Run 12 maneuver was found to have at least one useful data point 1430 (dashes) when compared to the training file data base 1410 (solid dots) and this useful point is illustrated (see Figure 14) using the same combination of variables. Figure 14 illustrates at least one useful data point using the statistical parameters Flight A Run 12 maneuver when compared to the training file database using the same combination of variables;
One of the major concerns when using the present invention's reduction methods involves the notion of having to gather significant amounts of redundant data until an envelope of performance is obtained. Figure 15 provides a simple example of data gathered (in alphabetical sequence) from three hypothetical flights. If we were to apply our pattern recognition data gathering optimization analysis to this situation it would most likely find that flights A and B contain significant amounts of redundant information and are candidates for elimination. However, flights A and B were most likely needed to accomplish flight C so no reduction in flight test program cost's could have been obtained by using this method and only performing flight C.
In response to this concern, the present invention's pattern recognition flight test program cost reduction techniques would probably not provide much useful information if flight test programs could be planned using a small, manageable number of variables or sensor inputs. Unfortunately, the number of independent and dependent variables being recorded in flight-test sensors and flight recorders grows larger and larger with subsequent flight test programs. Determining if an aircraft is operating in the
Figure 15 A, B, or C range becomes a difficult task when 100 variables resulting in 4950 combinations of two variables are interacting. As the number of variables increases, the need for using pattern recognition data gathering optimization analysis to assist the flight test planner in gathering flight test data in the most cost effective manner increases. In the above mentioned specific aircraft's store configuration flight test, 41.5% of the maneuvers showed no useful information. Most of these maneuvers were located at the beginning and at the end of the flight test program. Maneuvers located at the end of the test flight program, which provided no useful information, served little data gathering purpose and might have been considered for elimination had this tool been available. This would have reduced the flight test program by a conservative 20%. Figure 16 illustrates the percentage of accumulated maneuvers selected as candidates for elimination (Y axis) as the QA store configuration flight test program progressed (X axis). The accumulated percentage of maneuvers possessing redundant information (candidates for elimination) did not increase as the program progressed but fluctuated around 40 percent throughout the test program. As mentioned previously, when the present invention was applied to store configuration QA flight test data, 27 out of 65 or 41.54% of the maneuvers were deemed to provide an insignificant amount of new data and considered strong candidates for elimination providing a safety risk was not incurred (see Table 12A and B). The present invention allows this information to be gathered using flight simulation before the flight test program begins and if all of the 27 non-contributing maneuvers could be removed (i.e., never flown), the amount of data in the flight test program would be reduced by 41.1% and the number of maneuvers by 41.54%. In this case of the example, each removed data point has 28 independent variables associated with it. Figures 17A through 17H represent several store configuration 29 A flight test two variable before/after comparisons when all of the data was included and when the data from the 27 non-contributing maneuvers was removed. Significant points (points on the edge perimeters) remain even though 41.1 % of the data points have been removed.
Refer to Figure 18, which illustrates in flow chart form, the methodology of the present invention. In step 1, a request is submitted for a system's test, in this case, a flight test for a specific aircraft. Once submitted, the flight test data gathering goals and requirements are defined in step 2. The flight test, which is composed of numerous flights and maneuvers designed to meet the data gather goals and requirements, is proposed to the flight test engineering, analysis and design group in step 3. Using their experience, the flight test engineers will design the flight test. In step 4, the proposed flight, as designed by the flight test engineering group is executed, with the various flights and maneuvers, using a simulator and the flight data is gather for analysis.
In step 5, the flight simulation data is evaluated using a data gathering optimization statistical algorithm to determine which maneuvers provide little information and should be considered as candidates for removal from the flight test. Additionally, the flights and maneuvers will be prioritized in relationship to the data gather effectiveness. In step 6, non-data gathering maneuvers from the flight test are identified as possible candidates for removal from the flight test. These are presented to the flight test engineering, analysis and design group for evaluation. Next, in step 7, the flight test engineering group determines which maneuver candidates may be eliminated and re-design the flight test using the prioritized data gathering information determined in step 5, so that the flight test is conducted in the most efficient data gathering fashion.
In step 8, the actual flight test is conducted with the reduced number of flights and maneuvers to gather the actual flight test data. In step 9, the actual flight test data gathered in step 8 is analyzed while the test is in progress to determine the actual amount of significant data being gathered, using the data gathering optimization techniques in step 5. After the completion of the flight test, the completed flight test data is also analyzed to determine the accuracy of earlier estimations based on the flight simulation data.
Figure 19 illustrates step 5 of Figure 18 in greater detail. In step 5.1, each flight simulation data point, which comprises of a vector composed of numerous independent variables, such as velocity, altitude, and a flight time marker indicating when the data was gathered during the flight simulation's various flights and maneuvers is collected during the simulation of the flight. In step 5.2, the flight simulation data for each flight test maneuver is converted into a database spreadsheet file with the flight test flight number, maneuver number, and time marker placed in separate columns for identification purposes. A separate row in the spreadsheet can be assigned for each data vector gathered and a separate column in the spreadsheet can be assigned for each independent variable.
In step 5.3, the spreadsheet information for each maneuver is combined into one large database file. A typical large database spreadsheet file might contain over 50,000 rows and 50 columns. In step 5.4, the importance of each independent variable associated with each vector is determined. A determination is also made for the number of independent variable alarms criteria for identifying when a vector associated with a test file is considered to be outside the data region associated with a training file. Next, in step 5.5, a determination is made for the amount of significant data criteria used in identifying those maneuvers which are non-contributing data gatherers when compared to the other maneuvers of the test flight. The significant data criteria will comprise the percent of data in a test file, which is considered to be outside the data region associated with a training file.
Next, in step 5.6, one maneuver's worth of data information is removed from the main database file. This set of data is the test file. The remaining portion of the database information file is the training file. The test and training file is converted into text files for analysis and evaluation by a subsequent advanced pattern recognition algorithm such as the Multivariate State Estimation Technique (MSET) developed by Argonne National Laboratory.
In step 5.7, the basic description of determining which vectors within the test file are in variance with normal or outside the region of vectors and thus represent new data contained in the training file is determined. This is accomplished by setting up a system model that provides analytically derived values of all monitored sensors is integrated with a statistically-based hypothesis test that compares the analytically estimated signal value with the test file measured values to detect the development or presence of new data. Three essential models of the flight for the algorithm include a training algorithm for the collection of representative data from sensors during normal system operations, an empirically based model for system state estimation, and a statistically based fault detection system.
The training model is used to produce a training set or set points whose data encompasses all expected normal operating states of the system, ideally. The system-modeling module (the test file) is used to estimate the values of all signals that are present in the process that is being monitored. The fault detection model is used to detect disturbances through an examination of the difference between the estimated and measured signal values. An algorithm (MSET for example) is used to detect incipient faults between current and normal behavior. However, instead of detecting incipient faults, the algorithm will be estimating which vectors in the test file exhibit abnormal behavior or could be considered new or useful data points when compared with the vector data stored in the training file. In step 5.8, a test file is created. This is accomplished by using the data associated with one maneuver from the training file database. The data associated with the test file is removed in step 5.8.1 from the training file for the purpose of analyzing the maneuver. Next in step 5.8.2, the test file is deliberately seeded with certain data known to be outside the training file. Next, in step 5.8.3, a comparison of the test file to the training file is made using the advanced pattern recognition algorithm. After the comparison is made, then in step 5.8.4, the pattern recognition algorithm parameters, such as wrongful rejection probability, wrongful acceptance probability, sequential probability ratio test sensitivity, is adjusted to ensure that the algorithm is operating at optimal performance when detecting the data known to be outside the training file.
Next in step 5.9, once the algorithm parameters, the number of independent variable alarm criteria, and the non-data gathering criteria for maneuvers have been established, the test file, seeded with false data for determining optimal the algorithm parameter values, is returned back into the training master database in its original state, i.e., with the seeded data removed. At this point the algorithm is ready to begin the complete analysis of each maneuver of the simulated flight data or system test.
Step 5.10 runs the analysis of the flight tests maneuvers. In step 5.10.1, the data associated with a specific maneuver is removed from the training file and placed in a test file.
The training file is normalized or fit to a normal distribution. If the maneuver's test file is found to be a non-contributor of data, that maneuvers data is removed from the training file and the process of ranking the maneuvers in terms of data gathering capabilities starts all over again.
Next, in step 5.10.2, the test file is compared using the algorithm to the training file associated with the data of all remaining maneuvers combined to determine the variance of the maneuver from normal. Once the comparison is performed, in step 5.10.3, the data associated with the maneuver is returned to the training file and the next maneuver is removed and placed into the test file. This process is repeated until all maneuvers have had their data placed in the test file and compared to the training file comprising of data from all the other non-test file maneuvers. The resulting estimate of new information from each maneuver provides an indication of the individual maneuver's data gathering cost effectiveness, A non-contributing data-gathering maneuver contains predominantly redundant information when compared to the data from the rest of the maneuvers in the training file.
After the comparison is completed in step 5.10, a table, is prepared comprising of the flight test maneuvers and the percent of useful information contained within each maneuver. The percent of useful information is calculated by determining the number of vectors, which meet the number of independent variable alarm criteria, divided by the total number of vectors in the maneuver. The flight test engineers will then select the maneuver with the smallest percent contribution of useful information. If this value is less than the previously agreed upon value for the criteria for identifying a non-contributing data gathering maneuver, then the flight test engineer may remove the maneuver's data from the master training file.
Next, in step 5.12, a record is maintained of those maneuver vector values associated with the variable alarms for the non-contributing data gathering maneuver's data, which has been removed from the training file for future analysis. When a maneuver gathers only one significant data vector, which provides an indication that data is outside the training file range, the vector value may provide cπtical information and justify maintaining the maneuver in the flight test even though only one significant vector value was gathered.
After the first non-contributing data gathering maneuver has been identified and removed from the master training data file, then in step 5 13, the process of analyzing the simulated data should be repeated (from step 5.9 through 5 11). Subsequently, each time a non-contributing data-gathering maneuver is removed, the analysis of the training file against a test file for each maneuver as described above, should be repeated to ensure that all non-contπbutmg data is accurately analyzed and identified. Finally, in step 5 14, a listing of all non-contributing data gathering maneuvers and maneuver vector values associated with variable alarms are prepared for presentation to the flight test engineers This list will show the maneuver candidates for flight test elimination.
Figure 20 illustrates step 7 of Figure 18 m greater detail. In step 7.1, the flight test engineering, analysis and design groups receive a listing of non-data gatheπng maneuvers identified as possible candidates for removal from the flight test. The flight test engineers determine which non-data gathering maneuvers may be eliminated from the flight test m step 7 2. Criteria for elimination may include the lack of data gathered, the importance of the minimal data gathered, or flight test safety aspects if the maneuver is eliminated, as well as others.
Once the flight test engineers have made the determination as described in step 7 2, then a final framing file database composed of only those maneuvers selected for the final test flight is created in step 7.3. A table listing the percent of useful information contained within each final flight test maneuver using steps 5.9 through 5.10 is prepared in step 7.4. The list of maneuvers is prioritized showing which maneuvers gather the most useful data and the data gatheπng cost effectiveness of each maneuver The flight test engineers redesign the reduced size flight test to accommodate the elimination of the non-data gathering maneuvers chosen for removal. The prioritized list of maneuvers indicating each maneuver's data gatheπng effectiveness can also be utilized in the flight test redesign to ensure that the flight test redesign is conducted in the most efficient data gathering fashion.
Figure 21 illustrates step 9 of Figure 18 m greater detail. In step 9.1, actual flight test data with maneuver markers and time markers is translated into various database files and test files for subsequent processing. Each subsequent maneuver with a framing file composed of the previous maneuvers is compared to obtain an estimate of how much new and useful data is being gathered duπng the flight test in step 9.2. As the entire envelop of performance is mapped into the framing file, it becomes difficult for subsequent maneuvers to add any new information and the amount of new information being added should decrease.
Upon completion of the flight test and using the parameters set up from step 5, the percentage of useful dependent and independent information in each maneuver is determined, when compared to a training file composed of data from all flight test maneuvers in step 9.3. The independent variable useful information values obtained for each maneuver in actual flight test to those obtained in flight simulation to determine the accuracy of the flight.
Next, in step 9.4, graphs of various variable combinations depicting the training file data and test file data are prepared to confirm that the test file data identified as useful is visually outside the training file region. The correlation coefficient between independent variable new information obtained versus dependent variable new information is determined to obtain enhanced future estimates. Finally, in step 9.6, the test files of flight simulation data is compared to the training file of actual flight test data using the parameters set in step 5. Those regions of the simulation that need adjustment to better represent actual aircraft performance is highlighted, and the flight simulation is enhanced by adjusting the flight simulation parameters.
The following examples illustrate the present invention's ability to detect new information in a test file when using aircraft data distributions. Figure 4 shows a comparison of a theoretical distribution vs. a distribution of aircraft data. A fraining file composed of data variables (fan rotor speed vs. EPLA engine power level angle) taken from an aircraft flight is compared to a test file composed of the same two variables arranged in a diagonal line (moving randomly off center). As the test file line intersects the training file data, the portion of the data from the diagonal line test file (highlighted in pink) outside of the training file becomes apparent.
Comparisons between distributions of aircraft data demonstrate the invention's capability of determining the amount of useful in formation obtained in one distribution when compared to another. Each data point may have not two but hundreds of variables associated with it. One of the advantages in using a pattern recognition algorithm is the capability of analyzing data points associated with large numbers of variables in multi-dimensional space. For ease of illustration, comparison examples will be limited to graphs of two significant variables.
Having thus described our invention, what we claim as new and desire to secure by Letters Patent is set forth in the following claims.

Claims

WHAT IS CLAIMED IS:
1. A method for optimizing data gathering, the method comprising the steps of: designing a first system test of two or more maneuvers; compiling a one or more data variable for each of the two or more maneuvers from the first system test; pooling each data variable from each maneuver into a first set of data; removing a first maneuver's data variables from the pooled first set of data; comparing the first maneuver's data variables against the first set of data with the first maneuver's data variables removed; and determining whether to perform the first maneuver based on whether the maneuver provided new data.
2. The method is claim 1 further comprising the steps of: simulating the first system test by simulating each of the two or more maneuvers; and using the simulation to calculate each one or more data variable for each of the one or more maneuver.
3. The method of claim 1 further comprising the step of removing the first maneuver from the first system test if the maneuver provided no new data.
4. The method of claim 3 further comprising the steps of: removing a next successive maneuver's data variables from the pooled first set of data; comparing the next successive maneuver's data variables against the first set of data with the next successive maneuver's data variables removed; and determining whether to perform the next successive maneuver based on whether the next successive prearranged maneuver provided new data.
5. The method of claim 4 further comprising the step of removing the next successive maneuver from the first system test if the next successive maneuver provided no new data.
6. The method of claim 1 further comprising the steps of: normalizing the pooled first set of data; and determining a variance of a first maneuver data's from the normalized first set of data.
7. The method of claim 6 further comprising the step of establishing a predetermined set point, whereby if the variance exceeds the set point, the first maneuver variable data is considered new data.
8. The method of claim 1 wherein the first system test is a flight test.
9. The method of claim 1, wherein the comparing step further comprises the step of using the MSET algorithm.
10. The method of claim 1 wherein the comparing step further comprises the step of using a pattern recognition algorithm analysis to determine if the maneuver's variable data varies from the first set of data.
11. A method for optimizing data gathering, the method comprising the steps of: performing a first system test; compiling a first set of test data from the first system test; designing a second system test; compiling a second set of test data from the second system test; comparing the first data set to the second data set; and determining whether to perform the second test based on the comparing step.
12. The method of claim 11 further comprising the steps of: designing a series of one or more maneuvers for the second system test; performing each one or more maneuvers as part of the second system test; and compiling a one or more data variable from each of the one or more maneuvers.
13. The method of claim 12 wherein the comparing step further comprising the steps of: comparing a first maneuver's data variables against the first set of data; and determining whether to perform the first maneuver based on whether the maneuver provided new data.
14. The method is claim 12 further comprising the steps of: simulating the second system test by simulating each one or more maneuver; and calculating each one or more data variable for each of the one or more simulated maneuver.
15. The method of claim 13 further comprising the steps of: removing the first maneuver from the second system test if the maneuver provided no new data; comparing a next successive maneuver's data variables against the first set of data; and determining whether to perform the next successive maneuver based on whether the next successive maneuver provided new data.
16. The method of claim 11 further comprising the steps of: normalizing the first set of data; and determining the first maneuver variable data's variance from the normalized first set of data.
17. The method of claim 16 further comprising the step of setting a predetermined set point, whereby the first prearranged maneuver variable data is considered new data if the variance exceeds the set point,.
18. The method of claim 11 wherein the first system test is a flight test.
19. The method of claim 11, wherein the comparing step further comprises the step of usin the MSET algorithm.
20. The method of claim 11 wherein the comparing step further comprises the step of using statistically analysis to determine if the second set of prearranged maneuver's variable data varies from the first set of data.
21. A system for optimizing data gathering comprising of: a first set of data gathered from a first system test; an second set of data from a second system test; and an algorithm for comparing the first set of data to the second set of data, wherein the algorithm provides an optimization result.
22. An apparatus for optimizing data gathering comprising of: a computer; a first data set from a first system test; an second set of data from a second system test; and an algorithm running on the computer, wherein the algorithm compares the first set of data to the second set of data and provides a result.
PCT/US2002/033084 2001-10-15 2002-10-15 Method to optimize test data WO2003034167A2 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
EP02776224A EP1436679B1 (en) 2001-10-15 2002-10-15 Method to optimize test data
DE60208415T DE60208415T2 (en) 2001-10-15 2002-10-15 METHOD FOR OPTIMIZING TEST DATA

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US09/978,817 2001-10-15
US09/978,817 US7401013B2 (en) 2001-10-15 2001-10-15 Method to optimize test data

Publications (2)

Publication Number Publication Date
WO2003034167A2 true WO2003034167A2 (en) 2003-04-24
WO2003034167A3 WO2003034167A3 (en) 2004-02-26

Family

ID=25526416

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/US2002/033084 WO2003034167A2 (en) 2001-10-15 2002-10-15 Method to optimize test data

Country Status (4)

Country Link
US (1) US7401013B2 (en)
EP (1) EP1436679B1 (en)
DE (1) DE60208415T2 (en)
WO (1) WO2003034167A2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8906353B2 (en) 2008-09-23 2014-12-09 Nektar Therapeutics Compositions and methods for achieving sustained therapeutic drug concentrations in a subject

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1706718A2 (en) * 2004-01-14 2006-10-04 ABB Inc. Method and apparatus to diagnose mechanical problems in machinery
WO2007109694A2 (en) * 2006-03-20 2007-09-27 Vincent Granville Scoring quality of traffic to network sites using interrelated traffic parameters
US7769562B2 (en) * 2006-11-13 2010-08-03 Oracle America, Inc. Method and apparatus for detecting degradation in a remote storage device
US8042073B1 (en) 2007-11-28 2011-10-18 Marvell International Ltd. Sorted data outlier identification
FR3050351B1 (en) * 2016-04-15 2018-05-11 Thales AIRCRAFT AVIONICS INTEGRITY MONITORING METHOD, APPARATUS AND COMPUTER PROGRAM PRODUCT THEREOF
CN106114876B (en) * 2016-08-25 2019-04-02 中国商用飞机有限责任公司 The parameter identification method of Aircraft Flight Test data
CN109325059A (en) * 2018-12-03 2019-02-12 枘熠集成电路(上海)有限公司 A kind of data comparing method and device
US11379578B1 (en) 2020-10-16 2022-07-05 Trend Micro Incorporated Detecting malware by pooled analysis of sample files in a sandbox

Family Cites Families (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4442491A (en) * 1981-06-23 1984-04-10 General Dynamics Corporation Training evaluation process
US4646241A (en) * 1984-06-21 1987-02-24 United Technologies Corporation Solid-state flight data recording system
US5270642A (en) 1992-05-15 1993-12-14 Hewlett-Packard Company Partitioned boundary-scan testing for the reduction of testing-induced damage
GB9514008D0 (en) * 1995-07-10 1995-09-06 Secr Defence Sensor fault detection system
US5742513A (en) 1996-05-15 1998-04-21 Abb Power T&D Company Inc. Methods and systems for automatic testing of a relay
US5908176A (en) * 1997-01-14 1999-06-01 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration In-flight adaptive performance optimization (APO) control using redundant control effectors of an aircraft
WO1999003000A1 (en) * 1997-07-09 1999-01-21 Massachusetts Institute Of Technology Integrated flight information and control system
US6199013B1 (en) 1997-07-15 2001-03-06 Navigation Technologies Corp. Maneuver generation program and method
US6216063B1 (en) * 1998-05-06 2001-04-10 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration On-line μ method for robust flutter prediction in expanding a safe flight envelope for an aircraft model under flight test
DE19850612C1 (en) * 1998-11-03 2000-02-24 Daimler Chrysler Ag Method of matching simulation system to regulated reference system involves transferring non-feedback parameters of second loop to reference system and matching feedback parameters
US6446015B1 (en) * 2000-02-23 2002-09-03 The United States Of America As Represented By The Administrator Of The National Aeronautics And Space Administration Method and apparatus for flight data acquisition using an optimized multiple frequency waveform
US6609036B1 (en) * 2000-06-09 2003-08-19 Randall L. Bickford Surveillance system and method having parameter estimation and operating mode partitioning
US6529821B2 (en) * 2001-06-05 2003-03-04 The United States Of America As Represented By The Secretary Of The Navy Route planner with area avoidance capability

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
None

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8906353B2 (en) 2008-09-23 2014-12-09 Nektar Therapeutics Compositions and methods for achieving sustained therapeutic drug concentrations in a subject
US9801873B2 (en) 2008-09-23 2017-10-31 Nektar Therapeutics Methods for treating humans having brain cancer
US10525051B2 (en) 2008-09-23 2020-01-07 Nektar Therapeutics Compositions and methods for achieving sustained therapeutic drug concentrations in a subject

Also Published As

Publication number Publication date
US7401013B2 (en) 2008-07-15
DE60208415D1 (en) 2006-02-02
WO2003034167A3 (en) 2004-02-26
EP1436679A2 (en) 2004-07-14
US20030074169A1 (en) 2003-04-17
DE60208415T2 (en) 2006-09-07
EP1436679B1 (en) 2005-12-28

Similar Documents

Publication Publication Date Title
EP3540532B1 (en) Control system and control method
EP3547057A1 (en) Abnormality detection system, support device, and abnormality detection method
US20110035188A1 (en) Method and apparatus for analyzing time series data
CN111459700A (en) Method and apparatus for diagnosing device failure, diagnostic device, and storage medium
US7401013B2 (en) Method to optimize test data
US20200050218A1 (en) Computer-implemented method and system for evaluating uncertainty in trajectory prediction
US11423321B2 (en) Method and system for predicting system status
CN107944005B (en) Data display method and device
CN114266289A (en) Complex equipment health state assessment method
EP3190519A1 (en) Monitoring device and monitoring method thereof, monitoring system, and recording medium in which computer program is stored
Patel et al. Doctor for machines: a failure pattern analysis solution for industry 4.0
CN117171896A (en) Method and system for optimizing machining of input shaft of automobile steering gear
CN115659271A (en) Sensor abnormality detection method, model training method, system, device, and medium
CN113487141B (en) Bearing state evaluation method based on multisource information clustering fusion
JPS6126112A (en) System for diagnosing abnormality of system
US20190073841A1 (en) Method for testing the integrity of the avionics of an aircraft, associated device and computer program product
CN112379656A (en) Processing method, device, equipment and medium for detecting abnormal data of industrial system
Daouayry et al. Data-centric helicopter failure anticipation: The mgb oil pressure virtual sensor case
CN112559477B (en) Quantitative evaluation method, system, equipment and medium based on unmanned aerial vehicle flight log
CN110928269A (en) Degradation acceleration test optimization design method and system based on inertial navigation platform
JP2715904B2 (en) Computer system performance evaluation device
EP4300232A1 (en) Information processing device, information processing program, and information processing method
JPH0713617A (en) Cause estimating method for nonconformity event
CN117290747B (en) Eye movement data-based flight state monitoring method, storage medium and electronic equipment
WO2022180681A1 (en) Data generation system, data generation method, and data generation program

Legal Events

Date Code Title Description
AK Designated states

Kind code of ref document: A2

Designated state(s): AE AG AL AM AT AU AZ BA BB BG BY BZ CA CH CN CO CR CU CZ DE DM DZ EC EE ES FI GB GD GE GH HR HU ID IL IN IS JP KE KG KP KR LC LK LR LS LT LU LV MA MD MG MN MW MX MZ NO NZ OM PH PL PT RU SD SE SG SI SK SL TJ TM TN TR TZ UA UG UZ VC VN YU ZA ZM

AL Designated countries for regional patents

Kind code of ref document: A2

Designated state(s): GH GM KE LS MW MZ SD SL SZ UG ZM ZW AM AZ BY KG KZ RU TJ TM AT BE BG CH CY CZ DK EE ES FI FR GB GR IE IT LU MC PT SE SK TR BF BJ CF CG CI GA GN GQ GW ML MR NE SN TD TG

121 Ep: the epo has been informed by wipo that ep was designated in this application
DFPE Request for preliminary examination filed prior to expiration of 19th month from priority date (pct application filed before 20040101)
WWE Wipo information: entry into national phase

Ref document number: 2002776224

Country of ref document: EP

WWP Wipo information: published in national office

Ref document number: 2002776224

Country of ref document: EP

WWG Wipo information: grant in national office

Ref document number: 2002776224

Country of ref document: EP

NENP Non-entry into the national phase

Ref country code: JP

WWW Wipo information: withdrawn in national office

Country of ref document: JP