US8745215B2 - Network delay analysis including parallel delay effects - Google Patents

Network delay analysis including parallel delay effects Download PDF

Info

Publication number
US8745215B2
US8745215B2 US13/333,955 US201113333955A US8745215B2 US 8745215 B2 US8745215 B2 US 8745215B2 US 201113333955 A US201113333955 A US 201113333955A US 8745215 B2 US8745215 B2 US 8745215B2
Authority
US
United States
Prior art keywords
delay
components
group
application
delay components
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
US13/333,955
Other versions
US20130067073A1 (en
Inventor
Steven Niemczyk
Patrick J. Malloy
Alain J. Cohen
Russell Mark Elsner
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Riverbed Technology LLC
Opnet Technologies Inc
Original Assignee
Riverbed Technology LLC
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Riverbed Technology LLC filed Critical Riverbed Technology LLC
Priority to US13/333,955 priority Critical patent/US8745215B2/en
Assigned to MORGAN STANLEY & CO. LLC reassignment MORGAN STANLEY & CO. LLC SECURITY AGREEMENT Assignors: OPNET TECHNOLOGIES, INC., RIVERBED TECHNOLOGY, INC.
Publication of US20130067073A1 publication Critical patent/US20130067073A1/en
Assigned to OPNET TECHNOLOGIES, INC. reassignment OPNET TECHNOLOGIES, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ELSNER, RUSSELL MARK, MALLOY, PATRICK J., COHEN, ALAIN J., NIEMCZYK, STEVEN
Assigned to OPNET TECHNOLOGIES LLC reassignment OPNET TECHNOLOGIES LLC CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: OPNET TECHNOLOGIES, INC.
Assigned to RIVERBED TECHNOLOGY, INC. reassignment RIVERBED TECHNOLOGY, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: OPNET TECHNOLOGIES LLC
Assigned to RIVERBED TECHNOLOGY, INC. reassignment RIVERBED TECHNOLOGY, INC. RELEASE OF PATENT SECURITY INTEREST Assignors: MORGAN STANLEY & CO. LLC, AS COLLATERAL AGENT
Assigned to JPMORGAN CHASE BANK, N.A., AS ADMINISTRATIVE AGENT reassignment JPMORGAN CHASE BANK, N.A., AS ADMINISTRATIVE AGENT PATENT SECURITY AGREEMENT Assignors: RIVERBED TECHNOLOGY, INC.
Publication of US8745215B2 publication Critical patent/US8745215B2/en
Application granted granted Critical
Assigned to RIVERBED TECHNOLOGY, INC. reassignment RIVERBED TECHNOLOGY, INC. RELEASE OF SECURITY INTEREST IN PATENTS Assignors: BARCLAYS BANK PLC
Assigned to MORGAN STANLEY SENIOR FUNDING, INC., AS COLLATERAL AGENT reassignment MORGAN STANLEY SENIOR FUNDING, INC., AS COLLATERAL AGENT SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: RIVERBED TECHNOLOGY, INC.
Assigned to RIVERBED TECHNOLOGY, INC. reassignment RIVERBED TECHNOLOGY, INC. CORRECTIVE ASSIGNMENT TO CORRECT THE CONVEYING PARTY NAME PREVIOUSLY RECORDED ON REEL 035521 FRAME 0069. ASSIGNOR(S) HEREBY CONFIRMS THE RELEASE OF SECURITY INTEREST IN PATENTS. Assignors: JPMORGAN CHASE BANK, N.A.
Assigned to ALTER DOMUS (US) LLC, AS COLLATERAL AGENT reassignment ALTER DOMUS (US) LLC, AS COLLATERAL AGENT PATENT SECURITY AGREEMENT Assignors: RIVERBED TECHNOLOGY, INC.
Assigned to MACQUARIE CAPITAL FUNDING LLC reassignment MACQUARIE CAPITAL FUNDING LLC SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ATERNITY LLC, RIVERBED HOLDINGS, INC., RIVERBED TECHNOLOGY, INC.
Assigned to RIVERBED HOLDINGS, INC., ATERNITY LLC, RIVERBED TECHNOLOGY, INC. reassignment RIVERBED HOLDINGS, INC. RELEASE OF SECURITY INTEREST IN PATENTS RECORED AT REEL 056397, FRAME 0750 Assignors: MACQUARIE CAPITAL FUNDING LLC
Assigned to MORGAN STANLEY SENIOR FUNDING, INC., AS COLLATERAL AGENT reassignment MORGAN STANLEY SENIOR FUNDING, INC., AS COLLATERAL AGENT PATENT SECURITY AGREEMENT SUPPLEMENT - FIRST LIEN Assignors: ATERNITY LLC, RIVERBED HOLDINGS, INC., RIVERBED TECHNOLOGY, INC.
Assigned to ALTER DOMUS (US) LLC, AS COLLATERAL AGENT reassignment ALTER DOMUS (US) LLC, AS COLLATERAL AGENT PATENT SECURITY AGREEMENT SUPPLEMENT - SECOND LIEN Assignors: ATERNITY LLC, RIVERBED HOLDINGS, INC., RIVERBED TECHNOLOGY, INC.
Assigned to WILMINGTON TRUST, NATIONAL ASSOCIATION reassignment WILMINGTON TRUST, NATIONAL ASSOCIATION PATENT SECURITY AGREEMENT Assignors: ATERNITY LLC, RIVERBED TECHNOLOGY, INC.
Assigned to WILMINGTON TRUST, NATIONAL ASSOCIATION, AS U.S. COLLATERAL AGENT reassignment WILMINGTON TRUST, NATIONAL ASSOCIATION, AS U.S. COLLATERAL AGENT SECURITY INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ATERNITY LLC, RIVERBED TECHNOLOGY LLC (FORMERLY RIVERBED TECHNOLOGY, INC.)
Assigned to ATERNITY LLC, RIVERBED TECHNOLOGY, INC. reassignment ATERNITY LLC TERMINATION AND RELEASE OF SECURITY INTEREST IN PATENTS Assignors: MORGAN STANLEY SENIOR FUNDING, INC., AS COLLATERAL AGENT
Assigned to RIVERBED TECHNOLOGY, INC., ATERNITY LLC reassignment RIVERBED TECHNOLOGY, INC. TERMINATION AND RELEASE OF SECURITY INTEREST IN PATENTS Assignors: ALTER DOMUS (US) LLC, AS COLLATERAL AGENT
Assigned to RIVERBED TECHNOLOGY, INC., ATERNITY LLC reassignment RIVERBED TECHNOLOGY, INC. TERMINATION AND RELEASE OF SECURITY INTEREST IN PATENTS Assignors: WILMINGTON TRUST, NATIONAL ASSOCIATION, AS U.S. COLLATERAL AGENT
Assigned to RIVERBED TECHNOLOGY LLC reassignment RIVERBED TECHNOLOGY LLC CHANGE OF NAME (SEE DOCUMENT FOR DETAILS). Assignors: RIVERBED TECHNOLOGY, INC.
Assigned to RIVERBED TECHNOLOGY, INC., RIVERBED HOLDINGS, INC., ATERNITY LLC reassignment RIVERBED TECHNOLOGY, INC. RELEASE BY SECURED PARTY (SEE DOCUMENT FOR DETAILS). Assignors: ALTER DOMUS (US) LLC, AS COLLATERAL AGENT
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/08Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters
    • H04L43/0852Delays
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/04Processing captured monitoring data, e.g. for logfile generation
    • H04L43/045Processing captured monitoring data, e.g. for logfile generation for graphical visualisation of monitoring data
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/28Timers or timing mechanisms used in protocols

Definitions

  • This invention relates to the field of network analysis, and in particular to a system and method for analyzing and assessing the effects of parallel delays within an application.
  • Proper management of a network generally requires assuring that the network is performing satisfactorily for the users of the network, modifying the network to address performance issues or problems, and planning for future improvements to the network as demand increases and as newer technologies and alternatives become available.
  • a variety of tools have been developed, and continue to be developed, to facilitate the management of communication networks, and in particular for managing networks that provide communications among computer devices. Many of these tools are configured to model the network's performance under a variety of traffic conditions, both real and hypothesized, and in many cases, base this performance on data collected from the actual network.
  • One of the primary parameters for analyzing or assessing the performance of a network is the time it takes for messages to reliably reach their destination. This time is dependent upon a variety of factors.
  • the message is typically partitioned into transmission elements, herein termed packets for convenience.
  • packets Each packet must enter the network, and incurs a delay as it gains network access. When it enters the network, it incurs a delay that is dependent upon the bandwidth available at each link along its path to its destination. It may also incur queuing delays as it passes through intermediate nodes, particularly at congested links. Upon arrival at the receiving node, a delay may also be incurred as the proper receipt of the message is verified. Some of these factors are constant, while others vary over time, typically dependent on network loading.
  • the effectiveness of a network analysis system is based on a number of factors, one of which is the system's ability to distinguish the variety of causes of message delay, and another is the system's ability to assess the effect of potential network modifications on each of these classes of delay.
  • a variety of tools have been developed to distinguish the causes of message delay, including, for example, the techniques disclosed in copending U.S. patent application Ser. No. 11/776,736, “NETWORK CONGESTION DELAY ANALYSIS”, filed 12 Jul. 2007 for Steve Niemczyk, Patrick J. Malloy, Alain J. Cohen, and Russel Mark Elsner, attorney docket OP-060717, and incorporated by reference herein.
  • the various components of message delays are classified as bandwidth delay, propagation delay, protocol delay, congestion delay, and processing delay.
  • potential solutions to reduce these delays can be determined. For example, if a significant portion of the delay is attributed to congestion delay, the node that is causing the ‘bottleneck’ can be identified, and the routing of messages may be modified to provide a more balanced distribution of traffic, with a corresponding reduction in the amount of traffic through the bottleneck node.
  • additional channels between the nodes that are causing the bandwidth delay can be provided.
  • knowing the delay characteristics of a network provides opportunities for the developers of applications to optimize the applications by avoiding bottleneck paths, avoiding bursty traffic on bandwidth limited paths, and so on.
  • component delay is used herein to reference a delay in an application that can be eliminated by eliminating a single component, or type, of delay
  • parallel delay is used herein to refer to a delay in an application that can only be eliminated by two or more components of delay.
  • a trace file of an application's network events is processed to categorize the causes of delays incurred in the propagation and processing of these events.
  • the system identifies the amount of delay (component delay) that can be eliminated by eliminating each of the components of delay individually, as well as the amount of delay (parallel delay) that can be eliminated by eliminating combinations of the delay components.
  • a user interface displays the amount of reduction that can be achieved by eliminating each component delay individually and the amount of reduction that can be achieved by eliminating combinations of the individual component delays.
  • the interface allows the user to ‘drill down’ to view the individual delay components contained in each combination forming the parallel delays. In this manner, the user is provided a view of each of the delay components that would need to be addressed, either individually or in combination, to improve the overall application delay.
  • FIG. 1 illustrates an example display of a prior art delay analysis system.
  • FIGS. 2 and 4 illustrate example displays of a delay analysis system in accordance with this invention.
  • FIG. 3 illustrates an example flow diagram of a delay analysis system in accordance with this invention.
  • FIG. 5 illustrates an example block diagram of a timing analysis system in accordance with this invention.
  • FIG. 1 illustrates an example display of a prior art network timing analysis system, as presented in copending U.S. patent application Ser. No. 11/397,359, “APPLICATION DELAY ANALYSIS”, filed 4 Apr. 2006 for Russell Mark Elsner, John Wilson Strohm, and Michael S. Canney, attorney docket OP-050117, and incorporated by reference herein.
  • the timing analysis diagram of FIG. 1 presents summary statistics related to the source of delays in a segmented/exploded pie-chart format, each source herein termed a ‘delay component’.
  • delay component the delay that is related to the processing of the message (‘processing component’), such as the delay incurred between the time that a message is received and the time that a response to the message is produced, and component delays that are related to the transmission of the message, including each of the components of delay encountered as the message travels from its source to its destination, such as a bandwidth component, a latency component, a congestion component, and so on.
  • the network traversal delay generally includes a variety of components:
  • Latency/Propagation time the delay incurred for a bit of a message to travel from the source to the destination, generally determined by the distance between the source and destination, the speed of light/electromagnetic propagation, and the minimum forwarding time at each transfer point along the path.
  • Bandwidth/Transit time the delay incurred for all of the bits of the message to be transmitted and received, generally determined by the minimum bandwidth along the path of the message and the number of bits in the message.
  • Congestion/Protocol time the delay incurred to gain access to the channel to effect the transmission, generally determined by the difference between the time that the message was ready for transmission and the time that it was actually transmitted, including retransmissions.
  • the system may be configured to distinguish between protocol delay and congestion delay, and, optionally, may distinguish ‘self-congestion’ delay caused by other packets from the same application.
  • the protocol delay includes the overhead associated with conforming to a protocol, such as waiting for an acknowledgement of a prior packet before sending a next packet, and so on, whereas congestion delay includes the delays incurred when the message competes with other messages for network resources.
  • messages are generated, processed, and propagated among a client, an application server, and a database server, corresponding, for example, to an application wherein a user at a client device interacts with a server to request information, and the server obtains data from a database server to satisfy this request.
  • FIG. 1 illustrates the accumulated tier processing delay 110 , 120 , 130 for each of the example tiers/nodes (DB Server, Client, and App Server) in the example application, and the network traversal delay 140 , 150 for each of the active links (the App Server—DB Server link 140 , and the Client—App Server link 150 ).
  • the processing at the DB Server delay 110 accounts for about 58% of the total accumulated delay
  • the network traversal delay 140 between the App Server and the DB Server accounts for about 38% of the total delay.
  • each of the network traversal delays are preferably displayed, so that a network administrator and/or application developer can address the different causes of delay to effect an improvement.
  • the network traversal delay 140 between the App Server and the DB Server is illustrated as comprising a latency delay 141 , a bandwidth delay 142 , and a protocol/congestion delay 143 .
  • the network traversal delay 150 between the Client and App Server comprises primarily bandwidth delay, the other components being substantially negligible.
  • the display of FIG. 1 indicates that the performance of the application can be improved in a number of ways. It would appear, for example, that eliminating the processing delay 110 at the DB Server will substantially reduce the overall accumulated delay. In like manner, eliminating the bandwidth 142 or protocol/congestion delay 143 would appear to provide a substantial reduction in the overall accumulated delay. Such an interpretation, however, assumes that each of the delays is substantially independent and/or that the delays are strictly sequential. Such an assumption, however, is rarely valid in a typical network environment.
  • FIG. 1 corresponds to the delays experienced by an application that is structured to transfer data from the DB Server in parallel, at least in part, with the processing at the DB Server. That is, instead of waiting for the request for data to be entirely processed at the DB Server before the data is transmitted, the application is structured to provide data as it becomes available from the DB Server, such that some of the data is being transmitted at the same time that other data is being generated at the DB Server.
  • the system is configured to directly identify the reduction in overall delay that can be achieved by the elimination of delay components.
  • the system identifies the portions of the overall application delay that can be reduced by addressing each delay component individually and the portion of the overall application that can only be reduced by addressing multiple delay components. That is, the paradigm for presenting the information to the user is changed from the conventional identification of the contribution of each delay component to the overall delay to an identification of the reduction in the overall delay that can be achieved by eliminating each delay component, and each combination of delay components.
  • the system distinguishes between component delays that can be directly affected by a particular user's action, such as bandwidth and processing delay that can be reduced by providing higher-performance equipment, and component delays that cannot be directly affected by the particular user, such as latency and congestion delays, thereby further focusing the analysis on reductions in the application delay that can be achieved by the user.
  • FIG. 2 illustrates an example display of achievable delay reduction, using the same example application that was used in FIG. 1 .
  • FIG. 2 illustrates at segment 210 that the elimination of the processing delay at the DB Server will reduce the overall delay by approximately 26%.
  • Each of the other segments 220 , 230 , 240 ( 241 - 243 ), and 250 illustrates the amount of delay reduction that can be achieved by eliminating each of these delay components.
  • a new segment 290 identifies an amount of parallel delay, which indicates that approximately 36% of the overall delay cannot be reduced without addressing multiple delay components.
  • Any number of techniques can be used to determine/estimate the amount of delay that is reduced by eliminating each delay component.
  • a variety of network analysis programs allow users to perform “what-if” analyses.
  • a controller is configured to exercise such a network analysis program for each hypothesized elimination of delay.
  • FIG. 3 illustrates an example flow diagram for determining the amount of delay reduction that is achievable for each delay component and combination of delay components.
  • the selected analysis program is run to determine the application delay on the network as currently configured.
  • This application delay forms a measure from which the achievable reductions are determined.
  • the delay analysis system as described in the above referenced copending U.S. patent application Ser. No. 11/776,736, “NETWORK CONGESTION DELAY ANALYSIS” is used to process a trace file of events that occur on the network during the execution of the application to determine the delays incurred by the application in the network as currently configured.
  • the delay measure can be defined as the application delay that remains after eliminating the component delays that are uncontrollable/unaffectable by the particular user.
  • the term ‘application delay’ as used herein refers to a delay measure that is used as a reference for assessing an amount of delay that can be reduced from this measure by addressing one or more delay components.
  • the individual delay components that are to be assessed are identified.
  • the particular delay components will be determined based on the available delay analysis system.
  • the identified delay components include the processing delay at each tier of the network (e.g. client, server, and database processing delays), and the bandwidth, latency, protocol, and congestion delays across each link of the network (e.g. client-server and server-database links), although for ease of illustration a single protocol/congestion delay component is used. That is, the selection of display components need not correspond directly to the delay components of the analysis system, provided that the selected delay components can be controlled/eliminated using the components of the given analysis system.
  • a user may select from the individual delay components identified in the analysis program based on the components that are within the direct control of the user, and lump all other delay components as a single “other” component for subsequent assessment.
  • the typical delay-related elements that can be directly affected by a network administrator are the processing delays at select nodes, and the bandwidth delays at select links; each of the other delay components that may be identified in the analysis program may be identified as a single ‘beyond control’ component for assessment in this process.
  • the loop 320 - 340 exercises the given delay analysis program for each of the delay components under the what-if condition: “What if this delay component could be eliminated?”
  • the parameters of the delay analysis program are appropriately configured to effect the hypothesized elimination of the given delay component.
  • this elimination is effected by merely asserting that the value of the given delay component is zero.
  • a different set of parameters may be required to be set to particular values to effect an elimination of the given delay component. For example, in a flow analysis system, the bandwidth of a link may need to be set to a virtually infinite value to provide an elimination of the bandwidth delay of that link. Similarly, a distance between nodes may need to be set to zero to provide an elimination of latency delay, and so on.
  • the analysis program is executed using the hypothesized elimination of the given delay component to determine the application delay time that could be achieved if this delay component were eliminated from the application and/or network.
  • this determination of the resultant delay is not merely a subtraction of the conventionally identified amount of delay attributed to the particular delay component, due to parallel effects. If each event of the application occurs in a strictly sequential path, the resultant delay can be determined by subtracting the individual component delay, but few, if any, applications are purely sequential.
  • a conventional delay analysis program attributes the portions of delay caused by each delay component based on that component's presence on the critical-path of the delays causing the overall delay of the application.
  • the elimination of a delay component will often produce a new critical-path, and the resultant application delay will be the delay caused by the remaining delays along this new critical-path.
  • the new critical path is determined by eliminating the given component delay along each parallel path, and then applying networking-related constraints and/or assumptions to determine the resultant delay, as detailed in the above reference copending U.S. patent application Ser. No. 11/776,736, incorporated by reference herein. For example, if the elimination of a delay component along a path introduces a conflict for available network resources with another path, a congestion-delay component is assumed to occur in one of the paths to offset the concurrent demand for the resource.
  • the achievable reduction in the application delay time is determined by the difference between the application delay on the network as currently configured and this resultant delay on the network with a hypothesized elimination of the delay component, at 340 .
  • each delay component segment in FIG. 2 corresponds to the reduction determined for that component's elimination at 340 .
  • the segments illustrating the achievable reduction correspond to a percentage of the total application delay, although other forms of representation may be used, as detailed further below.
  • the amount of parallel delay i.e. the delay that can only be reduced by addressing multiple delay components, is determined as the delay that remains after accounting for the delays that can be reduced by addressing/eliminating the delay components individually, at 345 .
  • the sum of each achievable-reduction-segment 210 , 220 , 230 , 241 , 242 , 243 , 250 of individual delay components amounts to approximately 64%, leaving 36% of the total application delay that cannot be reduced by addressing any of the delay components individually, as illustrated by the parallel delay component 290 .
  • particular combinations of delay components contributing to the overall parallel delay component can be assessed to determine the reduction in application delay that can be achieved by addressing each combination. That is, a user may be interested in determining the reduction that could be achieved if the processing delay at both the client and server are addressed, or the reduction that could be achieved if both the bandwidth delay on the link between the server and database and the processing delay at the database server are addressed, and so on.
  • the particular combinations of delay components that are to be assessed are identified.
  • the number of possible combinations is dependent upon the number of individual delay components.
  • a default set of combinations are defined, and the user is provided a variety of options for making alternative selections of the combinations to be assessed.
  • the user may, for example, explicitly define each combination (e.g. client-server-bandwidth and server-processing; server-processing and client-processing; and so on).
  • the user may define a selection rule, such as ‘all two-component combinations’, ‘all two- or three-component combinations that include client-server-bandwidth’, ‘all combinations that do not include the “beyond control” delay component’, and so on).
  • a preferred user interface for this selection includes the ability to define the selection and/or the rules by selecting items on a graphic display, such as icons for each delay component and for common operations, such as ‘and’, ‘or’, ‘not’, ‘fewer than’, ‘more than’, and so on.
  • Common selection options such as ‘all n-component combinations’ can be defined wherein the user need only specify the desired “n” (e.g. “3” for selecting ‘all 3-component combinations’).
  • the loop 360 - 380 assesses each selected combination of component delays, using the same techniques as discussed above with respect to the loop 320 - 340 .
  • the appropriate parameters of the delay analysis system are set to effect an elimination of each delay component in the given combination, such as, in the above reference example, the elimination of both the client-server-bandwidth delay and the server-processing delay.
  • the analysis program determines the resultant delay under the hypothesized multiple delay component eliminations, and at 375 , the achieved reduction by this combined elimination is determined.
  • the user is continually provided the opportunity to define a particular combination and request a determination of the achievable reduction by that combination, either explicitly or implicitly. For example, if the user selects a displayed link, each of the two-component combinations that include this link may be assessed for subsequent display. Similarly, the user is provided the opportunity to refine a currently selected combination by adding or removing delay components from the combination, thereby facilitating progressive analyses.
  • the display of FIG. 2 an identification of the amount of application delay that can be reduced by the elimination of each of the individual delay components.
  • the display of FIG. 2 provides an indication of the delay components whose reduction is likely to provide a reduction in the overall delay.
  • an alternative interpretation of the parallel delay 290 is an indication of the efficiencies provided by the parallelism of the application/network. That is, the parallel delay 290 likely identifies proportions of time that are being efficiently utilized by including multiple concurrent delay components.
  • the parallel delay 290 likely identifies proportions of time that are being efficiently utilized by including multiple concurrent delay components.
  • the example application is structured to transmit data from the DB Server to the App Server while the DB Server continues to process the request for the data, some of the transmission delay that would have occurred at the end of processing the data at the DB Server will occur concurrent with the processing of the data at the DB Server. That is, if the delays were not occurring on parallel paths, the overall delay would likely be substantially larger, and thus parallel delays are generally indicative of a reduction of the delay that might have occurred if these parallel delays were not present.
  • the display of the components of an application's delay is interactive, allowing the user to obtain alternative views to facilitate the analysis of the achievable reduction in application delays.
  • the example protocol/congestion delay category in FIGS. 1 and 2 may be partitioned into a protocol delay category and a congestion delay category.
  • the protocol delay category can be further partitioned into sub-categories of delay such as delays caused by Retransmissions, Out of Sequence Packets, TCP-specific effects (In-Flight Data, Nagle's Delay, etc.), and so on.
  • the congestion delay category can be similarly further partitioned.
  • each of the characteristic delays associated with the “ftp downloads” during an application can be categorized as individual delay components, to facilitate an analysis of the effect of reducing such ftp-related delay components.
  • the user is provided the option of ‘drilling down’ into each displayed delay component to view statistics related to the variety of sub-categories within each category. These statistics may be presented in a variety of forms, including, for example, plots, bar-graphs, tables, and so on.
  • the user preferably drills-down to lower levels of detail by ‘double-clicking’ on the display of the delay component of interest. For example, double-clicking on the parallel delay segment 290 of FIG. 2 may provide a display of the reduction provided by each 2-component combination and the remainder reduction provided by the ‘greater-than-2’ component combinations.
  • Double-clicking on the ‘greater-than-2’ segment may provide a display of the reduction provided by each 3-component combination and the remainder reduction provided by the ‘greater-than-3’ component combinations.
  • double-clicking on the parallel delay segment 290 may provide a partitioning of the segment 290 into the reduction achievable by each set of combinations (all 2-component combinations, all 3-component combinations, all 4-component combinations, and so on).
  • FIG. 4 illustrates an example display of achievable delay reductions in a bar-chart format.
  • the user is provided the option of partitioning/grouping the delays to present different views of the delays to facilitate analysis and/or trouble-shooting.
  • the delays associated with a client-server-database application are illustrated in three bar graphs 410 , 411 , and 412 relative to the processing delay at each tier and the bandwidth delay of the client-server link.
  • Such a display may be provided, for example, in response to the user selecting the client-server bandwidth delay as a first parameter of the display presentation, and processing delays as a second parameter.
  • the first bar 410 illustrates that 1.0 seconds 411 of application delay can be reduced by eliminating the server processing delay; 1.25 seconds 412 can be reduced by eliminating the server-client bandwidth delay; and an additional 0.75 seconds 413 can be reduced by eliminating both the server processing delay and the server-client bandwidth delay.
  • the second bar 420 illustrates the same 1.25 second reduction 412 by eliminating the server-client bandwidth delay, a 0.5 second reduction 421 by eliminating the database processing delay, and an additional 0.1 second reduction 423 by eliminating both the database processing delay and the server-client bandwidth delay.
  • the third bar 430 illustrates the same 1.25 second reduction 412 by eliminating the server-client bandwidth delay, a 0.2 second reduction 431 by eliminating the client processing delay, and a minimal additional reduction 433 by eliminating both the client processing delay and the server-client bandwidth delay.
  • FIG. 5 illustrates an example block diagram of a timing analysis system for embodying this invention.
  • One of skill in the art will recognize that different processes and structures may be used to provide the desired determination and presentation of achievable application delay reductions based on the elimination of each delay component individually and the elimination of sets of multiple delay components.
  • An application extractor 520 extracts messages and events related to the application being analyzed from trace data 510 to create a sequence of application messages/events.
  • This trace data 510 is generally stored data, although it could be ‘live’ data from trace devices, such as one or more network ‘sniffers’ that monitor traffic on a network.
  • this timing analysis system may be coupled to a network simulator that models network behavior, and the trace data 510 may reflect simulated messages and events.
  • a controller 540 enables a delay analyzer 530 to perform a delay analysis of the application-specific events identified by the application extractor 520 to determine an application delay under current application/network conditions.
  • the determined application delay may only include select delay components that are controllable by the particular user.
  • the controller 540 subsequently receives a set of select delay components 540 for assessment.
  • These delay components 540 may be the same components as used in the delay analyzer 530 , or components that can be derived from or can control the delay components of the delay analyzer 530 , as discussed above.
  • the controller 540 controls the delay analyzer 530 to determine the effect of an elimination of each of the individual delay components 550 , preferably in the form of an identification of the reduction from the application delay that is provided by the sole elimination of each individual delay component. From the application delay and these determined reductions, the remainder reduction that can only be achieved by the elimination of multiple delay components is determined and identified as a composite parallel delay.
  • the determined reductions 560 including the reductions achievable by addressing multiple component delays (parallel delays) are preferably presented to the user via an interactive display 570 , and accessed via a user interface 580 .
  • the system embodying this invention also allows a user to define/select combinations of delay components 550 , as discussed above, based on which the controller 540 adjusts the parameters of delay analyzer 530 to effect the elimination of all of the delay components within each selected combination.
  • the delay analyzer 530 consequently determines that reduction in application delay 560 that can be achieved by eliminating all of the delay components of each combination.
  • the user interface 580 allows the user to view the determined delay reductions for each individual delay component and any or all of the selected combinations of these delay components, as well as allowing the user to identify other combinations 550 for assessment by the delay analyzer 530 .
  • the interactive display 570 is configured to interact with the user interface 580 to provide a display of the reductions of application delay that can be achieved by addressing each of the individual delay components, as well as the sets of selected combinations of the various delay components, as discussed above.

Abstract

The embodiments facilitate the analysis of application delays, including delays that occur on multiple paths. A trace file of an application's network events is processed to categorize the causes of delays incurred in the propagation and processing of these events. The system identifies the amount of delay that can be eliminated by eliminating each of the components of delay individually, as well as the amount of delay that can be eliminated by eliminating combinations of the delay components. A user interface displays the amount of reduction that can be achieved by eliminating various delays alone or in combination. The interface also allows the user to view the individual delay components contained in combinations of delay components. In this manner, the user is provided a view of each of the delay components that would need to be addressed, either individually or in combination, to improve the overall application delay.

Description

CROSS-REFERENCE TO RELATED APPLICATIONS
This application is a continuation of U.S. patent application Ser. No. 12/060,271, now U.S. Pat. No. 8,095,649, filed Apr. 1, 2008, entitled “Network Delay Analysis including Parallel Delay Effects,” which claims the benefit of priority of U.S. Provisional Application No. 60/916,842, filed May 9, 2007, which are incorporated by reference in their entirety.
BACKGROUND
Unfortunately, after manufacturing or assembly, many devices will remain idle or inactive while awaiting delivery or in transit to their destination. This leads to unnecessary incurring of costs to purchase and install an operating system. In addition, the idle and/or transit time leads to depreciation of the value of the device and its operating system.
This invention relates to the field of network analysis, and in particular to a system and method for analyzing and assessing the effects of parallel delays within an application.
Proper management of a network generally requires assuring that the network is performing satisfactorily for the users of the network, modifying the network to address performance issues or problems, and planning for future improvements to the network as demand increases and as newer technologies and alternatives become available.
A variety of tools have been developed, and continue to be developed, to facilitate the management of communication networks, and in particular for managing networks that provide communications among computer devices. Many of these tools are configured to model the network's performance under a variety of traffic conditions, both real and hypothesized, and in many cases, base this performance on data collected from the actual network.
One of the primary parameters for analyzing or assessing the performance of a network is the time it takes for messages to reliably reach their destination. This time is dependent upon a variety of factors. The message is typically partitioned into transmission elements, herein termed packets for convenience. Each packet must enter the network, and incurs a delay as it gains network access. When it enters the network, it incurs a delay that is dependent upon the bandwidth available at each link along its path to its destination. It may also incur queuing delays as it passes through intermediate nodes, particularly at congested links. Upon arrival at the receiving node, a delay may also be incurred as the proper receipt of the message is verified. Some of these factors are constant, while others vary over time, typically dependent on network loading.
The effectiveness of a network analysis system is based on a number of factors, one of which is the system's ability to distinguish the variety of causes of message delay, and another is the system's ability to assess the effect of potential network modifications on each of these classes of delay. A variety of tools have been developed to distinguish the causes of message delay, including, for example, the techniques disclosed in copending U.S. patent application Ser. No. 11/776,736, “NETWORK CONGESTION DELAY ANALYSIS”, filed 12 Jul. 2007 for Steve Niemczyk, Patrick J. Malloy, Alain J. Cohen, and Russel Mark Elsner, attorney docket OP-060717, and incorporated by reference herein. In this copending application, the various components of message delays are classified as bandwidth delay, propagation delay, protocol delay, congestion delay, and processing delay. By knowing the cause of the delays that a message incurs, potential solutions to reduce these delays can be determined. For example, if a significant portion of the delay is attributed to congestion delay, the node that is causing the ‘bottleneck’ can be identified, and the routing of messages may be modified to provide a more balanced distribution of traffic, with a corresponding reduction in the amount of traffic through the bottleneck node. In like manner, if a significant portion of the delay is attributed to bandwidth delay, additional channels between the nodes that are causing the bandwidth delay can be provided. In like manner, knowing the delay characteristics of a network provides opportunities for the developers of applications to optimize the applications by avoiding bottleneck paths, avoiding bursty traffic on bandwidth limited paths, and so on.
The use of conventional delay analysis and assessment techniques to identify potential improvements of performance, however, has significant limitations. Generally, the performance factors are not independent, such that an improvement in one delay factor is not necessarily reflected in the resultant delay. Conventional delay analysis techniques generally allocate/classify delays to the components in the ‘critical path’ of the message delay; that is, each component delay is determined by its direct effect on the overall message delay. Often, a reduction in one delay component merely reveals that another delay component that was not on the critical path is identified as being (another) major cause of the overall message delay. For example, if a particular communications link exhibits a significant bandwidth delay, the fact that there are delays caused by slow processing may be masked, particularly if a slow processor is providing data only slightly faster than the bandwidth-limited link can forward the data. Curing the bottleneck will not necessarily have a corresponding effect on the overall delay, because the data continues to be presented slowly, albeit into a wider bandwidth channel.
This lack of independence among delay components is particularly problematic for applications that employ parallelism. Consider, for example, an application that includes two tasks, one task that incurs a substantial processing delay, and another that incurs a substantial bandwidth delay. If these tasks are performed sequentially, the overall delay will correspond to the sum of these delays, and a reduction in either will be reflected in the overall delay. If these tasks are performed in parallel, however, the overall delay will correspond to the longer of the two delays, and a reduction in one of the delays will not necessarily affect the overall delay. Conventional delay analysis techniques that report the delays that are only on the critical path, and thus have a direct effect on the overall delay, provide little guidance as to the effect that a reduction of any delay component will have on the overall delay of an application that employs parallelism.
It should be noted that most network applications are affected by multiple delays on parallel paths, even if the application is not purposely designed to use parallelism. An event at one node may trigger, for example, parallel events on another node. Even though the application on the first node may be purely sequential, its response from the second node will be dependent upon the delays occurring on the parallel paths.
For ease of reference, the term ‘component delay’ is used herein to reference a delay in an application that can be eliminated by eliminating a single component, or type, of delay, and ‘parallel delay’ is used herein to refer to a delay in an application that can only be eliminated by two or more components of delay.
It would be advantageous to provide a method and system that identify parallel delays. It would also be advantageous to provide a method and system that facilitate the analysis of parallel delays. It would also be advantageous to provide a method and system that facilitate the identification of improvements that can be achieved by reducing one or more delay components within a network or within an application.
These advantages, and others, can be realized by a method and system that facilitate the analysis and assessment of application delays, including parallel delays. A trace file of an application's network events is processed to categorize the causes of delays incurred in the propagation and processing of these events. The system identifies the amount of delay (component delay) that can be eliminated by eliminating each of the components of delay individually, as well as the amount of delay (parallel delay) that can be eliminated by eliminating combinations of the delay components. A user interface displays the amount of reduction that can be achieved by eliminating each component delay individually and the amount of reduction that can be achieved by eliminating combinations of the individual component delays. To facilitate the analysis and assessment of these parallel delays, the interface allows the user to ‘drill down’ to view the individual delay components contained in each combination forming the parallel delays. In this manner, the user is provided a view of each of the delay components that would need to be addressed, either individually or in combination, to improve the overall application delay.
BRIEF DESCRIPTION OF THE DRAWINGS
The invention is explained in further detail, and by way of example, with reference to the accompanying drawings wherein:
FIG. 1 illustrates an example display of a prior art delay analysis system.
FIGS. 2 and 4 illustrate example displays of a delay analysis system in accordance with this invention.
FIG. 3 illustrates an example flow diagram of a delay analysis system in accordance with this invention.
FIG. 5 illustrates an example block diagram of a timing analysis system in accordance with this invention.
Throughout the drawings, the same reference numerals indicate similar or corresponding features or functions. The drawings are included for illustrative purposes and are not intended to limit the scope of the invention.
DETAILED DESCRIPTION
In the following description, for purposes of explanation rather than limitation, specific details are set forth such as the particular architecture, interfaces, techniques, etc., in order to provide a thorough understanding of the concepts of the invention. However, it will be apparent to those skilled in the art that the present invention may be practiced in other embodiments, which depart from these specific details. In like manner, the text of this description is directed to the example embodiments as illustrated in the Figures, and is not intended to limit the claimed invention beyond the limits expressly included in the claims. For purposes of simplicity and clarity, detailed descriptions of well-known devices, circuits, and methods are omitted so as not to obscure the description of the present invention with unnecessary detail.
FIG. 1 illustrates an example display of a prior art network timing analysis system, as presented in copending U.S. patent application Ser. No. 11/397,359, “APPLICATION DELAY ANALYSIS”, filed 4 Apr. 2006 for Russell Mark Elsner, John Wilson Strohm, and Michael S. Canney, attorney docket OP-050117, and incorporated by reference herein.
The timing analysis diagram of FIG. 1 presents summary statistics related to the source of delays in a segmented/exploded pie-chart format, each source herein termed a ‘delay component’. Typically, a distinction is made between the delay that is related to the processing of the message (‘processing component’), such as the delay incurred between the time that a message is received and the time that a response to the message is produced, and component delays that are related to the transmission of the message, including each of the components of delay encountered as the message travels from its source to its destination, such as a bandwidth component, a latency component, a congestion component, and so on.
The network traversal delay generally includes a variety of components:
Latency/Propagation time: the delay incurred for a bit of a message to travel from the source to the destination, generally determined by the distance between the source and destination, the speed of light/electromagnetic propagation, and the minimum forwarding time at each transfer point along the path.
Bandwidth/Transit time: the delay incurred for all of the bits of the message to be transmitted and received, generally determined by the minimum bandwidth along the path of the message and the number of bits in the message.
Congestion/Protocol time: the delay incurred to gain access to the channel to effect the transmission, generally determined by the difference between the time that the message was ready for transmission and the time that it was actually transmitted, including retransmissions.
Optionally, the system may be configured to distinguish between protocol delay and congestion delay, and, optionally, may distinguish ‘self-congestion’ delay caused by other packets from the same application. Generally, the protocol delay includes the overhead associated with conforming to a protocol, such as waiting for an acknowledgement of a prior packet before sending a next packet, and so on, whereas congestion delay includes the delays incurred when the message competes with other messages for network resources.
In the example of FIG. 1, messages are generated, processed, and propagated among a client, an application server, and a database server, corresponding, for example, to an application wherein a user at a client device interacts with a server to request information, and the server obtains data from a database server to satisfy this request.
FIG. 1 illustrates the accumulated tier processing delay 110, 120, 130 for each of the example tiers/nodes (DB Server, Client, and App Server) in the example application, and the network traversal delay 140, 150 for each of the active links (the App Server—DB Server link 140, and the Client—App Server link 150). In this example, the processing at the DB Server delay 110 accounts for about 58% of the total accumulated delay, while the network traversal delay 140 between the App Server and the DB Server accounts for about 38% of the total delay.
In a preferred embodiment of a network analysis system, the components of each of the network traversal delays are preferably displayed, so that a network administrator and/or application developer can address the different causes of delay to effect an improvement. The network traversal delay 140 between the App Server and the DB Server is illustrated as comprising a latency delay 141, a bandwidth delay 142, and a protocol/congestion delay 143. The network traversal delay 150 between the Client and App Server comprises primarily bandwidth delay, the other components being substantially negligible.
The display of FIG. 1 indicates that the performance of the application can be improved in a number of ways. It would appear, for example, that eliminating the processing delay 110 at the DB Server will substantially reduce the overall accumulated delay. In like manner, eliminating the bandwidth 142 or protocol/congestion delay 143 would appear to provide a substantial reduction in the overall accumulated delay. Such an interpretation, however, assumes that each of the delays is substantially independent and/or that the delays are strictly sequential. Such an assumption, however, is rarely valid in a typical network environment.
Consider, for example, an environment wherein FIG. 1 corresponds to the delays experienced by an application that is structured to transfer data from the DB Server in parallel, at least in part, with the processing at the DB Server. That is, instead of waiting for the request for data to be entirely processed at the DB Server before the data is transmitted, the application is structured to provide data as it becomes available from the DB Server, such that some of the data is being transmitted at the same time that other data is being generated at the DB Server.
Consider now a possible effect of improving the processing time at the DB Server; such as increasing the rate at which the DB Server is able to provide the data to the network. If the network is currently underutilized, such that there is excess bandwidth available, increasing the rate at which the data is provided to the network will have a direct effect on the overall delay time of the example application. If, however, the link between the DB Server and App Server is currently at or near its maximum bandwidth capability, increasing the rate that the DB Server is able to provide the data will not necessarily produce a corresponding reduction in the overall delay, because the data will merely have to wait in queues before the bandwidth-constrained link is able to accept the data. That is, most or all of the reduction in processing delay at the DB Server will merely be offset by a corresponding increase in congestion delay across the DB Server—App Server link. That is, in such an environment, although the processing delay component 110 of FIG. 1 would decrease, the protocol/congestion delay component 143 would increase, and the effect on the overall delay would be significantly less than the effect implied by the conventional display of FIG. 1. In like manner, reducing the bandwidth delay 142 may have little effect on the overall delay if a significant amount of the bandwidth delay is being incurred in parallel with the processing delay 110.
This simple example serves to demonstrate that although the processing delay at the DB Server is illustrated as contributing 58% of the overall delay using the conventional delay analysis display of FIG. 1, the elimination of the processing delay at the DP Server will not necessarily reduce the overall delay by 58%, because the amount of reduction in overall delay that is achievable is dependent upon other factors, such link utilization and others. In general, the conventional display of FIG. 1 provides a minimal amount of information regarding the amount of actual delay reduction that can be achieved by eliminating any of the illustrated delay components.
In accordance with an aspect of this invention, the system is configured to directly identify the reduction in overall delay that can be achieved by the elimination of delay components. In a preferred embodiment of this invention, the system identifies the portions of the overall application delay that can be reduced by addressing each delay component individually and the portion of the overall application that can only be reduced by addressing multiple delay components. That is, the paradigm for presenting the information to the user is changed from the conventional identification of the contribution of each delay component to the overall delay to an identification of the reduction in the overall delay that can be achieved by eliminating each delay component, and each combination of delay components. Optionally, the system distinguishes between component delays that can be directly affected by a particular user's action, such as bandwidth and processing delay that can be reduced by providing higher-performance equipment, and component delays that cannot be directly affected by the particular user, such as latency and congestion delays, thereby further focusing the analysis on reductions in the application delay that can be achieved by the user.
FIG. 2 illustrates an example display of achievable delay reduction, using the same example application that was used in FIG. 1. As contrast to the 58% segment 110 of overall delay being attributed to processing delay at the DB Server illustrated in FIG. 1, with no indication of the savings achievable by addressing this processing delay, FIG. 2 illustrates at segment 210 that the elimination of the processing delay at the DB Server will reduce the overall delay by approximately 26%. Each of the other segments 220, 230, 240 (241-243), and 250 illustrates the amount of delay reduction that can be achieved by eliminating each of these delay components. A new segment 290 identifies an amount of parallel delay, which indicates that approximately 36% of the overall delay cannot be reduced without addressing multiple delay components.
Any number of techniques can be used to determine/estimate the amount of delay that is reduced by eliminating each delay component. A variety of network analysis programs, for example, allow users to perform “what-if” analyses. In a preferred embodiment of this invention, a controller is configured to exercise such a network analysis program for each hypothesized elimination of delay.
FIG. 3 illustrates an example flow diagram for determining the amount of delay reduction that is achievable for each delay component and combination of delay components.
At 310, the selected analysis program is run to determine the application delay on the network as currently configured. This application delay forms a measure from which the achievable reductions are determined. In an example embodiment of this invention, the delay analysis system as described in the above referenced copending U.S. patent application Ser. No. 11/776,736, “NETWORK CONGESTION DELAY ANALYSIS” is used to process a trace file of events that occur on the network during the execution of the application to determine the delays incurred by the application in the network as currently configured. Optionally, the delay measure can be defined as the application delay that remains after eliminating the component delays that are uncontrollable/unaffectable by the particular user. For ease of reference, the term ‘application delay’ as used herein refers to a delay measure that is used as a reference for assessing an amount of delay that can be reduced from this measure by addressing one or more delay components.
At 315, the individual delay components that are to be assessed are identified. Generally, the particular delay components will be determined based on the available delay analysis system. In the example delay analysis system, the identified delay components include the processing delay at each tier of the network (e.g. client, server, and database processing delays), and the bandwidth, latency, protocol, and congestion delays across each link of the network (e.g. client-server and server-database links), although for ease of illustration a single protocol/congestion delay component is used. That is, the selection of display components need not correspond directly to the delay components of the analysis system, provided that the selected delay components can be controlled/eliminated using the components of the given analysis system. For example, a user may select from the individual delay components identified in the analysis program based on the components that are within the direct control of the user, and lump all other delay components as a single “other” component for subsequent assessment. For example, the typical delay-related elements that can be directly affected by a network administrator are the processing delays at select nodes, and the bandwidth delays at select links; each of the other delay components that may be identified in the analysis program may be identified as a single ‘beyond control’ component for assessment in this process.
The loop 320-340 exercises the given delay analysis program for each of the delay components under the what-if condition: “What if this delay component could be eliminated?”
At 325, the parameters of the delay analysis program are appropriately configured to effect the hypothesized elimination of the given delay component. In the example delay analysis program, this elimination is effected by merely asserting that the value of the given delay component is zero. In other conventional delay analysis systems, a different set of parameters may be required to be set to particular values to effect an elimination of the given delay component. For example, in a flow analysis system, the bandwidth of a link may need to be set to a virtually infinite value to provide an elimination of the bandwidth delay of that link. Similarly, a distance between nodes may need to be set to zero to provide an elimination of latency delay, and so on.
At 330, the analysis program is executed using the hypothesized elimination of the given delay component to determine the application delay time that could be achieved if this delay component were eliminated from the application and/or network.
It is significant to note that, as discussed above, this determination of the resultant delay is not merely a subtraction of the conventionally identified amount of delay attributed to the particular delay component, due to parallel effects. If each event of the application occurs in a strictly sequential path, the resultant delay can be determined by subtracting the individual component delay, but few, if any, applications are purely sequential.
Conceptually, a conventional delay analysis program attributes the portions of delay caused by each delay component based on that component's presence on the critical-path of the delays causing the overall delay of the application. In an application with parallel treads, the elimination of a delay component will often produce a new critical-path, and the resultant application delay will be the delay caused by the remaining delays along this new critical-path.
In the example delay analysis program, the new critical path is determined by eliminating the given component delay along each parallel path, and then applying networking-related constraints and/or assumptions to determine the resultant delay, as detailed in the above reference copending U.S. patent application Ser. No. 11/776,736, incorporated by reference herein. For example, if the elimination of a delay component along a path introduces a conflict for available network resources with another path, a congestion-delay component is assumed to occur in one of the paths to offset the concurrent demand for the resource.
Having determined the resultant delay under the hypothesized elimination of the given delay component, the achievable reduction in the application delay time is determined by the difference between the application delay on the network as currently configured and this resultant delay on the network with a hypothesized elimination of the delay component, at 340.
This achievable reduction for each delay component is used as the basis for the display of delay components in FIG. 2. That is, each delay component segment in FIG. 2 corresponds to the reduction determined for that component's elimination at 340. In FIG. 2, the segments illustrating the achievable reduction correspond to a percentage of the total application delay, although other forms of representation may be used, as detailed further below.
The amount of parallel delay, i.e. the delay that can only be reduced by addressing multiple delay components, is determined as the delay that remains after accounting for the delays that can be reduced by addressing/eliminating the delay components individually, at 345. In FIG. 2, for example, the sum of each achievable-reduction- segment 210, 220, 230, 241, 242, 243, 250 of individual delay components amounts to approximately 64%, leaving 36% of the total application delay that cannot be reduced by addressing any of the delay components individually, as illustrated by the parallel delay component 290.
In a preferred embodiment of this invention, particular combinations of delay components contributing to the overall parallel delay component can be assessed to determine the reduction in application delay that can be achieved by addressing each combination. That is, a user may be interested in determining the reduction that could be achieved if the processing delay at both the client and server are addressed, or the reduction that could be achieved if both the bandwidth delay on the link between the server and database and the processing delay at the database server are addressed, and so on.
At 350 of FIG. 3, the particular combinations of delay components that are to be assessed are identified. The number of possible combinations is dependent upon the number of individual delay components. In the current example, there are three processing delay components (client, server, database processing) and six link-related delay components (latency, bandwidth, and protocol/congestion at each of the two links (client-server, server-database)), which would amount to over a thousand possible combinations of two or more of these ten individual delay components.
Preferably, a default set of combinations are defined, and the user is provided a variety of options for making alternative selections of the combinations to be assessed. The user may, for example, explicitly define each combination (e.g. client-server-bandwidth and server-processing; server-processing and client-processing; and so on). Or, the user may define a selection rule, such as ‘all two-component combinations’, ‘all two- or three-component combinations that include client-server-bandwidth’, ‘all combinations that do not include the “beyond control” delay component’, and so on). A preferred user interface for this selection includes the ability to define the selection and/or the rules by selecting items on a graphic display, such as icons for each delay component and for common operations, such as ‘and’, ‘or’, ‘not’, ‘fewer than’, ‘more than’, and so on. Common selection options, such as ‘all n-component combinations’ can be defined wherein the user need only specify the desired “n” (e.g. “3” for selecting ‘all 3-component combinations’).
The loop 360-380 assesses each selected combination of component delays, using the same techniques as discussed above with respect to the loop 320-340. At 365, the appropriate parameters of the delay analysis system are set to effect an elimination of each delay component in the given combination, such as, in the above reference example, the elimination of both the client-server-bandwidth delay and the server-processing delay. At 370, the analysis program determines the resultant delay under the hypothesized multiple delay component eliminations, and at 375, the achieved reduction by this combined elimination is determined.
One of skill in the art will recognize that the selection of combinations and subsequent determination of the reduction achievable by eliminating all of the delay components in each combination need not occur as a loop that processes each of a set of defined combinations. In a preferred embodiment, the user is continually provided the opportunity to define a particular combination and request a determination of the achievable reduction by that combination, either explicitly or implicitly. For example, if the user selects a displayed link, each of the two-component combinations that include this link may be assessed for subsequent display. Similarly, the user is provided the opportunity to refine a currently selected combination by adding or removing delay components from the combination, thereby facilitating progressive analyses.
As noted above, the display of FIG. 2 an identification of the amount of application delay that can be reduced by the elimination of each of the individual delay components. Alternatively viewed, the display of FIG. 2 provides an indication of the delay components whose reduction is likely to provide a reduction in the overall delay.
In like manner, an alternative interpretation of the parallel delay 290 is an indication of the efficiencies provided by the parallelism of the application/network. That is, the parallel delay 290 likely identifies proportions of time that are being efficiently utilized by including multiple concurrent delay components. As noted above, if the example application is structured to transmit data from the DB Server to the App Server while the DB Server continues to process the request for the data, some of the transmission delay that would have occurred at the end of processing the data at the DB Server will occur concurrent with the processing of the data at the DB Server. That is, if the delays were not occurring on parallel paths, the overall delay would likely be substantially larger, and thus parallel delays are generally indicative of a reduction of the delay that might have occurred if these parallel delays were not present. That is, one could alternatively use the provided analysis to optimize efficiency by maximizing the proportion of parallel delay. In like manner, the occurrence of a parallel delay of a combination of a large number of individual components can often be viewed as a highly efficient occurrence of multiple concurrent delays.
In a preferred embodiment of this invention, the display of the components of an application's delay is interactive, allowing the user to obtain alternative views to facilitate the analysis of the achievable reduction in application delays.
As noted above, the various components of delay can be identified to different levels of detail. The example protocol/congestion delay category in FIGS. 1 and 2, for example, may be partitioned into a protocol delay category and a congestion delay category. The protocol delay category can be further partitioned into sub-categories of delay such as delays caused by Retransmissions, Out of Sequence Packets, TCP-specific effects (In-Flight Data, Nagle's Delay, etc.), and so on. In like manner, the congestion delay category can be similarly further partitioned.
In addition to partitioning delays into components based on the type of delay that is occurring, certain events or processes exhibit characteristic patterns of delays, and the delays attributable to such events or processes can be distinguished as individual delay components. For example, each of the characteristic delays associated with the “ftp downloads” during an application can be categorized as individual delay components, to facilitate an analysis of the effect of reducing such ftp-related delay components.
In accordance with an aspect of this invention, the user is provided the option of ‘drilling down’ into each displayed delay component to view statistics related to the variety of sub-categories within each category. These statistics may be presented in a variety of forms, including, for example, plots, bar-graphs, tables, and so on. As is common in the art, the user preferably drills-down to lower levels of detail by ‘double-clicking’ on the display of the delay component of interest. For example, double-clicking on the parallel delay segment 290 of FIG. 2 may provide a display of the reduction provided by each 2-component combination and the remainder reduction provided by the ‘greater-than-2’ component combinations. Double-clicking on the ‘greater-than-2’ segment may provide a display of the reduction provided by each 3-component combination and the remainder reduction provided by the ‘greater-than-3’ component combinations. Or, for example, double-clicking on the parallel delay segment 290 may provide a partitioning of the segment 290 into the reduction achievable by each set of combinations (all 2-component combinations, all 3-component combinations, all 4-component combinations, and so on).
FIG. 4 illustrates an example display of achievable delay reductions in a bar-chart format. In a preferred embodiment of this invention, the user is provided the option of partitioning/grouping the delays to present different views of the delays to facilitate analysis and/or trouble-shooting.
In this example, the delays associated with a client-server-database application are illustrated in three bar graphs 410, 411, and 412 relative to the processing delay at each tier and the bandwidth delay of the client-server link. Such a display may be provided, for example, in response to the user selecting the client-server bandwidth delay as a first parameter of the display presentation, and processing delays as a second parameter.
The first bar 410 illustrates that 1.0 seconds 411 of application delay can be reduced by eliminating the server processing delay; 1.25 seconds 412 can be reduced by eliminating the server-client bandwidth delay; and an additional 0.75 seconds 413 can be reduced by eliminating both the server processing delay and the server-client bandwidth delay.
The second bar 420 illustrates the same 1.25 second reduction 412 by eliminating the server-client bandwidth delay, a 0.5 second reduction 421 by eliminating the database processing delay, and an additional 0.1 second reduction 423 by eliminating both the database processing delay and the server-client bandwidth delay.
The third bar 430 illustrates the same 1.25 second reduction 412 by eliminating the server-client bandwidth delay, a 0.2 second reduction 431 by eliminating the client processing delay, and a minimal additional reduction 433 by eliminating both the client processing delay and the server-client bandwidth delay.
FIG. 5 illustrates an example block diagram of a timing analysis system for embodying this invention. One of skill in the art will recognize that different processes and structures may be used to provide the desired determination and presentation of achievable application delay reductions based on the elimination of each delay component individually and the elimination of sets of multiple delay components.
An application extractor 520 extracts messages and events related to the application being analyzed from trace data 510 to create a sequence of application messages/events. This trace data 510 is generally stored data, although it could be ‘live’ data from trace devices, such as one or more network ‘sniffers’ that monitor traffic on a network. Optionally, this timing analysis system may be coupled to a network simulator that models network behavior, and the trace data 510 may reflect simulated messages and events.
A controller 540 enables a delay analyzer 530 to perform a delay analysis of the application-specific events identified by the application extractor 520 to determine an application delay under current application/network conditions. As noted above, the determined application delay may only include select delay components that are controllable by the particular user.
The controller 540 subsequently receives a set of select delay components 540 for assessment. These delay components 540 may be the same components as used in the delay analyzer 530, or components that can be derived from or can control the delay components of the delay analyzer 530, as discussed above.
The controller 540 controls the delay analyzer 530 to determine the effect of an elimination of each of the individual delay components 550, preferably in the form of an identification of the reduction from the application delay that is provided by the sole elimination of each individual delay component. From the application delay and these determined reductions, the remainder reduction that can only be achieved by the elimination of multiple delay components is determined and identified as a composite parallel delay.
The determined reductions 560, including the reductions achievable by addressing multiple component delays (parallel delays) are preferably presented to the user via an interactive display 570, and accessed via a user interface 580.
Optionally, the system embodying this invention also allows a user to define/select combinations of delay components 550, as discussed above, based on which the controller 540 adjusts the parameters of delay analyzer 530 to effect the elimination of all of the delay components within each selected combination. The delay analyzer 530 consequently determines that reduction in application delay 560 that can be achieved by eliminating all of the delay components of each combination. The user interface 580 allows the user to view the determined delay reductions for each individual delay component and any or all of the selected combinations of these delay components, as well as allowing the user to identify other combinations 550 for assessment by the delay analyzer 530.
The interactive display 570 is configured to interact with the user interface 580 to provide a display of the reductions of application delay that can be achieved by addressing each of the individual delay components, as well as the sets of selected combinations of the various delay components, as discussed above.
The foregoing merely illustrates the principles of the invention. It will thus be appreciated that those skilled in the art will be able to devise various arrangements which, although not explicitly described or shown herein, embody the principles of the invention and are thus within its spirit and scope. For example, although the timing analysis system is presented for assessing the delays of a single application, one of skill in the art will recognize that the ‘application’ being analyzed may include multiple applications, to assess the parallel delays while multiple applications are run, or select parts of a single application, to assess the parallel delays that occur at these select parts. These and other system configuration and optimization features will be evident to one of ordinary skill in the art in view of this disclosure, and are included within the scope of the following claims.
In interpreting these claims, it should be understood that: a) the word “comprising” does not exclude the presence of other elements or acts than those listed in a given claim; b) the word “a” or “an” preceding an element does not exclude the presence of a plurality of such elements; c) any reference signs in the claims do not limit their scope; d) several “means” may be represented by the same item or hardware or software implemented structure or function; e) each of the disclosed elements may be comprised of hardware portions (e.g., including discrete and integrated electronic circuitry), software portions (e.g., computer programming), and any combination thereof; f) hardware portions may be comprised of one or both of analog and digital portions; g) any of the disclosed devices or portions thereof may be combined together or separated into further portions unless specifically stated otherwise; h) no specific sequence of acts is intended to be required unless specifically indicated; and i) the term “plurality of” an element includes two or more of the claimed element, and does not imply any particular range of number of elements; that is, a plurality of elements can be as few as two elements, and can include an immeasurable number of elements.
The features and attributes of the specific embodiments disclosed above may be combined in different ways to form additional embodiments, all of which fall within the scope of the present disclosure. Although the present disclosure provides certain embodiments and applications, other embodiments that are apparent to those of ordinary skill in the art, including embodiments, which do not provide all of the features and advantages set forth herein, are also within the scope of this disclosure. Accordingly, the scope of the present disclosure is intended to be defined only by reference to the appended claims.

Claims (18)

What is claimed is:
1. A method of providing network analysis information regarding execution of an application over a network based on varying individual components and at least one group of components affecting the application, said method comprising:
identifying a plurality of individual delay components affecting an application executing over a network;
identifying at least one group of delay components, wherein the at least one group of components comprises a combination of at least two or more of the individual delay components;
determining, by a processing system, respective contributions of each individual delay component to execution of the application;
determining, by the processing system, respective contributions of the identified at least one group of delay components to the application delay; and
outputting information to at least one of a display device and a computer-readable memory indicating respective effects of substantially varying individual delay components and the at least one group of delay components on the application delay.
2. The method of claim 1, further comprising displaying the outputted information to a user the effects of varying individual delay components in comparison to the effects of the at least one group of delay components.
3. The method of claim 1, wherein identifying the at least one group of delay components comprises identifying the at least one group in response to a request by a user.
4. The method of claim 1, wherein identifying the at least one group of delay components comprises identifying at least two components of a parallel delay associated with the application.
5. The method of claim 1, wherein identifying the at least one group of delay components comprises: categorizing each of the individual delay components; and grouping combinations of at least two or more delay components based on their respective category.
6. An analysis system configured to analyze execution of an application over a network, said analysis system comprising:
an application extractor, coupled to portions of the application and the network, configured to receive trace information related to execution of the application over the network;
a computer-readable memory that stores the trace information; and
a controller configured to:
identify delay components contributing to delay in the execution of the application delay,
identify at least one group of two or more of the delay components,
determine values of the delay components based on the trace information,
determine respective contributions of varying each delay component individually,
determine respective contributions of collectively varying the identified at least one group of delay components, and
provide information indicating reductions achieved based on varying the delay components individually and varying the at least one group of delay components collectively.
7. The analysis system of claim 6, further comprising an interactive display providing to a user the information indicating reductions achieved based on varying the delay components individually and varying the at least one group of delay components collectively.
8. The analysis system of claim 6, wherein the controller is configured to identify the at least one group of two or more delay components in response to an input by a user.
9. The analysis system of claim 6, wherein the controller is configured to automatically identify the at least one group of two or more delay components.
10. An analysis system configured to simulate execution of an application over a network, said analysis system comprising:
an application extractor configured to receive trace information related to execution of the application over the network;
a computer-readable memory that stores the trace information; and
a controller configured to:
identify delay components contributing to delay in the execution of the application delay,
identify at least one group of two or more of the delay components,
determine values of the delay components based on the trace information,
determine respective contributions of varying each delay component individually,
determine respective contributions of collectively varying the identified at least one group of delay components, and
provide information indicating anticipated reductions achieved based on varying the delay components individually and varying the at least one group of delay components collectively.
11. The analysis system of claim 10, further comprising an interactive display providing to a user the information indicating simulated reductions achieved based on varying the delay components individually and varying the at least one group of delay components collectively.
12. The analysis system of claim 10, wherein the controller is configured to identify the at least one group of two or more delay components in response to an input by a user.
13. The analysis system of claim 10, wherein the controller is configured to automatically identify the at least one group of two or more delay components.
14. A non-transitory computer-readable medium that includes a program that, when executed by a processor, causes the processor to provide network analysis information regarding execution of an application over a network based on varying individual components and at least one group of components affecting the application, by:
identifying a plurality of individual delay components affecting an application executing over a network;
identifying the at least one group of delay components, wherein the at least one group of components comprise a combination of at least two or more of the individual delay components;
determining respective contributions of each of the individual delay components to execution of the application;
determining respective contributions of the identified at least one group of delay components to the application delay; and
outputting information indicating respective effects of substantially varying individual delay components and the at least one group of delay components on the application delay.
15. The medium of claim 14, further comprising means for displaying the outputted information to a user the effects of varying individual delay components in comparison to the effects of varying the at least one group of delay components.
16. The medium of claim 14, wherein identifying the at least one group of delay components includes identifying the at least one group in response to a request by a user.
17. The medium of claim 14, wherein identifying the at least one group of delay components includes identifying the at least one group of at least two components of a parallel delay by the application.
18. The medium of claim 14, wherein identifying the at least one group of delay components includes categorizing each of the individual delay components and grouping combinations of at least two or more of the delay components based on their respective category.
US13/333,955 2007-05-09 2011-12-21 Network delay analysis including parallel delay effects Active US8745215B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US13/333,955 US8745215B2 (en) 2007-05-09 2011-12-21 Network delay analysis including parallel delay effects

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US91684207P 2007-05-09 2007-05-09
US12/060,271 US8095649B2 (en) 2007-05-09 2008-04-01 Network delay analysis including parallel delay effects
US13/333,955 US8745215B2 (en) 2007-05-09 2011-12-21 Network delay analysis including parallel delay effects

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
US12/060,271 Continuation US8095649B2 (en) 2007-05-09 2008-04-01 Network delay analysis including parallel delay effects

Publications (2)

Publication Number Publication Date
US20130067073A1 US20130067073A1 (en) 2013-03-14
US8745215B2 true US8745215B2 (en) 2014-06-03

Family

ID=39970541

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/060,271 Active 2029-09-02 US8095649B2 (en) 2007-05-09 2008-04-01 Network delay analysis including parallel delay effects
US13/333,955 Active US8745215B2 (en) 2007-05-09 2011-12-21 Network delay analysis including parallel delay effects

Family Applications Before (1)

Application Number Title Priority Date Filing Date
US12/060,271 Active 2029-09-02 US8095649B2 (en) 2007-05-09 2008-04-01 Network delay analysis including parallel delay effects

Country Status (1)

Country Link
US (2) US8095649B2 (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140059202A1 (en) * 2012-08-23 2014-02-27 Efrat Ben David Determining the type of a network tier

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8095649B2 (en) * 2007-05-09 2012-01-10 Opnet Technologies, Inc. Network delay analysis including parallel delay effects
US20100083184A1 (en) * 2008-09-30 2010-04-01 Apple Inc. Selective grouping and manipulation of chart components
US8756312B2 (en) 2010-05-27 2014-06-17 Riverbed Technology, Inc. Multi-tier message correlation
US8683568B1 (en) * 2011-09-22 2014-03-25 Emc Corporation Using packet interception to integrate risk-based user authentication into online services
US20130081001A1 (en) * 2011-09-23 2013-03-28 Microsoft Corporation Immediate delay tracker tool
CN105075323B (en) * 2013-03-29 2019-02-05 Vid拓展公司 Early stage packet loss detecting and feedback
US9191330B2 (en) * 2013-11-14 2015-11-17 International Business Machines Corporation Path selection for network service requests
US10277699B1 (en) * 2015-12-18 2019-04-30 Amazon Technologies, Inc. System for determining process relationships
CN109379247B (en) * 2018-09-27 2022-02-15 腾讯科技(深圳)有限公司 Method and device for detecting network delay of application program

Citations (50)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5375199A (en) 1991-06-04 1994-12-20 Digital Equipment Corporation System monitoring method and device including a graphical user interface to view and manipulate system information
US6061724A (en) 1997-01-29 2000-05-09 Infovista Sa Modelling process for an information system, in particular with a view to measuring performance and monitoring the quality of service, and a measurement and monitoring system implementing this process
US6313768B1 (en) 2000-03-31 2001-11-06 Siemens Information And And Communications Networks, Inc. System and method for trace diagnostics of telecommunications systems
US20020078195A1 (en) 2000-12-20 2002-06-20 Dowell Allen Method and system of telecommunications trace diagnostics by multiple users
US20020124070A1 (en) 2001-03-02 2002-09-05 Pulsipher Eric A. System for providing related information of a network error event in a hand-held device
US6449739B1 (en) 1999-09-01 2002-09-10 Mercury Interactive Corporation Post-deployment monitoring of server performance
US20020198985A1 (en) 2001-05-09 2002-12-26 Noam Fraenkel Post-deployment monitoring and analysis of server performance
US6529954B1 (en) 1999-06-29 2003-03-04 Wandell & Goltermann Technologies, Inc. Knowledge based expert analysis system
US20030065986A1 (en) 2001-05-09 2003-04-03 Fraenkel Noam A. Root cause analysis of server system performance degradations
US20030131098A1 (en) 2001-07-17 2003-07-10 Huntington Stephen G Network data retrieval and filter systems and methods
US20030135612A1 (en) 2001-07-17 2003-07-17 Huntington Stephen Glen Full time network traffic recording systems and methods
US20040049693A1 (en) 2002-09-11 2004-03-11 Enterasys Networks, Inc. Modular system for detecting, filtering and providing notice about attack events associated with network security
US20040054776A1 (en) 2002-09-16 2004-03-18 Finisar Corporation Network expert analysis process
US20040057389A1 (en) 2002-09-16 2004-03-25 Finisar Corporation Network analysis scalable analysis tool for multiple protocols
US20040059807A1 (en) 2002-09-16 2004-03-25 Finisar Corporation Network analysis topology detection
US20040143663A1 (en) * 2002-08-14 2004-07-22 Leedom David Arlen Method and apparatus for monitoring and controlling the allocation of network bandwidth
US6781959B1 (en) 2000-07-31 2004-08-24 Cisco Technology, Inc. Method and apparatus for determining troubleshooting information for completed calls in a telecommunications network
US6801940B1 (en) 2002-01-10 2004-10-05 Networks Associates Technology, Inc. Application performance monitoring expert
US20050064820A1 (en) 2001-12-07 2005-03-24 Park Sang Jin Network analyzing method and apparatus for optimal performance of network, and a recording medium having programs to conduct said method
US20050102402A1 (en) 2000-10-07 2005-05-12 Whitehead Stephen C. Remote communications with a vending machine using call back
US20050108379A1 (en) 2003-08-01 2005-05-19 West Ridge Networks, Inc. System and methods for simulating traffic generation
US20050195797A1 (en) 2004-03-04 2005-09-08 Kryuchkov Sergey V. System and method for facilitating network troubleshooting
US6975330B1 (en) 2001-08-08 2005-12-13 Sprint Communications Company L.P. Graphic display of network performance information
US7010593B2 (en) 2001-04-30 2006-03-07 Hewlett-Packard Development Company, L.P. Dynamic generation of context-sensitive data and instructions for troubleshooting problem events in a computing environment
US20060098586A1 (en) * 2001-03-09 2006-05-11 Farrell Craig A Method and apparatus for application route discovery
US20060146729A1 (en) * 2000-11-10 2006-07-06 Todd Krautkremer Application service level mediation and method of using the same
US20060190480A1 (en) 2005-02-22 2006-08-24 Transparency Software, Inc. Generation of names related to organization actions
US20060190488A1 (en) 2005-02-22 2006-08-24 Transparency Software, Inc. System and method for determining information related to user interactions with an application
US20060274684A1 (en) 2003-10-03 2006-12-07 Diener Neil R Automated real-time site survey in a shared frequency band environment
US20070067296A1 (en) * 2005-08-19 2007-03-22 Malloy Patrick J Network capacity planning
US7197559B2 (en) 2001-05-09 2007-03-27 Mercury Interactive Corporation Transaction breakdown feature to facilitate analysis of end user performance of a server system
US7290048B1 (en) 2002-03-29 2007-10-30 Hyperformix, Inc. Method of semi-automatic data collection, data analysis, and model generation for the performance analysis of enterprise applications
US7293287B2 (en) 2001-10-25 2007-11-06 General Dynamics C4 Systems, Inc. Method and system for modeling, analysis and display of network security events
US7392234B2 (en) 1999-05-18 2008-06-24 Kom, Inc. Method and system for electronic file lifecycle management
US20080209273A1 (en) * 2007-02-28 2008-08-28 Microsoft Corporation Detect User-Perceived Faults Using Packet Traces in Enterprise Networks
US7467202B2 (en) 2003-09-10 2008-12-16 Fidelis Security Systems High-performance network content analysis platform
US7509229B1 (en) 2002-07-23 2009-03-24 Opnet Technologies, Inc. Bayesian approach to correlating network traffic congestion to performance metrics
US7546368B2 (en) 2005-06-01 2009-06-09 Neustar, Inc. Systems and methods for isolating local performance variation in website monitoring
US20090204704A1 (en) 2002-05-06 2009-08-13 Paul Nicolas Muret System and method for tracking unique visitors to a website
US7577689B1 (en) 2005-06-15 2009-08-18 Adobe Systems Incorporated Method and system to archive data
US7593351B1 (en) 2005-06-30 2009-09-22 Opnet Technologies, Inc. Method and system for collecting and consolidating network traffic information
US7606165B2 (en) 2004-01-30 2009-10-20 Microsoft Corporation What-if analysis for network diagnostics
US7730172B1 (en) 1999-05-24 2010-06-01 Computer Associates Think, Inc. Method and apparatus for reactive and deliberative service level management (SLM)
US7891000B1 (en) 2005-08-05 2011-02-15 Cisco Technology, Inc. Methods and apparatus for monitoring and reporting network activity of applications on a group of host computers
US7925729B2 (en) 2004-12-07 2011-04-12 Cisco Technology, Inc. Network management
US7954144B1 (en) 2000-01-18 2011-05-31 Novell, Inc. Brokering state information and identity among user agents, origin servers, and proxies
US7979522B2 (en) 2005-05-27 2011-07-12 L-Cubed Medical Informatics, Llc System and method for monitoring and displaying radiology image traffic
US7984126B2 (en) 2002-01-22 2011-07-19 Siemens Medical Solutions Usa, Inc. Executable application network impact and load characteristic estimation system
US20110213869A1 (en) 2000-09-25 2011-09-01 Yevgeny Korsunsky Processing data flows with a data flow processor
US8095649B2 (en) * 2007-05-09 2012-01-10 Opnet Technologies, Inc. Network delay analysis including parallel delay effects

Family Cites Families (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP3537356B2 (en) * 1998-12-09 2004-06-14 株式会社日立製作所 Delay factor analysis method in job system
US6393480B1 (en) * 1999-06-21 2002-05-21 Compuware Corporation Application response time prediction
US7133911B1 (en) * 2000-03-06 2006-11-07 Compuware Corporation Response time analysis of network performance
US7130915B1 (en) * 2002-01-11 2006-10-31 Compuware Corporation Fast transaction response time prediction across multiple delay sources
US7401141B2 (en) * 2003-01-07 2008-07-15 International Business Machines Corporation Method and system for monitoring performance of distributed applications
US7624178B2 (en) * 2006-02-27 2009-11-24 International Business Machines Corporation Apparatus, system, and method for dynamic adjustment of performance monitoring
US8982709B2 (en) * 2007-01-31 2015-03-17 Hewlett-Packard Development Company, L.P. Selecting service nodes for an end-to-end service path from a reduced search space

Patent Citations (54)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5375199A (en) 1991-06-04 1994-12-20 Digital Equipment Corporation System monitoring method and device including a graphical user interface to view and manipulate system information
US6061724A (en) 1997-01-29 2000-05-09 Infovista Sa Modelling process for an information system, in particular with a view to measuring performance and monitoring the quality of service, and a measurement and monitoring system implementing this process
US7392234B2 (en) 1999-05-18 2008-06-24 Kom, Inc. Method and system for electronic file lifecycle management
US20080263112A1 (en) 1999-05-18 2008-10-23 Kom Inc. Method and system for electronic file lifecycle management
US7730172B1 (en) 1999-05-24 2010-06-01 Computer Associates Think, Inc. Method and apparatus for reactive and deliberative service level management (SLM)
US6529954B1 (en) 1999-06-29 2003-03-04 Wandell & Goltermann Technologies, Inc. Knowledge based expert analysis system
US6449739B1 (en) 1999-09-01 2002-09-10 Mercury Interactive Corporation Post-deployment monitoring of server performance
US7954144B1 (en) 2000-01-18 2011-05-31 Novell, Inc. Brokering state information and identity among user agents, origin servers, and proxies
US6313768B1 (en) 2000-03-31 2001-11-06 Siemens Information And And Communications Networks, Inc. System and method for trace diagnostics of telecommunications systems
US6781959B1 (en) 2000-07-31 2004-08-24 Cisco Technology, Inc. Method and apparatus for determining troubleshooting information for completed calls in a telecommunications network
US20110213869A1 (en) 2000-09-25 2011-09-01 Yevgeny Korsunsky Processing data flows with a data flow processor
US20050102402A1 (en) 2000-10-07 2005-05-12 Whitehead Stephen C. Remote communications with a vending machine using call back
US20060146729A1 (en) * 2000-11-10 2006-07-06 Todd Krautkremer Application service level mediation and method of using the same
US6871227B2 (en) 2000-12-20 2005-03-22 Siemens Communications, Inc. Method and system of telecommunications trace diagnostics by multiple users
US20020078195A1 (en) 2000-12-20 2002-06-20 Dowell Allen Method and system of telecommunications trace diagnostics by multiple users
US20020124070A1 (en) 2001-03-02 2002-09-05 Pulsipher Eric A. System for providing related information of a network error event in a hand-held device
US20060098586A1 (en) * 2001-03-09 2006-05-11 Farrell Craig A Method and apparatus for application route discovery
US7010593B2 (en) 2001-04-30 2006-03-07 Hewlett-Packard Development Company, L.P. Dynamic generation of context-sensitive data and instructions for troubleshooting problem events in a computing environment
US20030065986A1 (en) 2001-05-09 2003-04-03 Fraenkel Noam A. Root cause analysis of server system performance degradations
US7197559B2 (en) 2001-05-09 2007-03-27 Mercury Interactive Corporation Transaction breakdown feature to facilitate analysis of end user performance of a server system
US20020198985A1 (en) 2001-05-09 2002-12-26 Noam Fraenkel Post-deployment monitoring and analysis of server performance
US20030131098A1 (en) 2001-07-17 2003-07-10 Huntington Stephen G Network data retrieval and filter systems and methods
US20030135612A1 (en) 2001-07-17 2003-07-17 Huntington Stephen Glen Full time network traffic recording systems and methods
US6975330B1 (en) 2001-08-08 2005-12-13 Sprint Communications Company L.P. Graphic display of network performance information
US7293287B2 (en) 2001-10-25 2007-11-06 General Dynamics C4 Systems, Inc. Method and system for modeling, analysis and display of network security events
US20050064820A1 (en) 2001-12-07 2005-03-24 Park Sang Jin Network analyzing method and apparatus for optimal performance of network, and a recording medium having programs to conduct said method
US7218928B2 (en) 2001-12-07 2007-05-15 Sk Telecom Co., Ltd. Network analyzing method and apparatus for optimal performance of network, and a recording medium having programs to conduct said method
US6801940B1 (en) 2002-01-10 2004-10-05 Networks Associates Technology, Inc. Application performance monitoring expert
US7984126B2 (en) 2002-01-22 2011-07-19 Siemens Medical Solutions Usa, Inc. Executable application network impact and load characteristic estimation system
US7290048B1 (en) 2002-03-29 2007-10-30 Hyperformix, Inc. Method of semi-automatic data collection, data analysis, and model generation for the performance analysis of enterprise applications
US20090204704A1 (en) 2002-05-06 2009-08-13 Paul Nicolas Muret System and method for tracking unique visitors to a website
US7509229B1 (en) 2002-07-23 2009-03-24 Opnet Technologies, Inc. Bayesian approach to correlating network traffic congestion to performance metrics
US20040143663A1 (en) * 2002-08-14 2004-07-22 Leedom David Arlen Method and apparatus for monitoring and controlling the allocation of network bandwidth
US20040049693A1 (en) 2002-09-11 2004-03-11 Enterasys Networks, Inc. Modular system for detecting, filtering and providing notice about attack events associated with network security
US20040057389A1 (en) 2002-09-16 2004-03-25 Finisar Corporation Network analysis scalable analysis tool for multiple protocols
US20040054776A1 (en) 2002-09-16 2004-03-18 Finisar Corporation Network expert analysis process
US20040059807A1 (en) 2002-09-16 2004-03-25 Finisar Corporation Network analysis topology detection
US20050108379A1 (en) 2003-08-01 2005-05-19 West Ridge Networks, Inc. System and methods for simulating traffic generation
US7467202B2 (en) 2003-09-10 2008-12-16 Fidelis Security Systems High-performance network content analysis platform
US20060274684A1 (en) 2003-10-03 2006-12-07 Diener Neil R Automated real-time site survey in a shared frequency band environment
US7606165B2 (en) 2004-01-30 2009-10-20 Microsoft Corporation What-if analysis for network diagnostics
US20050195797A1 (en) 2004-03-04 2005-09-08 Kryuchkov Sergey V. System and method for facilitating network troubleshooting
US20110167145A1 (en) 2004-12-07 2011-07-07 Pure Networks, Inc. Network management
US7925729B2 (en) 2004-12-07 2011-04-12 Cisco Technology, Inc. Network management
US20060190488A1 (en) 2005-02-22 2006-08-24 Transparency Software, Inc. System and method for determining information related to user interactions with an application
US20060190480A1 (en) 2005-02-22 2006-08-24 Transparency Software, Inc. Generation of names related to organization actions
US7979522B2 (en) 2005-05-27 2011-07-12 L-Cubed Medical Informatics, Llc System and method for monitoring and displaying radiology image traffic
US7546368B2 (en) 2005-06-01 2009-06-09 Neustar, Inc. Systems and methods for isolating local performance variation in website monitoring
US7577689B1 (en) 2005-06-15 2009-08-18 Adobe Systems Incorporated Method and system to archive data
US7593351B1 (en) 2005-06-30 2009-09-22 Opnet Technologies, Inc. Method and system for collecting and consolidating network traffic information
US7891000B1 (en) 2005-08-05 2011-02-15 Cisco Technology, Inc. Methods and apparatus for monitoring and reporting network activity of applications on a group of host computers
US20070067296A1 (en) * 2005-08-19 2007-03-22 Malloy Patrick J Network capacity planning
US20080209273A1 (en) * 2007-02-28 2008-08-28 Microsoft Corporation Detect User-Perceived Faults Using Packet Traces in Enterprise Networks
US8095649B2 (en) * 2007-05-09 2012-01-10 Opnet Technologies, Inc. Network delay analysis including parallel delay effects

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20140059202A1 (en) * 2012-08-23 2014-02-27 Efrat Ben David Determining the type of a network tier
US9178771B2 (en) * 2012-08-23 2015-11-03 Hewlett-Packard Development Company, L.P. Determining the type of a network tier

Also Published As

Publication number Publication date
US20130067073A1 (en) 2013-03-14
US20080281961A1 (en) 2008-11-13
US8095649B2 (en) 2012-01-10

Similar Documents

Publication Publication Date Title
US8745215B2 (en) Network delay analysis including parallel delay effects
US7885200B2 (en) Application delay analysis
US8078691B2 (en) Web page load time prediction and simulation
US20050154576A1 (en) Policy simulator for analyzing autonomic system management policy of a computer system
US8195443B2 (en) Application level interface to network analysis tools
US8917624B2 (en) Click quality classification and delivery
US7616568B2 (en) Generic packet generation
US9300523B2 (en) System and method for performance management in a multi-tier computing environment
US6775644B2 (en) System load testing coordination over a network
US8387059B2 (en) Black-box performance control for high-volume throughput-centric systems
US20130046887A1 (en) Network capacity planning for multiple instances of an application
US20030117959A1 (en) Methods and apparatus for placement of test packets onto a data communication network
WO2003088576A1 (en) Method, system and computer program for determining network operational characteristics of software applications
US8121031B2 (en) Structure for a multi-scale network traffic generator
Wierman et al. Scheduling despite inexact job-size information
JP5112277B2 (en) Reproduction processing method, computer system, and program
US7130915B1 (en) Fast transaction response time prediction across multiple delay sources
US7839790B2 (en) Network congestion analysis
Bacigalupo et al. An investigation into the application of different performance prediction methods to distributed enterprise applications
US20030187617A1 (en) User perception tool
US20150046480A1 (en) Information processing device for executing load evaluation, and load evaluation method
CN115277436B (en) Micro-service software architecture identification method based on topological structure
Cardoso et al. HTTP traffic modeling: Development and application
KR102082486B1 (en) Method for transmitting and receiving data
Primorac Understanding and Mitigating Latency Variability of Latency-Critical Applications

Legal Events

Date Code Title Description
AS Assignment

Owner name: MORGAN STANLEY & CO. LLC, MARYLAND

Free format text: SECURITY AGREEMENT;ASSIGNORS:RIVERBED TECHNOLOGY, INC.;OPNET TECHNOLOGIES, INC.;REEL/FRAME:029646/0060

Effective date: 20121218

AS Assignment

Owner name: OPNET TECHNOLOGIES, INC., MARYLAND

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:NIEMCZYK, STEVEN;MALLOY, PATRICK J.;COHEN, ALAIN J.;AND OTHERS;SIGNING DATES FROM 20080318 TO 20080328;REEL/FRAME:030116/0104

AS Assignment

Owner name: OPNET TECHNOLOGIES LLC, MARYLAND

Free format text: CHANGE OF NAME;ASSIGNOR:OPNET TECHNOLOGIES, INC.;REEL/FRAME:030411/0273

Effective date: 20130401

AS Assignment

Owner name: RIVERBED TECHNOLOGY, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:OPNET TECHNOLOGIES LLC;REEL/FRAME:030462/0148

Effective date: 20130401

AS Assignment

Owner name: RIVERBED TECHNOLOGY, INC., CALIFORNIA

Free format text: RELEASE OF PATENT SECURITY INTEREST;ASSIGNOR:MORGAN STANLEY & CO. LLC, AS COLLATERAL AGENT;REEL/FRAME:032113/0425

Effective date: 20131220

AS Assignment

Owner name: JPMORGAN CHASE BANK, N.A., AS ADMINISTRATIVE AGENT, NEW YORK

Free format text: PATENT SECURITY AGREEMENT;ASSIGNOR:RIVERBED TECHNOLOGY, INC.;REEL/FRAME:032421/0162

Effective date: 20131220

Owner name: JPMORGAN CHASE BANK, N.A., AS ADMINISTRATIVE AGENT

Free format text: PATENT SECURITY AGREEMENT;ASSIGNOR:RIVERBED TECHNOLOGY, INC.;REEL/FRAME:032421/0162

Effective date: 20131220

STCF Information on status: patent grant

Free format text: PATENTED CASE

AS Assignment

Owner name: RIVERBED TECHNOLOGY, INC., CALIFORNIA

Free format text: RELEASE OF SECURITY INTEREST IN PATENTS;ASSIGNOR:BARCLAYS BANK PLC;REEL/FRAME:035521/0069

Effective date: 20150424

AS Assignment

Owner name: MORGAN STANLEY SENIOR FUNDING, INC., AS COLLATERAL AGENT, NEW YORK

Free format text: SECURITY INTEREST;ASSIGNOR:RIVERBED TECHNOLOGY, INC.;REEL/FRAME:035561/0363

Effective date: 20150424

Owner name: MORGAN STANLEY SENIOR FUNDING, INC., AS COLLATERAL

Free format text: SECURITY INTEREST;ASSIGNOR:RIVERBED TECHNOLOGY, INC.;REEL/FRAME:035561/0363

Effective date: 20150424

AS Assignment

Owner name: RIVERBED TECHNOLOGY, INC., CALIFORNIA

Free format text: CORRECTIVE ASSIGNMENT TO CORRECT THE CONVEYING PARTY NAME PREVIOUSLY RECORDED ON REEL 035521 FRAME 0069. ASSIGNOR(S) HEREBY CONFIRMS THE RELEASE OF SECURITY INTEREST IN PATENTS;ASSIGNOR:JPMORGAN CHASE BANK, N.A.;REEL/FRAME:035807/0680

Effective date: 20150424

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 4TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1551)

Year of fee payment: 4

AS Assignment

Owner name: ALTER DOMUS (US) LLC, AS COLLATERAL AGENT, ILLINOIS

Free format text: PATENT SECURITY AGREEMENT;ASSIGNOR:RIVERBED TECHNOLOGY, INC.;REEL/FRAME:055514/0249

Effective date: 20201231

AS Assignment

Owner name: MACQUARIE CAPITAL FUNDING LLC, NEW YORK

Free format text: SECURITY INTEREST;ASSIGNORS:RIVERBED HOLDINGS, INC.;RIVERBED TECHNOLOGY, INC.;ATERNITY LLC;REEL/FRAME:056397/0750

Effective date: 20210420

AS Assignment

Owner name: ATERNITY LLC, CALIFORNIA

Free format text: RELEASE OF SECURITY INTEREST IN PATENTS RECORED AT REEL 056397, FRAME 0750;ASSIGNOR:MACQUARIE CAPITAL FUNDING LLC;REEL/FRAME:057983/0356

Effective date: 20211012

Owner name: RIVERBED TECHNOLOGY, INC., CALIFORNIA

Free format text: RELEASE OF SECURITY INTEREST IN PATENTS RECORED AT REEL 056397, FRAME 0750;ASSIGNOR:MACQUARIE CAPITAL FUNDING LLC;REEL/FRAME:057983/0356

Effective date: 20211012

Owner name: RIVERBED HOLDINGS, INC., CALIFORNIA

Free format text: RELEASE OF SECURITY INTEREST IN PATENTS RECORED AT REEL 056397, FRAME 0750;ASSIGNOR:MACQUARIE CAPITAL FUNDING LLC;REEL/FRAME:057983/0356

Effective date: 20211012

AS Assignment

Owner name: ALTER DOMUS (US) LLC, AS COLLATERAL AGENT, ILLINOIS

Free format text: PATENT SECURITY AGREEMENT SUPPLEMENT - SECOND LIEN;ASSIGNORS:RIVERBED HOLDINGS, INC.;RIVERBED TECHNOLOGY, INC.;ATERNITY LLC;REEL/FRAME:057810/0559

Effective date: 20211013

Owner name: MORGAN STANLEY SENIOR FUNDING, INC., AS COLLATERAL AGENT, MARYLAND

Free format text: PATENT SECURITY AGREEMENT SUPPLEMENT - FIRST LIEN;ASSIGNORS:RIVERBED HOLDINGS, INC.;RIVERBED TECHNOLOGY, INC.;ATERNITY LLC;REEL/FRAME:057810/0502

Effective date: 20211013

AS Assignment

Owner name: WILMINGTON TRUST, NATIONAL ASSOCIATION, MINNESOTA

Free format text: PATENT SECURITY AGREEMENT;ASSIGNORS:RIVERBED TECHNOLOGY, INC.;ATERNITY LLC;REEL/FRAME:057943/0386

Effective date: 20211013

MAFP Maintenance fee payment

Free format text: PAYMENT OF MAINTENANCE FEE, 8TH YEAR, LARGE ENTITY (ORIGINAL EVENT CODE: M1552); ENTITY STATUS OF PATENT OWNER: LARGE ENTITY

Year of fee payment: 8

AS Assignment

Owner name: WILMINGTON TRUST, NATIONAL ASSOCIATION, AS U.S. COLLATERAL AGENT, MINNESOTA

Free format text: SECURITY INTEREST;ASSIGNORS:RIVERBED TECHNOLOGY LLC (FORMERLY RIVERBED TECHNOLOGY, INC.);ATERNITY LLC;REEL/FRAME:058486/0216

Effective date: 20211207

AS Assignment

Owner name: ATERNITY LLC, MASSACHUSETTS

Free format text: TERMINATION AND RELEASE OF SECURITY INTEREST IN PATENTS;ASSIGNOR:WILMINGTON TRUST, NATIONAL ASSOCIATION, AS U.S. COLLATERAL AGENT;REEL/FRAME:058593/0169

Effective date: 20211207

Owner name: RIVERBED TECHNOLOGY, INC., CALIFORNIA

Free format text: TERMINATION AND RELEASE OF SECURITY INTEREST IN PATENTS;ASSIGNOR:WILMINGTON TRUST, NATIONAL ASSOCIATION, AS U.S. COLLATERAL AGENT;REEL/FRAME:058593/0169

Effective date: 20211207

Owner name: ATERNITY LLC, MASSACHUSETTS

Free format text: TERMINATION AND RELEASE OF SECURITY INTEREST IN PATENTS;ASSIGNOR:ALTER DOMUS (US) LLC, AS COLLATERAL AGENT;REEL/FRAME:058593/0108

Effective date: 20211207

Owner name: RIVERBED TECHNOLOGY, INC., CALIFORNIA

Free format text: TERMINATION AND RELEASE OF SECURITY INTEREST IN PATENTS;ASSIGNOR:ALTER DOMUS (US) LLC, AS COLLATERAL AGENT;REEL/FRAME:058593/0108

Effective date: 20211207

Owner name: ATERNITY LLC, MASSACHUSETTS

Free format text: TERMINATION AND RELEASE OF SECURITY INTEREST IN PATENTS;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC., AS COLLATERAL AGENT;REEL/FRAME:058593/0046

Effective date: 20211207

Owner name: RIVERBED TECHNOLOGY, INC., CALIFORNIA

Free format text: TERMINATION AND RELEASE OF SECURITY INTEREST IN PATENTS;ASSIGNOR:MORGAN STANLEY SENIOR FUNDING, INC., AS COLLATERAL AGENT;REEL/FRAME:058593/0046

Effective date: 20211207

AS Assignment

Owner name: RIVERBED TECHNOLOGY LLC, CALIFORNIA

Free format text: CHANGE OF NAME;ASSIGNOR:RIVERBED TECHNOLOGY, INC.;REEL/FRAME:059232/0551

Effective date: 20211207

AS Assignment

Owner name: RIVERBED HOLDINGS, INC., CALIFORNIA

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC, AS COLLATERAL AGENT;REEL/FRAME:064673/0739

Effective date: 20211207

Owner name: ATERNITY LLC, MASSACHUSETTS

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC, AS COLLATERAL AGENT;REEL/FRAME:064673/0739

Effective date: 20211207

Owner name: RIVERBED TECHNOLOGY, INC., CALIFORNIA

Free format text: RELEASE BY SECURED PARTY;ASSIGNOR:ALTER DOMUS (US) LLC, AS COLLATERAL AGENT;REEL/FRAME:064673/0739

Effective date: 20211207