CA2388712A1 - Distributed network management system and method - Google Patents

Distributed network management system and method Download PDF

Info

Publication number
CA2388712A1
CA2388712A1 CA002388712A CA2388712A CA2388712A1 CA 2388712 A1 CA2388712 A1 CA 2388712A1 CA 002388712 A CA002388712 A CA 002388712A CA 2388712 A CA2388712 A CA 2388712A CA 2388712 A1 CA2388712 A1 CA 2388712A1
Authority
CA
Canada
Prior art keywords
server
primary
remote
hub
hub server
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
CA002388712A
Other languages
French (fr)
Inventor
John Carlson
Tim Hinderliter
Leigh Metcalf
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Internap Network Services Corp
Original Assignee
Internap Network Services Corporation
John Carlson
Tim Hinderliter
Leigh Metcalf
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Internap Network Services Corporation, John Carlson, Tim Hinderliter, Leigh Metcalf filed Critical Internap Network Services Corporation
Publication of CA2388712A1 publication Critical patent/CA2388712A1/en
Abandoned legal-status Critical Current

Links

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/02Standardisation; Integration
    • H04L41/0213Standardised network management protocols, e.g. simple network management protocol [SNMP]
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L1/00Arrangements for detecting or preventing errors in the information received
    • H04L1/22Arrangements for detecting or preventing errors in the information received using redundant apparatus to increase reliability
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L41/00Arrangements for maintenance, administration or management of data switching networks, e.g. of packet switching networks
    • H04L41/12Discovery or management of network topologies
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/08Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters
    • H04L43/0805Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters by checking availability
    • H04L43/0817Monitoring or testing based on specific metrics, e.g. QoS, energy consumption or environmental parameters by checking availability by checking functioning
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/40Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass for recovering from a failure of a protocol instance or entity, e.g. service redundancy protocols, protocol state redundancy or protocol service redirection
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/10Active monitoring, e.g. heartbeat, ping or trace-route
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/10Active monitoring, e.g. heartbeat, ping or trace-route
    • H04L43/103Active monitoring, e.g. heartbeat, ping or trace-route with adaptive polling, i.e. dynamically adapting the polling rate
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/10Active monitoring, e.g. heartbeat, ping or trace-route
    • H04L43/106Active monitoring, e.g. heartbeat, ping or trace-route using time related information in packets, e.g. by adding timestamps
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L43/00Arrangements for monitoring or testing data switching networks
    • H04L43/16Threshold monitoring

Abstract

A distributed network management system (10) and method of operation. The system (10) includes at least one hub server (12) and at least one remote server (16), where the hub server (12) and the remote server (16) communicate with each other. The remote server (16) additionally communicates with and monitors one or more network devices (20). In the event that the remote server (16) becomes inoperational, the hub server (12) assumes monitoring of the network device (20). For redundancy, primary (12) and secondary (14) hub servers can be provided, wherein the primary (12) and secondary (14) hub servers communicate with each other and are capable of communicating with the remote server (16). For further redundancy, primary (16) and secondary (18) remote servers can be provided, wherein the primary (16) and secondary (18) remote servers communicate with each other but independently monitor the network devices (20). In the peered remote configuration, the hub server (12) is capable of communicating with either of the remote servers (16, 18). Where both the hub servers (12, 14) and the remote servers (16,18) are peered, each hub server (12, 14) is capable of communicating with each remote server (16, 18).

Description

DISTRIBUTED NETWORK MANAGEMENT SYSTEM AND METHOD
BACKGROUND OF THE INVENTION
1. Field of the Invention This invention pertains generally to network communications, and more particularly to monitoring and managing network performance.
2. Description of the Background Art In the operation of interconnected networks, it is often desirable to have a mechanism for monitoring the state of equipment and devices in the network.
Traditionally, this has been accomplished using a centrally-based network management system, with a plurality of individual network management systems feeding up to the central network management system in a conventional tree hierarchy. Equipment and devices would similarly feed up to the individual network management systems in a conventional tree hierarchy.
Unfortunately, such a architecture for a network management system does not scale well and does not provide for propagation of state and configuration information among a set of cooperating systems.
BRIEF SUMMARY OF THE INVENTION
The present invention is a scalable distributed network management system with the potential for full redundancy at hub and remote levels. The remotes monitor state changes of network devices, and those state changes propagate bidirectionally between hubs and remotes. Furthermore, configuration changes for designating the monitoring parameters of the remotes propagate bidirectionally between remotes and hubs.
By way of example, and not of limitation, the system includes at least one hub server acid at least one remote server, where the hub server and the remote server communicate with each other. The remote server additionally communicates with and monitors one or more network devices. In the event that the remote server becomes inoperational, the hub server assumes monitoring of the network device(s).
According to another aspect of the invention, for redundancy, primary and secondary hub servers can be provided, wherein the primary and secondary hub servers communicate with each other. In this peered hub configuration, if the primary hub server becomes inoperational and the secondary hub server is operational, the secondary hub server communicates with the remote server. Additionally, in the peered hub configuration, if both WO 02/03211 . PCT/US00/23728 the primary hub server and the remote server are inoperational, the secondary hub server assumes monitoring of the network devices.
According to another aspect of the invention, for redundancy, primary and secondary remote servers can be provided, wherein the primary and secondary remote servers cormnunicate with each other but independently monitor the network devices. In the peered remote configuration, if the primary remote server becomes inoperational, the primary hub communicates with the secondary remote.
According to a still further aspect of the invention, if the remotes and the hubs are peered and the primary hub is inoperational, the secondary hub communicates with the primary remote thereby temporarily assuming the duties of the primary hub.
Also in the peered hub and peered remote configuration, if both the primary hub and primary remote are inoperational, the secondary hub communicates with the secondary remote. If both remotes are inoperational, then all active hubs assume monitoring of the network devices.
To facilitate monitoring of network devices, the invention derives state information from network devices using what is referred to herein as the Leigh/Tim Paradigm or LTP. In LTP, a plurality of pings is sent from an ICMP server to an interface address on a network device during a polling interval. The number of pings returned from said network device is recorded and converted to a percentage based on the ratio of the number of pings sent to the number of pings received. Next, an SNMP query is sent to the network device and the operational status of the network device, such as "up", "down" or "unknown" is determined from the SNMP query. Using the percentage of pings returned and the SNMP
status, a status percentage for the polling period is generated by multiplying the percentage pings returned by a constant associated with the operational status, where the constant has a first value if the operational status is "up", a second value if the operational status is down", and a third value if the operational status is "unknown". Next, a weighted average of the status percentages for the current and previous four polling periods is computed. Then, the state of the network device is determined from the weighted average.
An object of the invention is to provide a distributed network management system where configuration information propagates bidirectionally through the system.
Another object of the invention is to provide a distributed network management system where configuration information can be entered at one location and propagate through the system.

Another object of the invention is to provide a distributed network management system which can be accessed through a web server.
Another object of the invention is to provide a distributed network management system where state changes propagate bidirectionally through the system.
Another object of the invention is to provide a peered distributed network management system with automatic failover and resynchronization.
Another object of the invention is to provide a distributed network management system which consolidates multiple status notifications into a single notification one based on an interface hierarchy.
Another object of the invention is to provide a distributed network management system with a plug-in architecture of service, notification and utility modules.
Another object of the invention is to provide a distributed network management system that can serve as an information transport.
Further objects and advantages of the invention will be brought out in the following portions of the specification, wherein the detailed description is for the purpose of fully disclosing preferred embodiments of the invention without placing limitations thereon.
BRIEF DESCRIPTION OF THE DRAWINGS
The invention will be more fully understood by reference to the following drawings which are for illustrative purposes only:
FIG. 1 is a schematic diagram of the high level architecture of an embodiment of a distributed network management system according to the invention depicting the primary hub and the primary remote as being operational, and the primary hub as communicating with the primary remote.
FIG. 2 is a schematic diagram of the distributed network management system of FIG.
1 depicting the primary hub as being operational, the primary remote as being inoperational, the secondary remote as being operational, and the primary hub communicating with the secondary remote.
FIG. 3 is a schematic diagram of the distributed network management system of FIG.
1 depicting the primary hub as being inoperational, the secondary hub as being operational, the primary remote as being operational, and the secondary hub communicating with the primary remote.
FIG. 4 a schematic diagram of the distributed network management system of FIG. 1 depicting the primary hub as being inoperational, the secondary hub as being operational, the primary remote as being inoperational, the secondary remote as being operational, and the secondary hub communicating with the secondary remote.
FIG. 5 is a schematic diagram of the distributed network management system of FIG.
1 depicting the primary and secondary remotes as being inoperational, and the primary and secondary hubs communicating with the network devices.
FIG. 6 is a schematic diagram of an implementation of a distributed network management system according to the invention.
FIG. 7 is schematic diagram showing an alternative embodiment of the distributed network management system implementation of FIG. 6 wherein hubs are regionalized.
FIG. 8 is a functional block diagram of the internal architecture of a remote according to the present invention.
FIG. 9 is a functional block diagram of an alternative embodiment of the remote architecture of FIG. 8.
FIG. 10 is a functional block diagram of the dNMS kernel portion of a remote according to the present invention.
FIG. 11 is a schematic diagram of an integration server in the dNMS kernel of FIG. 10.
FIG. 12 is a schematic diagram of a monolithic server in the dNMS kernel of FIG. 10.
FIG. 13 is a schematic diagram showing data flow between the integration server of FIG. 1 l and the monolithic server of FIG. 12.
FIG. 14 is a schematic diagram depicting traffic flow between hubs and remotes through queuing according to the invention.
DETAILED DESCRIPTION OF THE INVENTION
Referring more specifically to the drawings, for illustrative purposes the present invention is embodied in the components, system and methods generally shown in FIG. 1 through FIG. 14. It will be appreciated that the invention may vary as to configuration and details without departing from the basic concepts as disclosed herein.
FIG. 1 is a schematic diagram of the high level architecture 10 of an embodiment of a distributed network management system according to the present invention. In the embodiment shown, the system comprises a primary hub 12 and a secondary hub 14, both of which can communicate with a primary remote 16 and a secondary remote 18. The remotes in turn communicate with a specific set of devices 20 on nodes 22 of the network 24, such as routers, to monitor network status. The network may be all or a portion of the Internet or other wide area network. The set of network devices is selected to provide an overall representation of the network being monitored.
Each hub is in active communication with the other hub through a full-time communications link 26 for redundancy, so that data received from one hub is continuously propagated to the other. Similarly, each remote is in active communication with the other remote through a full-time communications link 28 for redundancy and for continuously propagating data to the other remote. In addition, each remote is in constant communication with each network device. However, each remote preferably monitors the network devices independent of the other remote. As a result, the data acquired by a remote may disagree with the data acquired by the other remote, even though both remotes are monitoring the same network devices. Because the remotes operate independently of each other, the monitoring times could be different and a particular remote may observe a network condition that was not observed by the other remote. For example, one remote may monitor conditions thirty seconds into each minute, while another remote may monitor conditions forty-five seconds into each minute.
Primary hub 12 is in full-time communication with primary remote 16 through communication link 30 so that changes detected by primary remote 16 is continuously propagated to primary hub 12 as well as to secondary hub 14 through primary hub 12. In addition, configuration data such as which network devices to monitor can be propagated to primary remote 16 and to secondary remote 18 through primary remote 16. Note, however, that there is also a normally inactive communication link 32 between secondary hub 14 and secondary remote 18, a normally inactive communications link 34 between secondary hub 14 and primary remote 16, and a normally inactive communications link 36 between primary hub 12 and secondary remote 18. These communications links are not necessarily direct physical links, however. In the preferred embodiment of the invention, each remote and network device has an address, such as an Internet Protocol (IP) address. This allows the remote or network device to be accessed over a network such as, for example, the Internet. In addition, each hub can communicate directly with a network device as well.
With the architecture described above, the preferred communications hierarchy is as follows:
1. if the primary hub and the primary remote are operational, the primary hub communicates with the primary remote as shown in FIG. 1.
2. if the primary hub is operational, the primary remote is inoperational, and the secondary remote is operational, the primary hub communicates with the secondary remote as shown in FIG. 2.
3. if the primary hub is inoperational, the secondary hub is operational, and the primary remote is operational, the secondary hub communicates with the primary remote as shown in FIG. 3.
4. if the primary hub is inoperational, the secondary hub is operational, the primary remote is inoperational, and the secondary remote is operational, the secondary hub communicates with the secondary remote as shown in FIG. 4.
5. if both the primary and secondary remotes are inoperational, all active hubs assume monitoring of the remote network as shown in FIG. 5.
Referring now to FIG. 6, an example of a possible geographical configuration of a distributed networlc management system according to the invention is shown. In FIG. 6, a first set of hubs 38 is shown located in the vicinity of Seattle and a second set of hubs 40 is shown located in the vicinity of New York City. Also shown are several sets of remotes 42, 44, 46, 48, 50, 52, 54, 56, and 58, each of which monitors a portion of the overall network.
Note that hubs 38 monitor remotes 42, 44, 46, and 48, while hubs 40 monitor remotes 50, 52, 54, 56, and 58. A change of state monitored by, for example, remotes 50 will propagate to hubs 40 in New York City, and from hubs 40 to sister hubs 38 in Seattle so that both sets of hubs have the same state information.
While the foregoing configuration is scalable, the addition of a larger number of remotes or hubs can become more complex than necessary. In that event, an additional monitoring layer can be added above the hubs. In this way, not only are remotes assigned to regions of the network, but hubs are assigned to regions of the network as well. For example, referring to FIG. 7, three regions 60, 62 and 64 are shown. Each region would include a primary and secondary hub that would be responsible for that region. For example, primary hub 66 and secondary hub 68 would be responsible for region A, primary hub 70 and secondary hub 72 would be responsible for region 62, and primary hub 74 and secondary hub 76 would be responsible for region 64. In turn the hubs in a particular region would be responsible for several sets of primary and secondary remotes in that region, such as set 78, 78', 78" ... in region 60, set 80, 80', 80" ... in region 62, and set 82, 82' and 82" in region 64, and each set of remotes would be responsible for a portion of the network devices therein.
The data collected by the primary hubs in each region would be propagated to a primary hub aggregator 84, which in turn would propagate the data to a secondary hub aggregator 86 for redundancy. In this way, a mufti-level distributed system architecture can be achieved.
Referring now to FIG. l and FIG. 8, an embodiment of the internal architecture of primary 16 and secondary 18 remote is shown. Each remote includes a dNMS
kernel 88 that, in addition to other functions that will be described, acquires data from the network 24. Also shown is a scheduler 90, which is a plug-in service that notifies administrative personnel that a problem exists on the network being monitored.
Each remote is accessible through a client terminal 92 running a browser-based application interface. Note that data propagates from the network to each dNMS
kernel through a data path 94, and that configuration changes received from a hub (not shown) propagates to each dNMS kernel through a configuration path 96.
Optionally, the remotes can include a collector 98, which is also a plug-in service, to which data from the network propagates and is stored in data files 100 for billing or other purposes. Also shown is a module 102 for mining the stored data and a module 104 for collating the mined data into a central database 106 accessible by a client terminal 108 for billing. The details of those components are not described herein as they do not form a part of the invention and are shown solely to indicate additional ways in which the data acquired by a remote can be used. In the event that such additional uses of the data are made, processing overhead of the remotes may increase. In that event, it is preferred to reduce the load on the primary remote by moving the auxiliary data collection functions into a separate remote server 110 as shown in FIG. 9. The primary remote 16 is then dedicated to monitoring network conditions, while server 110 is dedicated to the auxiliary data collection functions. Secondary remote 18 can be configured as before, or unloaded in the same way.
Note that primary 12 and secondary 14 hubs in FIG. 1 would be configured in the same manner as the remotes. Note also that configuration information, as well as state information, propagates bidirectionally between hubs and remotes and between peers (e.g., hub to hub or remote to remote).

As can be seen, therefore, a critical element of a hub and a remote is the dNMS
lcernel 88. Referring now to FIG. 10, which shows primary remote 16 as an example, the high level architecture of dNMS kernel 88 comprises an integration server 112 and a monolithic server 114. Integration server 112 communicates with client terminal 92 and monolithic server 114 communicates with the network devices comiected to network 24.
In the case of a remote, state information relating to the network devices collected by monolithic server 114 is propagated to integration server 112 and then propagated to the integration server in primary hub 12, for example. Furthermore, in the case of a remote, configuration information such as the IP addresses of the network devices to be monitored is entered into integration server 110 from client terminal 92, from which it propagates down to monolithic server 112 as well as propagates up to the integration server in primary hub 12.
Alternatively, configuration information can be entered into a hub, in which case the configuration information propagates down to integration server and the monolithic server n the remotes. While configuration information is entered into a dNMS kernel by a client terminal, state information for the network devices is acquired. In the preferred embodiment of the invention, state information is derived using what will be referred to herein as LTP, which is an acronym developed by the inventors herein. LTP provides for simple real time monitoring of network devices and their interfaces using ICMP, SNMP or a combination thereof, and employs a sliding window to compensate for minor interruptions in Internet links or IP traffic.
In LTP according to the present invention, a polling interval is defined during which each ICMP server sends out a plurality of ICMP echo requests, or pings. While the polling interval and number of pings can vary, in the preferred embodiment ten pings are sent every sixty seconds, with each ping being separated by a one-second interval. The number of pings that are returned is converted to a percentage for that polling interval.
In addition, for that same polling interval, if the node is SNMP-enabled (which may not be the case for servers and other non-router equipment), an SNMP query is sent to the node on which the interface resides. The "operational status" of the interface is queried as to three possible states: "up", "down", and "unknown". An "unknown" operational status means that the SNMP request was never returned and, therefore, the system does not know the status.
Using the percentage of pings returned and the SNMP status, a single number is generated for the polling period. This number is generated by multiplying the percentage of pings returned by a constant that is assigned depending on the result of the SNMP query;
namely, a value of one if the query returned "up", a value of zero if the query returned "down", and a value of 0.4 if the query returned "unknown. In essence, the SNMP query returned "up", we simply use the percentage of returned ICMP packets. If the query returned "down", we discard the ICMP information and take the time period as being zero percent. If the query returned "unknown", we assume that there is a routing problem and multiply the percentage ICMP packets by an arbitrary value of four tenths (0.4). For example, if ten out of ten pings are returned during a polling interval, but we were unable to obtain SNMP
information for that interface during that time period, the ratio for that time period would be forty percent (40%). Table 1 shows examples of various network conditions, given different SNMP and ICMP values, including the total ratio computed for the time period.
Once the percentage is computed in this manner, the next step is to compute a weighted average of the percentages for current and previous four time periods. This is preferably carried out by with a five element table with a sliding window. The percentage for the current time is inserted in the rightmost (e.g., current period) slot. If the current period slot is not empty, all values in the table are shifted to the left by one slot (i.e., the oldest data is dropped). Therefore, each position in the table represents a different time period's ratio.
The leftmost slot contains data that is four polling intervals old and, as the table is transversed to the right, the data is more recent.
Each position in the table is also assigned a weight, which affects the extent to which that position in the table will influence the final percentage; that is, the state of the interface.
Higher weights are assigned to the more recent polling intervals, as they are more indicative of the current state. Note, however, that the weights should not be too high;
otherwise, the result will be over-notification of problems with the interface. In other words, if the weights are set too high, the normal intermittency in the Internet will result in unnecessary notification. By keeping the weights low, some flapping of the interface is allowed without over notification. Therefore, the weights can vary and are typically set using empirical data.
Table 2 shows an example of a completely filled in sliding window for an interface that, while having an "up" operational state as far as the router is concerned, is dropping a considerable number of ICMP packets. Table 3 shows the relationship between the percentage for the polling period and the "total ratio" once the weights are applied. To arrive at the forty-five percent (45%) total ratio, we take all of the positions in the table into account. The position percentage is multiplied by the weight for all positions to arrive at the resulting percentage for all positions. The resulting percentages are then added and divided by the sum of the weights. Given this total percentage, the final state of the interface is computed. Referring to Table 4, if the percentage is greater than sixty percent (60%), the interface is considered "up". If the percentage is between forty percent (40%) and sixty percent (60%), the state is either intermittent or unknown. However, it is unknown if and only if the last SNMP poll came back as "unknown"; otherwise, it is intermittent. If the ratio is less than forty percent (40%), the interface is "down".
It can be appreciated at this point that a hub and remote each comprise software executing on hardware. The hardware comprises one or more conventional computers and associated peripherals and communications interfaces. The dNMS kernel is a software engine executable on a computer that is integral to a hub or a remote. Preferably, the engine is never modified; instead, for flexibility and scalability, the invention employs "plug-ins" to implement specific functions. A "plug-in" as the term is used herein is a software module that carries a unique file name. Additionally, the only information that need be changed in the dNMS kernel is the configuration information that controls the functioning of a plug-in service, such as LTP described above. The dNMS kernel sends the configuration information, such as device addresses and how often a plug-in should perform a specified function on one or more devices, to the plug-ins and the monolithic server, and the monolithic server monitors the network devices based on the configuration information acquired by the plug-ins.
Monolithic server processing according to the invention can be summarized in terms of nodes (e.g., routers, servers, or topological containers for the same), interfaces (e.g., physical interfaces, IP addresses), services and notifiers. While nodes and interfaces have states, neither a node nor an interface knows how to determine its own state.
Nodes and interfaces only have states because they are associated with services that have a state.
Therefore, state information is derived from services; namely, an action performed on a node or interface that returns information. A service has a state by definition and is the only object that determines state on its own. An example of a service, as described above, is LTP.
In the present invention, a notifier is a plug-in that routes state information to another service, such as scheduler 90 in FIG. 8. If a service has determined that a change of state has to talcen place, a notifier is called. Therefore, a notifier is called when the state of a service is changed. In contrast, states of interfaces and nodes are determined by their owned services, but a notifier is not called when the state of an interface or node changes.
Note, however, that generally speaking the state change of a service will cause a change of state for the corresponding interface or node.
Note, however, that the state of an interface is defined as the worst state of any of its services, and that the state of a node is defined as the worst state of its interfaces, sub-nodes, and services. This means that a state change of a node or an interface is dictated by a downstream state change, which may not represent all objects on that node or interface.
Accordingly, to manage the amount of notifications resulting from state changes on a node or interface, the present invention employs a "toggle notification flag"
associated with nodes and interfaces. By setting the flag, an object will be ignored in an upstream state determination.
For example, if a node contains multiple interfaces, the state of one or more of the interfaces can be ignored for purposes of determining the state of the node. Notifiers are not called for interfaces or nodes who have their "toggle notification flag" set.
Referring now to FIG. 11 and FIG. 12, the preferred embodiment of the lower level architecture of dNMS kernel 88 is shown. At the outset, it should be noted that this architecture is common to all dNMS kernels, whether they reside in a hub or a remote. In FIG. 11, the architecture of the integration server is shown, while the architecture of the monolithic server is shown in FIG. 12. Note that the basic architecture is the same; however, the functions are different.
A primary function of integration server 112 is to manage the configuration information for the network it is configured to represent, such as network 24.
An integration server includes "placeholders" for each of the plug-in services, with each placeholder having a unique name that corresponds to the plug-in service that monitors the network. These placeholders are not operational services, however; they only represent configuration information that is passed to operational plug-ins located in monolithic server 114. The integration server manages this configuration information since it is connected to other integration servers in other dNMS kernels and, as discussed previously, configuration information propagates bidirectionally through the system. Therefore, the integration servers manage and route the configurations of all of the monitoring and collection services for the distributed network management system of the invention.
The monolithic server shares the same architecture of the integration server as can be seen in FIG. 12. Here, however, the services are operational and determine the state of downstream objects on the network. Note that the numbers and types of services are not limited. One such service is LTP as described above. Other services include, but are not limited to, monitoring bandwidth thresholding, temperature, power supply status, disk space, and environmental conditions. The system may optionally include one or more utility modules, such as an auto discovery module that knows how a router works and can talk to router to automatically add interfaces. Essentially, any software module that is not in the dNMS kernel itself can be "plugged-in" to the dNMS kernel to provide a service.
As indicated previously, each service has a unique identification (e.g., service or file name). Referring to FIG. 13, these identifiers permit the integration server and monolithic server to communicate through a conduit 116, which is an internal bus or other communications link. This allows state information from the monolithic server to be propagated to the corresponding service placeholder in the integration server for further propagation to another dNMS kernel. It also allows for configuration information to be propagated from the integration server to the monolithic server, whether the configuration information originates from the same or a different dNMS kernel (e.g., from the hub or remote in which the dNMS kernel resides, or from another hub or remote).
It will be appreciated that assigning a unique identifier to every service also allows for dNMS kernel to dNMS kernel communication. In addition to every service having a unique identifier, each identifier has a relative timestamp that denotes the last time that the service was changed. For example, when a "change" message such as an "add service"
message is transmitted it would indicate that the change was made one-thousand (1000) seconds ago.
This helps resolve time-based synchronization problems.
Note also that every attribute type for the various objects has a change message type, such as polling rate, node name, etc. The reason fox the time stamping is that, if two changes for the same attribute of the object are received, the most recent is used.
More simply, if a more recent type change is received than what is currently recorded, the more recent information is kept instead. Note that the sender of the change does not care how the recipient handles the message, only that it was received.
Referring to FIG. 14, the invention also includes a mechanism to control traffic between hubs and remotes. Each time a change message is sent, it is placed into a queue. For example, primary remote 16 sends a message to primary hub 12 through queue 118, and messages from primary hub 12 to primary remote 16 are sent through queue 120.
The message is then sent to the appropriate recipient. When the recipient acknowledges receipt, the message is dropped out of the queue. If the recipient does not have sufficient storage to accept the message, it will not send an acknowledgement. In that event, the message will stay in the queue indefinitely until an acknowledgment is received. For example, a remote could keep the message in the queue and not take the message until it has room to receive the message. Note that there are two reasons for a hub or remote to send a change message;
when that hub or remote generates the change message, or when propagating a change message for another hub or remote. An example would be where a secondary remote generates a change message. The secondary remote would send it to the primary remote and, in turn the primary remote would propagate it up to a hub.
The use of queues and acknowledgement controls will also keep the hubs from becoming overloaded when all or a part of the system returns from a system failure. Suppose, for example, that a secondary hub comes on line after a failure and thinks that it last received change information from the primary hub thirty (30) seconds ago. Also assume that the primary hub thinks that it last spoke to the secondary hub twelve-hundred (1200) seconds ago. In this instance, the primary hub would send a batch change representing a list of all changes in the past twelve-hundred (1200) seconds to the secondary hub, since that is the oldest timestamp. This can occur in either direction. The queues exist to accommodate batch transactions, rather than real-time transactions.
Another aspect of the invention involves knowing if a peer is operational; for example, a primary hub knowing that its corresponding secondary hub is operational and vice versa. In the present invention, this is not determined simply by testing connectivity. Here, all systems connected to each other send "keep alive" signals at specified intervals and look for "keep alive" signals from their peers at specified intervals. For example, every forty (40) seconds a "keep alive" signal is sent from the primary hub to the secondary hub. If a "keep alive" signal is not received by the secondary hub within one-hundred and eighty (180) seconds, the primary hub is considered to be down. Additionally, if a system tries to communicate with its peer, but cannot, the peer is deemed to be down. Other polling periods could be used, but the foregoing empirically have been found to provide the best results.

Also, with regard to the anatomy of a message, each message includes a unique identifier, a timestamp, a change type (e.g., node add, node remove, IP
address); message ID, and information specific to the change type (e.g., node name or IP address).
To prevent looping in the system, each time a system sends a message it puts a host name in the message and will never send a message to a system whose name is already in the message.
Lastly, it will be appreciated by those skilled in the art that a possible system configuration might involve monitoring a plurality of devices through one physical cable to all devices. In the event that the cable becomes inoperational, each of those devices may be reported as being inoperational. To reduce the need for "redundant" reporting of multiple devices experiencing an outage when the outage is due to a cable or other common device being inoperational, we can collate all devices into one and simply report that the common interface is inoperational.
Although the description above contains many specificities, these should not be construed as limiting the scope of the invention but as merely providing illustrations of some of the presently preferred embodiments of this invention. Thus the scope of this invention should be determined by the appended claims and their legal equivalents.
Therefore, it will be appreciated that the scope of the present invention fully encompasses other embodiments which may become obvious to those skilled in the art, and that the scope of the present invention is accordingly to be limited by nothing other than the appended claims, in which reference to an element in the singular is not intended to mean "one and only one" unless explicitly so stated, but rather "one or more." All structural, chemical, and functional equivalents to the elements of the above-described preferred embodiment that are known to those of ordinary skill in the art are expressly incorporated herein by reference and are intended to be encompassed by the present claims. Moreover, it is not necessary for a device or method to address each and every problem sought to be solved by the present invention, for it to be encompassed by the present claims. Furthermore, no element, component, or method step in the present disclosure is intended to be dedicated to the public regardless of whether the element, component, or method step is explicitly recited in the claims. No claim element herein is to be construed under the provisions of 35 LJ.S.C. 112, sixth paragraph, unless the element is expressly recited using the phrase "means for."

Table 1 Examples for filling out one entry in the LTP sliding window SINGLE DESCRIPTION ICMP SNMP RESULTING
OF

ROW FROM SITUATION PERCENTAGE STATUS PERCENTAGE

LTP RECEIVED FOR TIME

VIEWER PERIOD

-4 min up normal up 100% up (lx) 100%

(100%) interface, passing traffic (100%
ICMP

x1 SNMP = 100%) -4 min down normal down 0% down (Ox)0%
(0%) interface, not passing anything (0% ICMP x SNMP = 100%) -4 min up major packet 40% up (lx) 40%
(40%) loss to interface, but interface is still up (40% ICMP
x 1 SNMP = 40%) -4 min sump- interface 90% unknown 36%
(36%) passing unknown most traffic, (no but problem gathering response) sump info (.4x) (likely an sump-renumber issue) (90%
ICMP

x .4 SNMP=36%) -4 min down routing problem60% down (Ox)0%
(0%) causing pings to go through anyway, even through interface is down (or, an snmp-renumber issue) (60% ICMP
x 0 SNMP= 0%) -4 min undefinednormal pings 70% - 70%
on an (100%) interface with no SNMP (web server, etc.), (70%

ICMP=70%) -4 min up sump-only - up (lx) 100%
(-) monitoring of un-numbered interface, no ICMP status at all ( 1 SNMP
=

100%) Table 2 Example output for entire window of data TIME PERIOD PERCENTAGE SNMP STATE WEIGHT

-4 min 33% up 2x -3 min 33% up 2x -2 min 0% up 3x -1 min 100% up 3x 0 min 50% up 4x Total ratio calculation for LTP view in Table 2 PERCENTAGE RECEIVED FOR TIME PERIOD WEIGHT RESULTING
PERCENTAGE

33% 2x +66%

33% 2x +66%

0% 3x +0%

100% 3x +300%

50% 4x +200%

632% / 14 = 45%

Table 4 Mapping of total ratio percentage to final state of LTP
TOTAL RATIO LEVEL RESULTING STATE

ratio < 40 down 40 < ratio < 60 unknown or intermittent ratio > 60 up

Claims (31)

What is claimed is:
1. A distributed network management system, comprising:
(a) a hub server; and (b) a remote server;
(c) said remote server capable of communicating with a network device and said hub;
(d) said hub server capable of communicating with said remote server and said network device;
(e) wherein (i) if said hub server and said remote server are operational, said hub server communicates with said remote server, and (ii) if said hub server is operational and said remote server is inoperational, said hub server communicates with said network device.
2. A distributed network management system, comprising:
(a) a primary hub server;
(b) a secondary hub server; and (c) a remote server;
(d) said remote server capable of communicating with a network device, said primary hub server and said secondary hub server;
(e) said primary hub server capable of communicating with said remote server and said secondary hub server;
(f) said secondary hub server capable of communicating with said remote server and said primary hub server;
(g) wherein (i) if said primary hub server and said remote server are operational, said primary hub server communicates with said remote server, and (ii) if said primary hub server is inoperational, said secondary hub server is operational, and said remote server is operational, said secondary hub server communicates with said remote server.
3. A system as recited in claim 2, wherein said primary hub server is capable of communicating with said network device, and wherein if said primary hub server is operational and said remote server is inoperational, said primary hub server communicates with said network device.
4. A system as recited in claim 2, wherein said secondary hub server is capable of communicating with said network device, and wherein if said primary hub server is inoperational and said remote server is inoperational, said secondary hub server communicates with said network device.
5. A distributed network management system, comprising:
(a) a hub server;
(b) a primary remote server; and (c) a secondary remote server;
(e) said primary remote server capable of communicating with a remote network, said secondary remote server, and said hub server;
(f) said secondary remote server capable of communicating with said remote network, said primary remote server, and said hub server;
(g) said hub server capable of communicating with said primary remote server and said secondary remote server;
(h) wherein (i) if said hub server and said primary remote server are operational, said hub server communicates with said primary remote server, and (ii) if said hub server is operational, said primary remote server is inoperational, and said secondary remote server is operational, said hub server communicates with said secondary remote server.
6. A system as recited in claim 5, wherein said hub server is capable of communicating with said network, and wherein if said hub server is operational and said primary and said secondary remote servers are inoperational, said hub server communicates with said network device.
7. A distributed network management system, comprising:
(a) a primary hub server;
(b) a secondary hub server;
(c) a primary remote server; and (d) a secondary remote server;
(e) said primary remote server capable of communicating with a remote network, said secondary remote server, said primary hub server and said secondary hub server;
(f) said secondary remote server capable of communicating with said remote network, said primary remote server, said primary hub server and said secondary hub server;
(g) said primary hub server capable of communicating with said secondary hub server, said primary remote server, said secondary remote server, and said remote network;
(h) said secondary hub server capable of communicating with said primary hub server, said primary remote server, said secondary remote server, and said remote network;
(i) wherein (i) if said primary hub server and said primary remote server are operational, said primary hub server communicates with said primary remote server, (ii) if said primary hub server is operational, said primary remote server is inoperational, and said secondary remote server is operational, said primary hub server communicates with said secondary remote server, (iii) if said primary hub server is operational and said primary and secondary remote servers are inoperational, said primary hub server communicates with said remote network, (iv) if said primary hub server is inoperational, said secondary hub server is operational, and said primary remote server is operational, said secondary hub server communicates with said primary remote server, (v) if said primary hub server is inoperational, said secondary hub server is operational, said primary remote server is inoperational, and said secondary remote server is operational, said secondary hub server communicates with said secondary remote server, (vi) if said primary hub server is inoperational, said secondary hub server is operational, and said primary and secondary remote servers are inoperational, said secondary hub server communicates with said remote network, and (vii) if said primary hub server is operational, said secondary hub server is operational, and said primary and secondary remote servers are inoperational, said primary hub server and said secondary hub server communicate with said remote network.
8. A distributed network management system, comprising:

(a) a hub server;
(b) a remote server;
(c) said remote server capable of communicating with a network device and said hub;
(d) said hub server capable of communicating with said remote server and said network device; and (e) programming associated with at least one of said servers for carrying out the operations of (i) if said hub server and said remote server axe operational, causing said hub server to communicate with said remote server, and (ii) if said hub server is operational and said remote server is inoperational, causing said hub server to communicate with said network device.
9. A distributed network management system, comprising:
(a) a primary hub server;
(b) a secondary hub server;
(c) a remote server;
(d) said remote server capable of communicating with a network device, said primary hub server and said secondary hub server;
(e) said primary hub server capable of communicating with said remote server and said secondary hub server;
(f) said secondary hub server capable of communicating with said remote server and said primary hub server; and (g) programming associated with at least one of said servers for carrying out the operations of (i) if said primary hub server and said remote server are operational, causing said primary hub server to communicate with said remote server, and (ii) if said primary hub server is inoperational, said secondary hub server is operational, and said remote server is operational, causing said secondary hub server to communicate with said remote server.
10. A system as recited in claim 9, wherein said primary hub server is capable of communicating with said network device, and further comprising programming for carrying out the operation of causing said primary hub server to communicate with said network device if said primary hub server is operational and said remote server is inoperational.
11. A system as recited in claim 9, wherein said secondary hub server is capable of communicating with said network device, and further comprising programming for carrying out the operation of causing said secondary hub server to communicate with said network device if said primary hub server is inoperational and said remote server is inoperational.
12. A distributed network management system, comprising:
(a) a hub server;
(b) a primary remote server;
(c) a secondary remote server;
(e) said primary remote server capable of communicating with a remote network, said secondary remote server, and said hub server;
(f) said secondary remote server capable of communicating with said remote network, said primary remote server, and said hub server;
(g) said hub server capable of communicating with said primary remote server and said secondary remote server; and (h) programming associated with at least one of said servers for carrying out the operations of (i) if said hub server and said primary remote server are operational, causing said hub server to communicate with said primary remote server, and (ii) if said hub server is operational, said primary remote server is inoperational, and said secondary remote server is operational, causing said hub server to communicate with said secondary remote server.
13. A system as recited in claim 12, wherein said hub server is capable of communicating with said network, and further comprising programming for carrying out the operation of causing said hub server to communicate with said network device if said hub server is operational and said primary and said secondary remote servers are inoperational.
14. A distributed network management system, comprising:
(a) a primary hub server;
(b) a secondary hub server;
(c) a primary remote server;
(d) a secondary remote server;
(e) said primary remote server capable of communicating with a remote network, said secondary remote server, said primary hub server and said secondary hub server;
(f) said secondary remote server capable of communicating with said remote network, said primary remote server, said primary hub server and said secondary hub server;
(g) said primary hub server capable of communicating with said secondary hub server, said primary remote server, said secondary remote server, and said remote network;
(h) said secondary hub server capable of communicating with said primary hub server, said primary remote server, said secondary remote server, and said remote network;
and (i) programming associated with at least one of said servers for carrying out the operations of (i) if said primary hub server and said primary remote server are operational, causing said primary hub server to communicate with said primary remote server, (ii) if said primary hub server is operational, said primary remote server is inoperational, and said secondary remote server is operational, causing said primary hub server to communicate with said secondary remote server, (iii) if said primary hub server is operational and said primary and secondary remote sewers are inoperational, causing said primary hub server to communicate with said remote network, (iv) if said primary hub server is inoperational, said secondary hub server is operational, and said primary remote server is operational, causing said secondary hub server to communicate with said primary remote server, (v) if said primary hub server is inoperational, said secondary hub server is operational, said primary remote server is inoperational, and said secondary remote server is operational, causing said secondary hub server to communicate with said secondary remote server, (vi) if said primary hub server is inoperational, said secondary hub server is operational, and said primary and secondary remote servers are inoperational, causing said secondary hub server to communicate with said remote network, and (vii) if said primary hub server is operational, said secondary hub server is operational, and said primary and secondary remote servers are inoperational, causing said primary hub server and said secondary hub server to communicate with said remote network.
15. A distributed network management system, comprising:
(a) a hub server; and (b) a remote server;
(c) said remote server capable of communicating with a network device and said hub server;
(d) wherein configuration parameters for said remote server to communicate with said network device can be propagated between said hub server and said remote server bidirectionally.
16. A distributed network management system, comprising:
(a) a network server capable of communicating with a network device; and (b) means associated with said network server for deriving state information from said network device using LTP.
17. A system as recited in claim 16, wherein said LTP comprises:
(a) defining a polling interval;
(b) sending, from an ICMP server, a plurality of pings to an interface address on said network device during said polling interval;
(c) monitoring the number of pings returned from said network device and converting said number to a percentage based on the number of pings sent;
(d) sending an SNMP query to said network device and determining operational status of said network device from said SNMP query, said operational status comprising "up", "down", and "unknown";
(e) using the percentage of pings returned and the SNMP status, generating a status percentage for the polling period by multiplying the percentage pings returned by a constant value associated with said operational status, said constant value comprising a first value if the operational status is "up", a second value if the operational status is down", and a third value if the operational status is "unknown"; and (f) computing a weighted average of the status percentages for current and previous four polling periods and determining the state of the network device from the weighted average.
18. A system as recited in claim 16, further comprising:
(a) means for defining a polling interval;
(b) means for sending, from an ICMP server, a plurality of pings to an interface address on said network device during said polling interval;
(c) means for monitoring the number of pings returned from said network device and converting said number to a percentage based on the number of pings sent;
(d) means for sending an SNMP query to said network device and determining operational status of said network device from said SNMP query, said operational status comprising "up", "down", and "unknown";
(e) means for using the percentage of pings returned and the SNMP status, generating a status percentage for the polling period by multiplying the percentage pings returned by a constant value associated with said operational status, said constant value comprising a first value if the operational status is "up", a second value if the operational status is down", and a third value if the operational status is "unknown"; and (f) means for computing a weighted average of the status percentages for current and previous four polling periods and determining the state of the network device from the weighted average.
19. A system as recited in claim 16, further comprising programming associated with said network server for carrying out the functions of:
(a) defining a polling interval;
(b) sending, from an ICMP server, a plurality of pings to an interface address on said network device during said polling interval;
(c) monitoring the number of pings returned from said network device and converting said number to a percentage based on the number of pings sent;
(d) sending an SNMP query to said network device and determining operational status of said network device from said SNMP query, said operational status comprising "up", "down", and "unknown";
(e) using the percentage of pings returned and the SNMP status, generating a status percentage for the polling period by multiplying the percentage pings returned by a constant value associated with said operational status, said constant value comprising a first value if the operational status is "up", a second value if the operational status is down", and a third value if the operational status is "unknown"; and (f) computing a weighted average of the status percentages for current and previous four polling periods and determining the state of the network device from the weighted average.
20. A system for deriving state information from a network device, comprising:
(a) a computer; and (b) programming associated with said computer for carrying out the operations of (i) defining a polling interval;
(ii) sending, from an ICMP server, a plurality of pings to an interface address on said network device during said polling interval;
(iii) monitoring the number of pings returned from said network device and converting said number to a percentage based on the number of pings sent;
(iv) sending an SNMP query to said network device and determining operational status of said network device from said SNMP query, said operational status comprising "up", "down", and "unknown";
(v) using the percentage of pings returned and the SNMP status, generating a status percentage for the polling period by multiplying the percentage pings returned by a constant value associated with said operational status, said constant value comprising a first value if the operational status is "up", a second value if the operational status is down", and a third value if the operational status is "unknown"; and (vi) computing a weighted average of the status percentages for current and previous four polling periods and determining the state of the network device from the weighted average.
21. A method for distributed network management, comprising:
(a) providing a hub server;
(b) providing a remote server;
(c) said remote server capable of communicating with a network device and said hub;
(d) said hub server capable of communicating with said remote server and said network device;
(e) if said hub server and said remote server are operational, causing said hub server to communicate with said remote server; and (f) if said hub server is operational and said remote server is inoperational, causing said hub server to communicate with said network device.
22. A method for distributed network management, comprising:
(a) providing a primary hub server;
(b) providing a secondary hub server;
(c) providing a remote server;

(d) said remote server capable of communicating with a network device, said primary hub server and said secondary hub server;
(e) said primary hub server capable of communicating with said remote server and said secondary hub server;
(f) said secondary hub server capable of communicating with said remote server and said primary hub server;
(g) if said primary hub server and said remote server are operational, causing said primary hub server to communicate with said remote server, and (h) if said primary hub server is inoperational, said secondary hub server is operational, and said remote server is operational, causing said secondary hub server to communicate with said remote server.
23. A system as recited in claim 22, wherein said primary hub server is capable of communicating with said network device, and further comprising causing said primary hub server to communicate with said network device if said primary hub server is operational and said remote server is inoperational.
24. A system as recited in claim 22, wherein said secondary hub server is capable of communicating with said network device, and further comprising causing said secondary hub server to communicate with said network device if said primary hub server is inoperational and said remote server is inoperational.
25. A method for distributed network management, comprising:
(a) providing a hub server;
(b) providing a primary remote server;
(c) providing a secondary remote server;
(e) said primary remote server capable of communicating with a remote network, said secondary remote server, and said hub server;
(f) said secondary remote server capable of communicating with said remote network, said primary remote server, and said hub server;
(g) said hub server capable of communicating with said primary remote server and said secondary remote server;

(h) if said hub server and said primary remote server are operational, causing said hub server to communicate with said primary remote server; and (i) if said hub server is operational, said primary remote server is inoperational, and said secondary remote server is operational, causing said hub server to communicate with said secondary remote server.
26. A method as recited in claim 25, wherein said hub server is capable of communicating with said network, and further comprising causing said hub server to communicate with said network device if said hub server is operational and said primary and said secondary remote servers are inoperational.
27. A method for distributed network management, comprising:
(a) providing a primary hub server;
(b) providing a secondary hub server;
(c) providing a primary remote server;
(d) providing a secondary remote server;
(e) said primary remote server capable of communicating with a remote network, said secondary remote server, said primary hub server and said secondary hub server;
(f) said secondary remote server capable of communicating with said remote network, said primary remote server, said primary hub server and said secondary hub server;
(g) said primary hub server capable of communicating with said secondary hub server, said primary remote server, said secondary remote server, and said remote network;
(h) said secondary hub server capable of communicating with said primary hub server, said primary remote server, said secondary remote server, and said remote network;
(i) if said primary hub server and said primary remote server are operational, causing said primary hub server to communicate with said primary remote server;
(j) if said primary hub server is operational, said primary remote server is inoperational, and said secondary remote server is operational, causing said primary hub server communicates with said secondary remote server;
(k) if said primary hub server is operational and said primary and secondary remote servers are inoperational, causing said primary hub server to communicate with said remote network;

(l) if said primary hub server is inoperational, said secondary hub server is operational, and said primary remote server is operational, causing said secondary hub server to communicates with said primary remote server;
(m) if said primary hub server is inoperational, said secondary hub server is operational, said primary remote server is inoperational, and said secondary remote server is operational, causing said secondary hub server to communicate with said secondary remote server;
(n) if said primary hub server is inoperational, said secondary hub server is operational, and said primary and secondary remote servers are inoperational, causing said secondary hub server communicates with said remote network; and (o) if said primary hub server is operational, said secondary hub server is operational, and said primary and secondary remote servers are inoperational, causing said primary hub server and said secondary hub server to communicate with said remote network.
28. A method for distributed network management, comprising:
(a) providing a hub server;
(b) providing a remote server;
(c) said remote server capable of communicating with a network device and said hub server; and (d) propagating configuration parameters for said remote server to communicate with said network device between said hub server and said remote server bidirectionally.
29. A method for distributed network management, comprising:
(a) providing a network server capable of communicating with a network device;
and (b) deriving state information from said network device using LTP.
30. A method as recited in claim 29, wherein said LTP comprises:
(a) defining a polling interval;
(b) sending, from an ICMP server, a plurality of pings to an interface address on said network device during said polling interval;

(c) monitoring the number of pings returned from said network device and converting said number to a percentage based on the number of pings sent;
(d) sending an SNMP query to said network device and determining operational status of said network device from said SNMP query, said operational status comprising "up", "down", and "unknown";
(e) using the percentage of pings returned and the SNMP status, generating a status percentage for the polling period by multiplying the percentage pings returned by a constant value associated with said operational status, said constant value comprising a first value if the operational status is "up", a second value if the operational status is down", and a third value if the operational status is "unknown"; and (f) computing a weighted average of the status percentages for current and previous four polling periods and determining the state of the network device from the weighted average.
31. A method for deriving state information from a network device, comprising:
(a) defining a polling interval;
(b) sending, from an ICMP server, a plurality of pings to an interface address on said network device during said polling interval;
(c) monitoring the number of pings returned from said network device and converting said number to a percentage based on the number of pings sent;
(d) sending an SNMP query to said network device and determining operational status of said network device from said SNMP query, said operational status comprising "up", "down", and "unknown";
(e) using the percentage of pings returned and the SNMP status, generating a status percentage for the polling period by multiplying the percentage pings returned by a constant value associated with said operational status, said constant value comprising a first value if the operational status is "up", a second value if the operational status is down", and a third value if the operational status is "unknown"; and (f) computing a weighted average of the status percentages for current and previous four polling periods and determining the state of the network device from the weighted average.
CA002388712A 2000-06-30 2000-08-29 Distributed network management system and method Abandoned CA2388712A1 (en)

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
US60918000A 2000-06-30 2000-06-30
US09/609,180 2000-06-30
PCT/US2000/023728 WO2002003211A1 (en) 2000-06-30 2000-08-29 Distributed network management system and method

Publications (1)

Publication Number Publication Date
CA2388712A1 true CA2388712A1 (en) 2002-01-10

Family

ID=24439673

Family Applications (1)

Application Number Title Priority Date Filing Date
CA002388712A Abandoned CA2388712A1 (en) 2000-06-30 2000-08-29 Distributed network management system and method

Country Status (7)

Country Link
US (1) US7490144B2 (en)
EP (1) EP1295208A1 (en)
JP (1) JP3593528B2 (en)
AU (1) AU2000270865A1 (en)
CA (1) CA2388712A1 (en)
MX (1) MXPA02004027A (en)
WO (1) WO2002003211A1 (en)

Families Citing this family (44)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6763372B1 (en) * 2000-07-06 2004-07-13 Nishant V. Dani Load balancing of chat servers based on gradients
US7379994B2 (en) * 2000-10-26 2008-05-27 Metilinx Aggregate system resource analysis including correlation matrix and metric-based analysis
US20020083163A1 (en) * 2000-10-26 2002-06-27 Metilinx Multi-platform optimization model
US7126920B2 (en) * 2001-08-08 2006-10-24 General Instrument Corporation Performance of lifetest using CMTS as a proxy
US7525923B2 (en) * 2002-06-28 2009-04-28 Ntt Docomo, Inc. Catprobe
US7729268B2 (en) * 2002-06-28 2010-06-01 Ntt Docomo, Inc. Method and apparatus for quality of service determination
US7203965B2 (en) * 2002-12-17 2007-04-10 Sony Corporation System and method for home network content protection and copy management
US8011015B2 (en) 2002-12-17 2011-08-30 Sony Corporation Content access in a media network environment
CN101778096A (en) * 2002-12-17 2010-07-14 索尼电影娱乐公司 Method and apparatus for access control in an overlapping multiserver network environment
CN101635625B (en) * 2002-12-17 2012-02-29 索尼电影娱乐公司 Method and apparatus for access control in an overlapping multiserver network environment
US20080259813A1 (en) * 2004-03-09 2008-10-23 Johnny Mikhael Matta Method and apparatus for quality of service determination
WO2005089235A2 (en) 2004-03-13 2005-09-29 Cluster Resources, Inc. System and method providing object messages in a compute environment
US20070266388A1 (en) 2004-06-18 2007-11-15 Cluster Resources, Inc. System and method for providing advanced reservations in a compute environment
US8176490B1 (en) 2004-08-20 2012-05-08 Adaptive Computing Enterprises, Inc. System and method of interfacing a workload manager and scheduler with an identity manager
US20060168170A1 (en) * 2004-10-25 2006-07-27 Korzeniowski Richard W System and method for analyzing information relating to network devices
US7408441B2 (en) * 2004-10-25 2008-08-05 Electronic Data Systems Corporation System and method for analyzing user-generated event information and message information from network devices
US7408440B2 (en) * 2004-10-25 2008-08-05 Electronics Data Systems Corporation System and method for analyzing message information from diverse network devices
WO2006053093A2 (en) 2004-11-08 2006-05-18 Cluster Resources, Inc. System and method of providing system jobs within a compute environment
US8863143B2 (en) * 2006-03-16 2014-10-14 Adaptive Computing Enterprises, Inc. System and method for managing a hybrid compute environment
US9075657B2 (en) 2005-04-07 2015-07-07 Adaptive Computing Enterprises, Inc. On-demand access to compute resources
US9231886B2 (en) 2005-03-16 2016-01-05 Adaptive Computing Enterprises, Inc. Simple integration of an on-demand compute environment
US7353034B2 (en) 2005-04-04 2008-04-01 X One, Inc. Location sharing and tracking using mobile phones or other wireless devices
US9282081B2 (en) 2005-07-28 2016-03-08 Vaporstream Incorporated Reduced traceability electronic message system and method
US7610345B2 (en) 2005-07-28 2009-10-27 Vaporstream Incorporated Reduced traceability electronic message system and method
US8797870B2 (en) * 2006-01-05 2014-08-05 Cisco Technology, Inc. Method and system for calculation of QOV metrics
BRPI0621898A2 (en) * 2006-07-28 2012-10-16 Zte Corp distributed master and auxiliary administration method and network element based system
US8661101B2 (en) * 2007-06-01 2014-02-25 Avaya Inc. Method of IP address de-aliasing
US8392923B2 (en) * 2007-09-28 2013-03-05 International Business Machines Corporation Dynamic polling for completion of asynchronous tasks based on historical task completion times
US7548545B1 (en) 2007-12-14 2009-06-16 Raptor Networks Technology, Inc. Disaggregated network management
US8658576B1 (en) 2009-10-21 2014-02-25 Encore Wire Corporation System, composition and method of application of same for reducing the coefficient of friction and required pulling force during installation of wire or cable
US10877695B2 (en) 2009-10-30 2020-12-29 Iii Holdings 2, Llc Memcached server functionality in a cluster of data processing nodes
US11720290B2 (en) 2009-10-30 2023-08-08 Iii Holdings 2, Llc Memcached server functionality in a cluster of data processing nodes
US8422358B2 (en) * 2010-11-23 2013-04-16 International Business Machines Corporation Best-path evaluation based on reliability of network interface layers
US20120290707A1 (en) * 2011-05-10 2012-11-15 Monolith Technology Services, Inc. System and method for unified polling of networked devices and services
US9352371B1 (en) 2012-02-13 2016-05-31 Encore Wire Corporation Method of manufacture of electrical wire and cable having a reduced coefficient of friction and required pulling force
US11328843B1 (en) 2012-09-10 2022-05-10 Encore Wire Corporation Method of manufacture of electrical wire and cable having a reduced coefficient of friction and required pulling force
US9344908B2 (en) * 2012-12-13 2016-05-17 Devicescape Software, Inc. Systems and methods for quality of experience measurement and wireless network recommendation
US10056742B1 (en) 2013-03-15 2018-08-21 Encore Wire Corporation System, method and apparatus for spray-on application of a wire pulling lubricant
US20150029892A1 (en) * 2013-07-26 2015-01-29 Ideaware Inc. Apparatus for detecting a periodicity, a method thereof and a recording medium thereof
US9525638B2 (en) 2013-10-15 2016-12-20 Internap Corporation Routing system for internet traffic
US9917728B2 (en) 2014-01-14 2018-03-13 Nant Holdings Ip, Llc Software-based fabric enablement
US10212101B2 (en) 2014-01-14 2019-02-19 Nant Holdings Ip, Llc Low level provisioning of network fabrics
US10826796B2 (en) 2016-09-26 2020-11-03 PacketFabric, LLC Virtual circuits in cloud networks
JP7027074B2 (en) * 2017-09-04 2022-03-01 株式会社日立システムズ Device management system and device management method

Family Cites Families (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US5710885A (en) * 1995-11-28 1998-01-20 Ncr Corporation Network management system with improved node discovery and monitoring
US6173311B1 (en) * 1997-02-13 2001-01-09 Pointcast, Inc. Apparatus, method and article of manufacture for servicing client requests on a network
US5948108A (en) * 1997-06-12 1999-09-07 Tandem Computers, Incorporated Method and system for providing fault tolerant access between clients and a server
US6338092B1 (en) * 1998-09-24 2002-01-08 International Business Machines Corporation Method, system and computer program for replicating data in a distributed computed environment
US6389459B1 (en) * 1998-12-09 2002-05-14 Ncr Corporation Virtualized storage devices for network disk mirroring applications
US6453468B1 (en) * 1999-06-30 2002-09-17 B-Hub, Inc. Methods for improving reliability while upgrading software programs in a clustered computer system
US6564336B1 (en) * 1999-12-29 2003-05-13 General Electric Company Fault tolerant database for picture archiving and communication systems
US7028083B2 (en) * 2000-05-26 2006-04-11 Akomai Technologies, Inc. Method for extending a network map

Also Published As

Publication number Publication date
MXPA02004027A (en) 2004-03-10
US20040210632A1 (en) 2004-10-21
JP3593528B2 (en) 2004-11-24
AU2000270865A1 (en) 2002-01-14
JP2004519113A (en) 2004-06-24
US7490144B2 (en) 2009-02-10
WO2002003211A1 (en) 2002-01-10
EP1295208A1 (en) 2003-03-26

Similar Documents

Publication Publication Date Title
US7490144B2 (en) Distributed network management system and method
US8099488B2 (en) Real-time monitoring of service agreements
US20060179059A1 (en) Cluster monitoring system with content-based event routing
US7099879B2 (en) Real-time monitoring of service performance through the use of relational database calculation clusters
US20030120764A1 (en) Real-time monitoring of services through aggregation view
US8489722B2 (en) System and method for providing quality of service in wide area messaging fabric
US6665262B1 (en) Distributed fault management architecture
US8601155B2 (en) Telemetry stream performance analysis and optimization
AU740146B2 (en) A telecommunications performance management system
CN106998263B (en) System and method for maintaining network service level
US20030039212A1 (en) Method and apparatus for the assessment and optimization of network traffic
US20020087696A1 (en) Automatic management systemfor communications networks
US20030046426A1 (en) Real time traffic engineering of data-networks
US20060069777A1 (en) Request message control method for using service and service providing system
US11296947B2 (en) SD-WAN device, system, and network
Kim et al. Efficacy of techniques for responsiveness in a wide-area publish/subscribe system
EP1548980A1 (en) A method of monitoring a network
Mansour et al. Load balancing in the presence of services in named-data networking
US20060179342A1 (en) Service aggregation in cluster monitoring system with content-based event routing
KR20080065284A (en) Independent message stores and message transport agents
Jerzak et al. Fail-aware publish/subscribe
Liang et al. Cost estimation queuing model for large-scale file delivery service
Hao et al. Distributed MQTT Brokers at Network Edges: A Study on Message Dissemination
Wang Management of Temporally and Spatially Correlated Failures in Federated Message Oriented Middleware for Resilient and QoS-Aware Messaging Services.
Abelsen Adaptive gridstat information flow mechanisms and management for power grid contingencies

Legal Events

Date Code Title Description
EEER Examination request
FZDE Discontinued