|Publication number||US5757640 A|
|Application number||US 08/590,933|
|Publication date||May 26, 1998|
|Filing date||Jan 24, 1996|
|Priority date||Jan 24, 1996|
|Publication number||08590933, 590933, US 5757640 A, US 5757640A, US-A-5757640, US5757640 A, US5757640A|
|Inventors||Robert J. Monson|
|Original Assignee||Ag-Chem Equipment Co., Inc.|
|Export Citation||BiBTeX, EndNote, RefMan|
|Patent Citations (33), Non-Patent Citations (6), Referenced by (36), Classifications (6), Legal Events (8)|
|External Links: USPTO, USPTO Assignment, Espacenet|
This invention relates to product application control systems. More particularly, this invention relates to a modular expandable product application control system having a distributed process manager using one or more distributed networks for use on vehicles either self-propelled or towed.
Present mobile product application control systems are generally designed for simple closed loop control of a multiple loop network or are otherwise directed generally toward a single open loop distributed network. At present, the maximum amount of loops that can be added to a multiple closed loop network typically would approach ten. Beyond this value, the overall packet structure and network control would become cumbersome, requiring significant redesign of the overall closed loop system.
A typical closed loop system is laid out with a host controller located within the cab of a vehicle, made up of a processor with associated input and output devices. This host controller is typically directly linked to a second controller located within the cab, which is responsible for all communication to devices on the chassis of the vehicle. Such a system is shown in U.S. Pat. No. 4,630,773 to Ortlip, issued Dec. 23, 1986 entitled "Method and Apparatus for Spreading Fertilizer." Another system is shown in U.S. Pat. No. 5,220,876 to Monson et al., issued Jun. 23, 1993 entitled "Variable Rate Application System" and assigned to the assignee of the present invention. Both the '773 patent and the '876 are incorporated herein by reference in their entirety.
In closed loop systems such as those stated above, the addition of loops as a result of further control requirements adds great complexity to the system wiring. This results in a realistic limit to the amount of loops that can be added to a single system. Beyond this limit, the amount of wiring becomes too great and routing the wiring is a considerable challenge. At present, it is felt that ten loops is the approximate limit for a single host system for the type described above.
Software construction also places many constraints on the design and application of a control system. The common top down design has certain flow characteristics that define a specific flow path within the operating system. This also limits the effective amount of devices that can be controlled with a single system, as eventually the communication link to the devices becomes the limiting factor.
The aforementioned software construction implies that any changes in the requirements, such as the inclusion of an additional loop on the vehicle, often causes sweeping changes within the code. This also has obvious trickle-down effects that are not desired within the flow structure of the program.
Thus, debugging and testing become required parts of the system maintenance, requiring large amounts of capital and time to assure a simple change has not caused greater errors elsewhere in the system (often occurring in totally unrelated places).
One solution to the many problems stated herein before was addressed in U.S. Pat. No. 5,453,924, to Monson et al., issued Sep. 26, 1995, entitled "Mobile Control System Responsive to Land Area Maps", and assigned to the assignee of the present invention. The '924 patent is incorporated by reference in its entirety herein. The inventive system disclosed and claimed by Monson et al. in the '924 patent effectively eliminated the requirement for multiple closed loops by utilizing a single open loop distributed network. This was accomplished with the use of modem hardware and software capabilities. In particular, the system utilized object oriented software design as well as windows interfacing, distributed network controls and Echelon LonWorks network technology.
For the aforementioned reasons, it becomes immediately apparent that existing closed loop product application control system designs are by no means desirable, and entirely new approaches are required. While the solution provided by Monson et al. in the '924 patent addressed many of the problems associated with closed loop product application control systems, a need still exists for a system architecture which, to a great extent, is independent of the media and network protocol utilized. The present invention described in detail hereinafter offers an additional solution to the stated problems.
The present inventive product application control system provides a novel method and apparatus which overcomes many of the shortcomings and attendant disadvantages of past control systems utilized by the agronomics industry. Several problems are considered unavoidable within the industry, some of which have been discussed previously. The present invention, however, overcomes many of these problems with a new approach to controls design.
Object-oriented design was developed within the software industry to create a system capable of expansion or retraction with very little effect on the system as a whole. The present invention utilizes this principle in a manner previously unknown or contemplated by those skilled in the art. Instead of expanding on the use of a single distributed network system, as known in the art, the present invention expands the control system into a multiple distributed network system, with any expansion having only a transparent effect on the system as a whole. As stated previously, this has been accomplished through the use of several state-of-the-art technologies including the aforementioned object-oriented software design, as well as windows interfacing, distributed network controls and modern network technology, e.g., Echelon™ LonWorks™ or CANs network technology.
Object-oriented software design is critical in an ever changing system such as that displayed by the present invention. With the use of objects, the software is translated from a top-down flowchart into a dynamic system that reacts to the world around it, in much the same manner one would expect the control system to operate. This is accomplished with a message-based system such as the windows interface which increases usability, but also allows a new approach to task completion.
The objects are designed in a fashion that allows ease of expansion and upgradability, easing maintenance in the future. Because of the packaging of objects, the trickle-down effect is removed from the system when upgrades are incorporated. Thus, the flow of the system is not so completely interrelated, and a change in one area will not cause unexpected problems elsewhere, as was the case in past closed loop systems.
The aforementioned inventive control system also greatly simplifies system testing since only the area of interest need be tested. Because the actual flow path of other areas is not disturbed in any way, it is not necessary to perform redundant and excessively costly testing of the complete system. Thus, system testing becomes an option and not a necessity.
Distributed network controls were selected for the present inventive system for much the same reason as object-oriented software. Use of a distributed network is intended to create the simplest method of network expansion or retraction without excessively burdening the system with overhead costs. In other words, with a standard closed loop system, it is easy to include ten control loops on every machine in order to create a common fleet of vehicles. Unfortunately, this means unnecessary costs are incurred on vehicles requiring only one loop. Thus, a distributed network control system has been developed which only allows required hardware as necessary components.
The control system design allows ease of expansion as discussed previously, but also brings a great enhancement in the area of manufacture. The present invention further incorporates a single cable which is run into the cab for each distinct network which is embodied on the vehicle. Each cable preferably contains a twisted pair of wires through which the network communicates with the host and a second pair to twisted wires providing power to the network. All other network wiring is external to the cab, itself, a great improvement over existing systems. Thus, the majority of the wiring has been removed beyond the cab to the location of the actuation and sensing. This means that complex, machine dependent wiring occurs only at the actuator or point of sensing and in a modular fashion.
Proper development of a multiple distributed network, also allows the node hardware to be identical from node to node, which implies that each machine control system is identical, with the only differences being in the number of nodes and the machine dependent wiring such as that displayed by the present invention.
Placing the node control at the location of the actuator reduces 1) attenuation of the signal, 2) the incorporation of noise 3) the packaging of loop-dependent controllers, and 4) continuation of the signal. Thus, the intelligence of the controllers is distributed throughout the vehicle. This also allows a great enhancement of the associated diagnostic system. With the control loop separated, and each node capable of self-sustained control, each loop can also be responsible for diagnostics, data-logging, communication and even the real generation of expert systems, if so desired, thereby allowing the main processor to simultaneously orchestrate command processing and control of more than a single distributed network.
The present inventive system additionally incorporates intelligent network technology such as Echelon™ LonWorks™ technology or CANs technology to support the aforementioned benefits. The system is capable of incorporating multiple intelligent networks. For example, the control system may include a product application control network, a vehicle location sensing network, a chassis control network and/or a product development sensing network for providing data relating to plant development. Each of the networks is capable of operation independently of the others. Further, each distributed network may communicate with the host using its own communication technology independent of that used by the other networks. This allows the use of Echelon™ LonWorks™ or CANs network technology, for example, which may be developed in the future.
In one embodiment the invention is a mobile product application control system which includes a vehicle and at least one distributed network coupled to the vehicle. The distributed network includes at least one independent control module located at a node within the distributed network and having independent processing capability. The intelligent control module controls at least one actuator device in response to control setpoints generated from objects processed within the control system. The actuators are controlled in a manner that causes at least one predetermined product to be applied to a predetermined geographic land area at variable rates determined by the control setpoints. The system includes a central processor coupled to the at least one distributed network. The central processor has a process distribution controller which monitors the level of processing activity of the central processor. It is the process distribution controller which determines whether objects are processed within the central processor or the independent control modules. The responsibility for processing objects in maintained within the central processor so long as the level of processing activity in the central processor does not exceed a predetermined target level. If that predetermined target level is exceeded the process distribution controller causes the responsibility for processing one or more objects to be downloaded to at least one independent control module. The objects are downloaded based upon a predetermined priority level. Preferably, objects with the lowest priority are downloaded first. The system further includes at least one data input device coupled to either the distributed network or the central processor. The data in used in processing the objects.
The independent networks may include a product application control network, a vehicle location sensing network, a chassis control network or a product developments sensing network. Further, the system of the present invention may be used in a way that provides a totally hands-free control system where independent control modules control actuators responsible for product application, vehicle speed, vehicle gear ratio and vehicle steering. Thus, operator involvement in hands-on control is kept to a minimum.
In a further embodiment the invention is a method of controlling application of a product to a predetermined geographic land area. The method comprises providing a vehicle having at least one input actuator device, a central processor and at least one independent sensing and control module. The vehicle is caused to travel over the surface of the geographic land area. Objects are processed within the control system to determine control setpoints. The objects are processed within the central processor so long as the level of processing activity in the central processor does not exceed a predetermined target level. If the predetermined target level is exceeded the responsibility for processing at least one object is downloaded to at least one independent processing and control module. The at lest one actuator device is controlled in response to the control setpoints generated from the objects processed within the control system such that at least one predetermined product is applied to the predetermined geographic land area at variable rates determined by the control setpoints. The method may include controlling actuators which control vehicle speed, vehicle gear ratio and vehicle steering based upon setpoints generated from the processed objects.
From the foregoing, it is clear the present inventive control system performance is greatly enhanced over existing systems. Other features of the present inventive product application control system include ease of use, manufacture, enhanced serviceability, maintainability, upgradability, and enhanced expansion and diagnostics capability.
Other objects and features of the present invention and many of the attendant advantages of the present invention will be readily appreciated as the same become better understood by reference to the detailed description when considered in connection with the accompanying drawings in which like reference numerals designate like parts throughout the figures thereof and wherein:
FIG. 1 is a simplified block diagram of a closed loop multiple loop control system for which reliance on processing resides entirely within a host, known to those skilled in the art.
FIG. 2 is a simplified block diagram of a known distributed network control system wherein significant processing takes place in the nodes.
FIG. 3 illustrates a simplified depiction of the control system of FIG. 2, where map and GPS data structures typically reside in the host, and a network carries only setpoint information to the node(s).
FIG. 4 is a simplified block diagram of one embodiment of a product application control system in accordance with the present invention where the GPS data structure(s) reside in a host, and map data structure(s) reside within the network node(s).
FIG. 5 is a simplified block diagram of another present product application control system, where the map data structure(s) and the GPS data structures reside entirely within the network node(s).
FIG. 6 is a simplified block diagram of yet another embodiment for the present product application control system, where the map data structure(s) reside in a host and the GPS data structure(s) reside entirely within the network node(s).
FIG. 7 is a simplified block diagram of another embodiment for the present product application control system, where the GPS data structure(s) reside entirely within the network node(s) and where the map data structure(s) reside in both a host as well as within the network node(s).
FIG. 8 is a simplified block diagram of a multiple network product application control system in accordance with the present invention.
FIG. 9 is a flow diagram for the host system tasks for operating the present product application control system shown in FIG. 8.
FIG. 10 is a flow diagram of a multiple network server start-up for operating the present product application control system shown in FIG. 8.
FIG. 11 is a flow diagram of a single network server for operating the present product application control system shown in FIG. 8.
FIG. 12 is a flow diagram of internal node control for operating the present product application control system shown in FIG. 8.
FIG. 13 is a flow diagram of a typical time cycle of the distributed processing system of the present invention.
FIG. 14 is a flow diagram of the distributed of the processing of objects between the host system and a single node.
Referring to FIG. 1, a presently known product application control system 100 includes a host controller 18 located within the cab of a vehicle (not shown), made up of a processor/computer 14 with associated input and output devices including monitor 10 and keyboard 12. The host controller 18 is responsible for all communication to devices such as actuators 20 including actuator feedback 24 to controller 18 and, for example, a radar device 22 on the chassis of the vehicle. As shown, and as previously discussed, the addition of loops 26 to the control requirements adds great complexity to the overall control system 100 wiring.
In FIG. 2, a known embodiment 200 for a product application control system similar to that described in the '924 patent to Monson et al. is shown. The system includes a processor/computer 14 with associated input and output devices including a monitor 10 and a keyboard 12, located within the cab of the vehicle (not shown). Computer 14 is connected to a single distributed network 210 via a single twisted pair of wires 204 which is run into the cab (not shown). All other system wiring is external to the cab.
The distributed network control system 200 includes multiple nodes 202, theoretically unlimited in number. It can thus be seen that the majority of the wiring displayed in the control system 100 illustrated in FIG. 1 has been moved beyond the cab to the location of the actuators 20 and sensing, feedback 24. It can be seen, therefore, as stated herein before and as specifically described in the '924 patent, that the complex common machine dependent wiring only occurs at the actuators 20 and in a modular fashion.
As stated herein before, node 202 hardware is generally identical from node 202 to node 202, thereby implying that each machine product application control system 200 is identical, with the only differences being in the number of nodes 202 and the machine (not shown) dependent wiring.
Keeping in mind the known art described above, the present inventive product application control system will now be discussed in detail with the aid of FIGS. 3-12 and following definitions.
Object-Oriented Software--software development using the object-oriented methodology which models the problem domain as a collection of real-world objects, and translates these objects into executable program code in a manner that promotes reusability, maintainability, and extendibility, among others.
Windows Interface--a graphical user interface extension of an operating system, built upon a message-based multi-tasking system, as opposed to a classical system built on a single thread of execution.
Distributed Network--a hardware design methodology that takes advantage of network communications and applies them in the control domain.
Echelon™ LonWorks™ Technology--a hardware system that fully exploits the distributed network concept, and greatly eases the more complex communication requirements that are brought about by the use of distributed processors.
CANs Technology--yet another hardware system that fully exploits the distributed network concept, and greatly eases the more complex communication requirements that are brought about by the use of distributed processors.
The system architecture basically refers to the overall structure and method of accomplishing a task in the environment of the network. That is to say, that once the network is in place, there are many methods and data structures for achieving the same end. If the network use is properly structured, the system design will, to a great extent, become independent of the media and network protocol used. Although Echelon™ network technology is presently preferred, the same structure as present utilizing Echelon™ network technology would transpose almost directly into similar network environments, e.g., CANs technology. In other words, the present multiple network control system may be used with any existing or contemplated network technology.
Preferably, the design of the present inventive multiple network control system creates a reliance on processing in the node(s) 202, by virtue of requiring complete PID capabilities and self-diagnostics, and processing capability within the host, to provide commands at predetermined intervals, e.g., one second, to each of the operating loops. This clear separation of processing allows the present inventive system to be created with very little reliance on the communication method between the host and particular network nodes. As stated herein before, Echelon network technology simplifies this communication, but other network technology, e.g., CANs could also be used.
FIGS. 3-7 are simplified block diagrams showing a host connected to a single network. These figures illustrate some of the alternatives which exist for computing and data storage capabilities of the system. Although a single network is shown the alternatives are equally applicable to a multiple network system.
FIG. 3 illustrates a simplified depiction of the control system 200 in FIG. 2, a known control system used in the agricultural industry for the application of agricultural products, where map 220 and GPS (global positioning system) 230 data structures typically reside in the host 201, and a network 210 carries only setpoint information to the node(s) 202. Each node may be connected to devices such as actuator 20 and feedback 24 as well as to additional nodes. Although control system 200 overcomes many of the attendant disadvantages inherent within known closed loop multiple loop control systems such as that illustrated in FIG. 1, it is clearly seen that such a system design continues to place a great burden on the host system. What is still needed within the agricultural industry as well as other industries which utilize self-propelled vehicle product application control systems, is even more flexibility to complete a desired task in several different manners with the same hardware, dependent only on the software and firmware (data structures) utilized. The present inventive control system will now be described in detail with reference to FIGS. 4-12.
Referring to FIG. 4, one preferred embodiment 400 for the present inventive product application control system includes a host 402 with associated input and output devices (e.g., 12, 14 located within the cab of the vehicle). Embodiment 400 includes a nodal control module 406 located at each node 404. Placing a nodal control module 406 at the location of each actuator device 20 provides for downloading a digital description of a land area 420, for example, thereby making the nodes 404 more independent than nodes 202 utilized in prior known distributed network product application control systems for use in conjunction with self-propelled or towed vehicles. In this embodiment, location data 430 from a GPS system is the only information communicated on the network 408 (along with status data). Therefore, this embodiment requires a great deal of memory in the node(s) 404. Each node 404 has associated with it devices including, for example, an actuator 20 and feedback 24 as described with respect to FIG. 2.
FIG. 5 is yet another preferred embodiment 500 for the present inventive product application control system. The system includes a host 502 with associated input and output devices, e.g., 12, 14 located within the cab of the vehicle. Embodiment 500 includes a nodal control module 506 located at each node 504. Each nodal control module 506 includes data structures at the location of each actuator device 20 which provide for downloading for both the map 520 and location data 530, for example, resulting in the host system 502 acting solely as a supervisor, with the nodes 504 acting in an autonomous fashion. It should be noted that GPS data 530 can be present on any node 504 within the network 508 with the same result. Such an embodiment of the present inventive control system requires the greatest processing capabilities for the nodes 504, as well as the greatest memory requirements.
A further embodiment 600 of the present inventive product application control system is shown in FIG. 6. The product application control system includes a host 602 with associated input and output devices, e.g., 12, 14 located within the cab of the vehicle. Embodiment 600 includes a nodal control module 606 located at each node 604. Each nodal control module 606 includes data storage transfer structures at the location of each actuator device 20 which provide for downloading network based GPS signals 630, for example, in order to simplify the main host 602 design. Map data 620 is stored in the host. Typically, an extra serial port is required for such a GPS input for the control module 606. This embodiment 600 is preferred for any network capable system which does not increase processing power within the nodes 604. Although this embodiment would not increase the processing power for any particular node 604, it would require the node 604 to transmit its location to the host 602 before receiving a command setpoint. Thus, this embodiment 600 could be somewhat slower, and would increase the network traffic over that expected with other embodiments of the present inventive system described above.
FIG. 7 shows a preferred embodiment 700 for distribution of data storage and computer capabilities for the present inventive control system. Embodiment 700 includes a nodal control module 706 located at each node 704. Each nodal control module 706 includes data storage and transfer structures at the location of each actuator device which provides for downloading network based GPS signals 730 similar to that described with respect to FIG. 6. In this embodiment map data storage and transfer structures reside in both a host 702 as well as within the network nodes 704. The distinction of embodiment 700 is, therefore, that map data 720, could reside to some extent in more than one location. As disclosed in U.S. Pat. No. 5,220,876 issued to Monson et al., there may be at least two maps present in the system. For example, one map may correspond to the digital description of the field and the other may describe how product application rates are related to the colors on the map. This can be a one dimensional map. In the case where the digital description of the field resides in the host, a look up table can reside in the node 704 with setpoint data. This results in network traffic being comprised of the host 702 telling the nodes 704 at what array location the machine is presently located. Embodiment 700 is advantageous in that communication in the network is minimized.
One of the advantages of downloading data and control responsibilities to the nodes is that the host is then free from the task of performing nodal control functions. By freeing the host from time critical activities of the control system it is possible to create a system capable of running several control/status networks simultaneously on a vehicle. The host is thus capable of monitoring and serving several independent networks. Such a multiple network control system is shown in FIG. 8. The control system includes two or more independent networks 810 and 910. Networks 810 and 910 communicate with a processor/computer 14 having associated input and output devices, e.g., monitor 10 and a keyboard 12 (not shown) through a network interface 50. Network interface 50 communicates with distributed networks 810 and 910 via cables 808 and 908, respectively. Cables 808 and 908 each include a twisted pair of conductors for carrying network data between network interface 50 and independent distributed networks 810 and 910 but the present invention is not so limited and other communication cables or methods of communication, including wireless, may be used within the scope of the invention. Each distributed network requires only one cable per network to enter the cab for full operation of each network and all devices. This is a great simplification from the traditional layout where dozens of wires enter the cab. Thus, troubleshooting and maintenance are greatly simplified and the network has provided a method of self-diagnostics by design. Obviously, if feedback is not received at the node, the problem lies between the node and device, rather than somewhere between the controller and the device.
Expansion of either distributed network 810 or 910 is easily accomplished by simply extending the network to a new node and providing the necessary node/device wiring. This allows the design to be completed in a completely modular fashion as each network remains independent of the others.
Each distributed network 810 and 910 includes multiple nodes 802 and 902, respectively. Each of the nodes has associated with it at least one actuator 20 and, preferably, a feedback device 24, although it will be appreciated that feedback is not always necessary. Each of the nodes 802 and 902 has a node controller which functions in a manner similar to those described in connection with FIGS. 4-7.
The operation of the product application control system can be understood generally with respect to FIGS. 9-12 which are flow diagrams of a preferred algorithm used to operate the various functional levels of the control system. FIG. 9 is a flow diagram of the system tasks performed by the host during operation. FIG. 10 is a flow diagram of the algorithm which operates the multiple network server start-up. FIG. 11 is a flow diagram of the operation of each network server. FIG. 12 is a flow diagram of internal node control operations. It will be understood that the control system operates simultaneously at each of these levels in a coordinated manner.
As seen in FIG. 9, the host operating system is booted up at step 920. An expert system is also started at this time. The expert system is the artificial intelligence device which processes all of the complex calculations associated with the control of product application. It is a system defmed by the use of one or more relationships between one or more factors that relate to crop production. The relationships may be mathematical, spatial, user defined or others; examples include recommendation equations for a given nutrient and crop as published by most state universities. Similar expert systems have been described in the '876 patent to Monson et al. and the '924 patent to Monson et al. The expert system of the present invention performs similar functions. It will be appreciated, however, that in the present multiple network system much of the responsibility for data storage and performance of calculations has been shifted downward to the node level. Therefore, the expert system as contemplated herein may include components within the host as well as the node controllers.
At step 921 an examination of network server applications is made. A determination is made at decision block 922 whether all servers are accounted for. If not, the algorithm loops back to step 921. When all servers are accounted for an examination of all data returned on the network is made at step 923. This data is then analyzed at step 924 to define inputs available, outputs required, and dependencies. The algorithm then proceeds, at step 925, to define setpoints for outputs required at each node based upon the spatial position of the vehicle. In the algorithm shown in FIG. 9, the calculation of setpoints at step 925 assumes that the map resides in the host. As previously discussed with respect to FIGS. 4-7 it is possible to download map information to the nodes or to include map information at both the host and node level. Therefore, depending upon the location of the map data the setpoint calculation may be performed in the host, at the node controller level or some combination thereof.
After the setpoints have been calculated a determination is made at decision block 926 whether independent feedback inputs exist. These independent feedback inputs may consist of the input from one or more sensors which sense data related to plant development. Such sensors may be used to calculate from soil samples taken periodically during the procedure certain characteristics such as nitrogen level, potassium level, phosphorous level, soil type, pH level, soil temperature, and herbicide, pesticide or fungicide levels, for example. These sensors may operate on a continual or intermittent basis to provide data on a real time basis which is used by the control system for updating the data used in product application control. This data may originally be comprised of pre-loaded maps which may then be updated in real time by the sensor data as the product application process proceeds. The information obtained from the sensors is analyzed for validity and, if valid, is utilized to update the existing data to provide the highest degree of accuracy at the present time. If the sensed data is invalid or inconclusive it is not used. In that case the system would utilize existing data. Therefore, even in the case of a sensor failure where no data is forthcoming the system would continue to be as accurate as the existing map data, which means, even in a worst case scenario, that the system is as accurate as those in use today.
If feedback is determined to exist at decision block 926 it is analyzed at step 929 to determine its impact and is communicated to the expert system. If no independent feedback inputs exist or if the feedback information has been analyzed and communicated to the expert system then an examination of network status is made at step 929. The status of the networks is communicated to the expert system at step 928 which then loops back to step 922 to start the process over again.
FIG. 10 is a flow diagram of the algorithm used for the multiple network server start-up. At step 930 the operating system is booted up. At decision block 931 a determination is made of the presence of the network. If the network is not sensed a display of system status is made at step 932. If the network is sensed network feedback is interrogated at step 933. A determination of network status from the feedback is made at step 934 and is displayed at step 935. At decision block 936 a determination is made whether all feedback is accounted for. If not, the algorithm loops back to decision block 931 to repeat the process until all feedback has been accounted for.
When all feedback is accounted for a determination is made at decision block 937 of the existence of multiple networks. If multiple networks are not sensed a single network server is begun at step 939. If multiple networks are sensed multiple independent single network servers are begun at step 938, one for each network on the system.
FIG. 11 is a flow diagram of the algorithm used for the single network server. Preliminary node configuration data is entered at step 940. The configuration data may be comprised of a short piece of code that is sent via the network to each specific node to configure node firmware for operation. The data defines pin assignments, constants, control technique, system variables and reporting requirements, among others. All executable software on the node is transmitted to the node by the host. The process effectively starts the node control software into operation.
At step 941 the status of the node(s) is examined. The nodes are initialized at step 942 according to the node configuration data received from the host. At decision block 943 a determination is made whether all nodes have been initialized. If not, the algorithm loops back to step 941 for the process to continue until the initialization of all nodes has been completed. When the node initialization process is complete the host system requirements are examined at step 944. It is at this step that the setpoints which were calculated at step 925 of the host system tasks algorithm (FIG. 9) are communicated to the network server. The calculated setpoints for the required outputs are utilized to defme the node requirements at step 945. The setpoints for the required outputs are communicated to the nodes at step 946. A determination is made at decision block 947 whether setpoint requirements have been communicated to all nodes. If not, the algorithm loops back to step 945 until setpoint communication to all nodes has been completed. Once completed, the system records pertinent network data. This may consist of updating maps or to otherwise provide a current status of system operation.
FIG. 12 is a flow diagram of the algorithm used for internal node control. Setpoint data is received from the host at step 949. This is the data that is communicated by the network server at step 946 of FIG. 11. The setpoint data is used to calculate new setpoint PID variables at step 950. The PID logic is programmed into the node firmware (software resident in the node non-volatile memory). This allows better control than centralized traditional control. Since the processor is located at the site of the control it is required only to maintain the particular node setpoint to which it has been assigned.
At step 951 the actuator is activated in response to host system setpoints. Loop feedback is interrogated at step 952 and the status of the node is determined from the feedback at step 953. At decision block 954 a determination is made whether actuator status has changed. If the status has not changed the algorithm loops back to step 949 to receive additional setpoint data from the host. If the actuator status has changed, for example, as the result of a watch dog alarm known to those of skill in the art, the actuator status change is displayed at step 955. New actuator status configuration data is entered at step 956. At step 957 new actuator status with new loop configuration data is entered and the new configuration data is communicated to the host at step 958. The algorithm then loops back to step 949 to receive new setpoint data from the host.
It is apparent from the foregoing that the multiple-network control system described above may be used to accurately control product application in an accurate and dependable manner.
In one preferred embodiment the present invention includes a distributed process manager which functions as a process distribution controller. The process manager is embedded in the host central processing unit. Through the use of specialized software the process manager is able to monitor the work or activity being performed by the system and to redistribute the work according to the load at which the host CPU is running.
As the size of the control system and the number of nodes and independent networks increases a greater burden is placed upon the host CPU. The size of the system thus dictates that at a certain level the host CPU will exceed its processing capacity and the system will slow down. This could result in delayed control, inaccurate control or even system failure. Recognizing the potential for this problem the present system utilizes a process manager. It is the function of the process manager to monitor the level of processing activity of the host CPU. When the host processing activity exceeds a certain predetermined target level, for example, 30%, the process manager begins to download processing responsibility to the nodes based upon predetermined process priorities. Thus, the processing activity of the host CPU is maintained at or below a desired target level which allows sufficient capacity to free up the host to handle the multitude of communications which are required between different parts of the system in an efficient and timely manner.
The distributed process manager can be best understood with respect to FIGS. 13 and 14. FIG. 13 is s flow diagram of a typical time cycle of the distributed processing system. FIG. 14 is a flow diagram of the distribution of the processing of objects between the host system and a single node. Box 960 represents the start of a typical time cycle during operation of the control system. Those of skill in the art will recognize that the lower the time cycle of the control system the more accurate the control will be. Preferably, a time cycle under 100 milliseconds is desirable. During each time cycle certain objects are required to be performed. Through the use of object-oriented software it is not necessary that the objects be performed in a top-down manner. Rather, the objects may be performed simultaneously with the order of performance being dependent only upon the interrelationship of the objects being performed. For example, the commencement of a particular object may depend upon the result of another object, each being performed during the same time cycle. In FIG. 13, boxes 961 through 965 represent typical objects to be performed during a typical time cycle. Each object represents an individual program the results of which are communicated between the host and independent control modules through the use of messages. Once the time cycle is complete the system moves to the next cycle represented by box 967.
FIG. 14 is a flow diagram of the distributed process management system of the present invention between the host system and a single node. It will be appreciated that the system functions similarly with respect to each individual node. At box 968 the objects for control systems operation are defined. Those objects are placed in executable position in both the host and the node at box 969. The process begins at box 970 with all host objects being activated. This means that when the system is initially activated the host processes all objects. At decision box 971 a determination is made whether the host processor is exceeding its target usage. For example, the target usage of the total processing capacity of the host CPU may be 30%. Obviously, the target may be varied depending upon system requirements and the desires of the system designer. Since the nodes are constructed with intelligent control modules which are self-sufficient, it is possible to download all of the processing activity to the individual nodes, the host then being utilized only for system communication. Alternatively, the system allows the designer to make primary use of the host and to download processing activity only when an overload situation is being encountered.
If the host processor is not exceeding target usage the process continues at box 972 to the next time cycle with no processing activity being downloaded to the node. At this point the algorithm loops back to decision box 971 at which time another determination is made whether the host processor is exceeding target usage. As long as the target usage is not exceeded this loop will continue and the host processor will continue to process all of the objects. If the host processor is exceeding target usage the process of lowest priority is activated in the node at box 973. This represents the download of processing responsibility for the object of lowest priority to the node. At box 974 the host process is deactivated once the node process is on-line. Thus, the object will be processed by the node and not at the host. At box 975 the process manager is updated. This allows the process manager to keep track of the location of objects such as by hand shake communication. The algorithm then loops back to decision box 971 at the next cycle to determine once again if the host processor is exceeding target usage. Objects will continue to be downloaded to the nodes on the basis of the lowest priority activity until the host processor reaches a level that does not exceed target usage.
Such a distributed process manager is ideal for use with a control system having multiple independent networks, each including one or more independent intelligent nodes. Since the system allows processing requirements to be downloaded to the node level the host CPU is able to efficiently perform as a communication manager between the various parts of the control system. Such a system provides opportunity for use of multiple networks of diverse function. For example, the control system may include a product application control network, a vehicle location sensing network, a chassis control network and/or a product development sensing network for providing data relating to plant development. Each of the networks is capable of operation which is essentially independent of the others but in a coordinated and controlled manner which allows continual communication of relevant data between each separate portion of the system. Such a control system can effectively be used to control not only the application of product but also the vehicle as it travels over the field. In particular, based upon sensed data which may include vehicle speed, engine RPM, transmission status and terrain topography the control system may include various nodes and actuators which control vehicle speed, steering, and transmission gear ratio. The result is that the mobile product application system becomes a generally hands-off system with no active intervention required by the operator. Such a system may be remotely operated or may include an operator present in the vehicle to monitor system status and to assume manual control should it be necessary.
Having thus described the preferred embodiments in sufficient detail to permit those of skill in the art to practice the present invention without undue experimentation, those of skill in the art will readily appreciate other useful embodiments within the scope of the claims hereto attached. For example, although the present invention has been described as useful for the agronomic industry, those of skill in the art will readily understand and appreciate that the present invention has substantial use and provides many benefits in other industries as well. Some of those may include practicing the present invention to spread products such as sand, gravel or salt on roads or highways. In general, the agronomic community would find the present invention useful in applying products such as fertilizers including starter fertilizers, insecticides, herbicides, as well as seeds. It is to be further appreciated that the present invention may apply such products in either liquid or solid form. The invention is also applicable to areas of the agronomic industry in addition to application of products. For example, it is known to be desirable to adjust the depth of tillage equipment in dependence on soil moisture. However, no presently used equipment is able to control tillage depth on a real time basis. By utilizing the control system of the present invention it is possible to use real time data collected by a soil moisture sensor to adjust tillage depth by controlling an actuator(s) which is responsible for depth control. Thus, the present invention is applicable to any situation where it is desirable to control an agronomic input (e.g., product application, tillage depth, etc.) based upon the real time measurement of data related to crop development.
|Cited Patent||Filing date||Publication date||Applicant||Title|
|US4015366 *||Apr 11, 1975||Apr 5, 1977||Advanced Decision Handling, Inc.||Highly automated agricultural production system|
|US4176395 *||Nov 16, 1977||Nov 27, 1979||Clemar Manufacturing Corporation||Interactive irrigation control system|
|US4209131 *||May 12, 1978||Jun 24, 1980||Motorola, Inc.||Computer-controlled irrigation system|
|US4277022 *||Dec 14, 1978||Jul 7, 1981||Dennis W. Holdsworth||Mobile material distribution system|
|US4350293 *||Sep 16, 1980||Sep 21, 1982||Lestradet M C J||Vehicle equipped with a liquid spreader device|
|US4392611 *||May 15, 1981||Jul 12, 1983||Dickey-John Corporation||Sprayer control system|
|US4527353 *||Apr 4, 1983||Jul 9, 1985||Newby John C||Irrigation/fertilization control and distribution system|
|US4545396 *||Feb 25, 1985||Oct 8, 1985||Miller Richard N||System for optimum irrigating and fertilizing|
|US4553702 *||Feb 4, 1983||Nov 19, 1985||Imperial Chemical Industries Plc||Spraying system|
|US4588127 *||Jul 30, 1982||May 13, 1986||Ehrat Arthur H||Material-spreading field vehicle having means for on-site metering and mixing of soil-treating chemicals|
|US4630773 *||Feb 5, 1986||Dec 23, 1986||Soil Teq., Inc.||Method and apparatus for spreading fertilizer|
|US4714196 *||Aug 8, 1986||Dec 22, 1987||Agrobotics, Inc.||Farm chemical delivery system|
|US4721245 *||May 12, 1986||Jan 26, 1988||Multinorm B.V.||Method and an apparatus for spraying a liquid|
|US4755942 *||May 17, 1985||Jul 5, 1988||The Standard Oil Company||System for indicating water stress in crops which inhibits data collection if solar insolation exceeds a range from an initial measured value|
|US4803626 *||Sep 15, 1987||Feb 7, 1989||Dickey-John Corporation||Universal controller for material distribution device|
|US4805088 *||Mar 23, 1987||Feb 14, 1989||Cross Equipment Company, Inc.||Method and apparatus for microprocessor controlled sprayer|
|US4823268 *||Jun 23, 1987||Apr 18, 1989||Clemson University||Method and apparatus for target plant foliage sensing and mapping and related materials application control|
|US4895303 *||Jul 15, 1988||Jan 23, 1990||Freyvogel Frank C||Sprinkler system fertilization regulator|
|US4925096 *||Sep 5, 1989||May 15, 1990||Nomix Manufacturing Company Limited||Apparatus for delivering a liquid|
|US4967957 *||Nov 2, 1989||Nov 6, 1990||Dickey-John Corporation||Injection mixer|
|US4992942 *||Jan 25, 1989||Feb 12, 1991||Bahm, Inc.||Apparatus and method for controlling a system, such as nutrient control system for feeding plants, based on actual and projected data and according to predefined rules|
|US5014914 *||Mar 4, 1988||May 14, 1991||Wallenaas Anders||Dose control apparatus for agricultural tube sprayers for spreading pesticides on fields and plants|
|US5021939 *||Mar 16, 1989||Jun 4, 1991||Demaco Mfg. Group, Ltd.||Computerized sprinkler control system|
|US5077653 *||Nov 27, 1989||Dec 31, 1991||Christian Barlet||Process and device for programmed spreading of an active product on the surface of the ground|
|US5170820 *||May 3, 1991||Dec 15, 1992||David P. Ward||Management system for the application of anhydrous ammonia fertilizer|
|US5184420 *||Dec 14, 1988||Feb 9, 1993||Labbate Climate Control System Inc.||Computerized fertilizer injection system|
|US5220876 *||Jun 22, 1992||Jun 22, 1993||Ag-Chem Equipment Co., Inc.||Variable rate application system|
|US5246164 *||Dec 16, 1991||Sep 21, 1993||Mccann Ian R||Method and apparatus for variable application of irrigation water and chemicals|
|US5260875 *||Aug 20, 1991||Nov 9, 1993||Micro-Trak System, Inc.||Networked agricultural monitoring and control system|
|US5313578 *||Dec 23, 1990||May 17, 1994||Motorola, Inc.||Portable interprocess communication facility|
|US5355815 *||Mar 19, 1993||Oct 18, 1994||Ag-Chem Equipment Co., Inc.||Closed-loop variable rate applicator|
|US5453924 *||May 16, 1994||Sep 26, 1995||Ag-Chem Equipment Company, Inc.||Mobile control system responsive to land area maps|
|US5463735 *||Feb 3, 1994||Oct 31, 1995||Johnson Service Company||Method of downloading information stored in an arching device to destination network controller through intermediate network controllers in accordance with routing information|
|1||"An Efficient Recovery Protocol for Distributed Network Planning Information with Network Partitoning and Equipment Failure", by Farnham et al., Globecom '95 Communications for Global Harmony, IEEE Global Telecommunications, May 1995, pp. 952-957.|
|2||"Data Transfer Bottlenecks over SPARC-Based Computer Networks", by E. Saulnier and B. Bortscheller, IEEE Local Computer Networks, 1995 20th Conference, pp. 289-295.|
|3||"Resource Allocation in a Flexible Manufacturing System by Graph Matching", by Shen et al., Proceedings of the 1991 IEEE International Conference on Robotics and Automation, Apr. 1991, pp. 1315-1320.|
|4||*||An Efficient Recovery Protocol for Distributed Network Planning Information with Network Partitoning and Equipment Failure , by Farnham et al., Globecom 95 Communications for Global Harmony, IEEE Global Telecommunications, May 1995, pp. 952 957.|
|5||*||Data Transfer Bottlenecks over SPARC Based Computer Networks , by E. Saulnier and B. Bortscheller, IEEE Local Computer Networks, 1995 20th Conference, pp. 289 295.|
|6||*||Resource Allocation in a Flexible Manufacturing System by Graph Matching , by Shen et al., Proceedings of the 1991 IEEE International Conference on Robotics and Automation, Apr. 1991, pp. 1315 1320.|
|Citing Patent||Filing date||Publication date||Applicant||Title|
|US5924371 *||Sep 23, 1997||Jul 20, 1999||Case Corporation||Global controller and distributed local controller(s) for an agricultural implement|
|US5978578 *||Jan 30, 1997||Nov 2, 1999||Azarya; Arnon||Openbus system for control automation networks|
|US6085135 *||Feb 18, 1998||Jul 4, 2000||Claas Kgaa||Method for agricultural map image display|
|US6198986 *||Apr 30, 1999||Mar 6, 2001||Ag-Chem Equipment Co., Inc.||Pre-charged multi-variable rate crop input applicator machine|
|US6216614||Apr 28, 1999||Apr 17, 2001||Ag-Chem Equipment Co., Inc.||Boom dispensing point control system|
|US6230091||Apr 30, 1999||May 8, 2001||Ag-Chem Equipment Co., Inc.||Variable flow spray nozzle system|
|US6246938 *||Oct 10, 1997||Jun 12, 2001||Giesecke & Devrient Gmbh||Vehicle for spreading products on the road surface, in particular de-icing products|
|US6253691||Apr 30, 1999||Jul 3, 2001||Ag-Chem Equipment Co., Inc.||All wheel steer variable load carrying tractor vehicle|
|US6330587||Dec 21, 1998||Dec 11, 2001||Ford Global Technologies, Inc.||Real-time multiprocessing computer infrastructure for automated testing|
|US6356830 *||Aug 11, 1999||Mar 12, 2002||Purdue Research Foundation||System and method for automated measurement of soil pH|
|US6394011||Mar 21, 2001||May 28, 2002||Ag-Chem Equipment Company, Inc.||All wheel steer variable load carrying tractor vehicle|
|US6542076||Apr 17, 2000||Apr 1, 2003||Raymond Anthony Joao||Control, monitoring and/or security apparatus and method|
|US6542077||Aug 20, 2001||Apr 1, 2003||Raymond Anthony Joao||Monitoring apparatus for a vehicle and/or a premises|
|US6549130||Mar 29, 1999||Apr 15, 2003||Raymond Anthony Joao||Control apparatus and method for vehicles and/or for premises|
|US6584919 *||Mar 15, 2002||Jul 1, 2003||Agco Corporation||All wheel steer variable load carrying tractor vehicle|
|US6587046||Oct 30, 2002||Jul 1, 2003||Raymond Anthony Joao||Monitoring apparatus and method|
|US6708631||Apr 30, 1999||Mar 23, 2004||Agco Corporation||Variable payload tractor vehicle with coordinated crop input management system|
|US6990459||Aug 21, 2001||Jan 24, 2006||Deere & Company||System and method for developing a farm management plan for production agriculture|
|US7380733||Jul 21, 2006||Jun 3, 2008||Barron & Brothers International||Plural bin metering system|
|US7814187 *||Aug 6, 2001||Oct 12, 2010||Siemens Aktiengesellschaft||Method and processing system for determining the spatial structure of a control system|
|US8326954||Dec 4, 2012||Caterpillar Inc.||System and method for synchronizing configurations in a controller network|
|US8428777 *||Aug 1, 2012||Apr 23, 2013||Google Inc.||Methods and systems for distributing tasks among robotic devices|
|US8522700||Jul 15, 2008||Sep 3, 2013||Donald K. Landphair||Air seeder with GPS based on/off control of individual row units|
|US9075136||Mar 1, 1999||Jul 7, 2015||Gtj Ventures, Llc||Vehicle operator and/or occupant information apparatus and method|
|US20020035431 *||Oct 17, 2001||Mar 21, 2002||Todd Ell||System and method for creating application maps for site-specific farming|
|US20020040300 *||Jun 5, 2001||Apr 4, 2002||Agco||System and method for creating controller application maps for site-specific farming|
|US20020103688 *||Aug 21, 2001||Aug 1, 2002||Schneider Gary M.||System and method for developing a farm management plan for production agriculture|
|US20030181993 *||Aug 6, 2001||Sep 25, 2003||Stefan Linke||Method and processing system for determining the spatial structure of a control system|
|US20030208319 *||Jun 5, 2001||Nov 6, 2003||Agco||System and method for creating demo application maps for site-specific farming|
|US20050062662 *||Jun 22, 2004||Mar 24, 2005||Mitsumi Electric Co. Ltd||Antenna unit having a wide band|
|US20070034721 *||Jul 21, 2006||Feb 15, 2007||Steve Owenby||Plural bin metering system|
|US20100017073 *||Jul 15, 2008||Jan 21, 2010||Landphair Donald K||Air seeder with gps based on/off control of individual row units|
|US20100023605 *||Jan 28, 2010||Caterpillar Inc.||System and method for synchronizing configurations in a controller network|
|WO1998036518A2 *||Jan 29, 1998||Aug 20, 1998||Arnon Azarya||Openbus system for control automation networks incorporating fuzzy logic control|
|WO1998036518A3 *||Jan 29, 1998||Nov 12, 1998||Arnon Azarya||Openbus system for control automation networks incorporating fuzzy logic control|
|WO1999016007A1 *||Sep 18, 1998||Apr 1, 1999||Case Corporation||Global controller and distributed local controller(s) for an agricultural implement|
|U.S. Classification||700/2, 701/1, 701/50|
|Jan 24, 1996||AS||Assignment|
Owner name: AG-CHEM EQUIPMENT COMPANY, MINNESOTA
Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:MONSON, ROBERT J.;REEL/FRAME:007858/0800
Effective date: 19960123
|Apr 27, 2001||AS||Assignment|
Owner name: COOPERATIVE CENTRALE RAIFFEISEN-BOERENLEENBANK, B.
Free format text: SECURITY INTEREST;ASSIGNOR:AG-CHEM EQUIPMENT CO., INC.;REEL/FRAME:011783/0088
Effective date: 20010417
|Dec 18, 2001||REMI||Maintenance fee reminder mailed|
|May 24, 2002||FPAY||Fee payment|
Year of fee payment: 4
|May 24, 2002||SULP||Surcharge for late payment|
|Mar 12, 2004||AS||Assignment|
Owner name: RABOBANK INTERNATIONAL, NEW YORK
Free format text: SECURITY AGREEMENT;ASSIGNOR:AGCO CORPORATION;REEL/FRAME:016480/0354
Effective date: 20040105
|Sep 23, 2005||FPAY||Fee payment|
Year of fee payment: 8
|Nov 20, 2009||FPAY||Fee payment|
Year of fee payment: 12