US 20050033505 A1
Traffic control systems, including an automated vehicle traffic assessment system. The system may measure and report on wait time and may include video cameras and image recognition software. The system may be used in connection with any flow-regulated area, including border crossings, toll roads and ports of entry.
1. A traffic assessment system comprising:
a. a vessel monitor configured to monitor vessels in one or more open traffic lanes that are waiting to pass through one or more flow regulation controls associated with the one or more open traffic lanes;
b. a flow monitor configured to monitor the flow of vessels passing through the one or more flow regulation controls; and
c. an analyzer configured to determine the likely effect of opening one or more closed traffic lanes or closing one or more of the open traffic lanes on the waiting time of a waiting vessel based on information from the vessel monitor and the flow monitor.
2. A traffic control system comprising:
a. a traffic assessment system configured to assess the waiting time a vessel entering an area controlled by one or more flow regulation controls is likely to need to wait before passing through the controlled area; and
b. a transmitter for transmitting information about the waiting time to vessels not in the controlled area.
3. A traffic assessment system comprising:
a. a camera system configured to create an image of at least a portion of an area that receives vessels and that is controlled by one or more vessel flow regulation controls;
b. a vessel recognition system configured to analyze information about vessels within a defined area of the image; and
c. a user interface configured to allow a user to define the area of the image from the camera system that should be analyzed by the vessel recognition system.
This application is based upon and claims priority to U.S. Provisional Application Ser. No. 60/431,117, filed Dec. 5, 2002, entitled TRAFFIC SURVEILLANCE & REPORT SYSTEM, the entire content of which is incorporated herein by reference.
The invention is directed generally to traffic control systems, and particularly to an automated vehicle traffic assessment system, and more particularly to wait-time measuring and data reporting system that includes video cameras and image recognition software used for monitoring roadways that are flow-regulated, for example, at a border crossing or port of entry. The system can also be used to monitor boat traffic in ports of entry or monitor toll-crossing road stops.
The system comprises video cameras and image recognition software to monitor traffic at a predetermined location for the purpose of determining and reporting the wait-time for vehicles waiting in line at locations, such as a border crossing, port of entry, stop light or other locations where traffic may be stopped. In the preferred embodiment, the system also provides data about virtual conditions, including reporting data about the implications of opening or closing one or more lanes of traffic and the effects of a change in the rate of arrival of new vehicles into waiting traffic lines. For example, the system can predict how the total wait time would change if one lane of a border or toll crossing was closed or opened. In another example, the change in the total wait time could be estimated where the current arrival rate of vehicles increased—to simulate rush-hour traffic, for instance. Additionally, the system may provide data about the waiting vehicles including the color, make, and speed of each vehicle waiting in line. Further, the preferred system can provide data about the width, length, and height of each vehicle waiting in line, and this data may be used to identify specific vehicles types.
The system can be used by an administrator at border crossings or port of entry where the administrator may desire to know the length of time a vehicle will wait in line if it arrives under the current traffic conditions. This information can provide the administrator with data that is helpful in deciding whether to open or close additional lanes, or make other changes in the operation of the crossing or port. This can have significant impact on commercial and private traffic. In the preferred embodiment, the system may use a variety of commonly known data delivery techniques including instant phone messaging, e-mail, facsimile, etc., and data from the system may also be made available to drivers in vehicles enroute to the area, to allow drivers to find alternative routes during periods of long wait-times.
The system uses data provided by video cameras and processed by image recognition software to determine the flow rate of vehicles passing through gated areas or exit lanes. The system automatically detects the end of the line of vehicles and calculates the total number of waiting vehicles and the wait time for the vehicles. The system thus allows for the capture of live images of traffic conditions while image recognition and processing software analyzes the video to produce real-time or archived data about current traffic conditions. One of the unique aspects of the preferred embodiment of the system is its use of two primary components to determine the overall wait-time. These components include fixed or movable cameras that monitor the exit lanes to determine the vehicle flow rate of traffic through the border crossing, and these cameras work in combination with the second component which includes a combination of fixed and movable cameras to automatically search for and find the end of the current traffic line and determine the total number of waiting vehicles. The system may also use image recognition or other methods to deliver data about the total number of vehicles waiting in the targeted lanes.
In a preferred embodiment of the present invention, the system also uses two methods to reduce number of cameras needed to find the end of the line and count the number of vehicles waiting (and determine total wait time) along a roadway. These methods also improve the useful range-of-view of the cameras.
The first method is automatic image registration. This method allows the system to use movable cameras with high telephoto lenses that are focused on very distant objects. Because of errors in most common camera pan/tilt mechanisms, use of image subtraction techniques for finding new or moving targets is normally not practical at high telephoto (where mechanical errors by pan/tilt mechanisms are magnified). By using automatic image registration, the present invention can automatically correct for errors in the field of view. This allows for the use of image subtraction techniques because the computer can automatically make a pixel-to-pixel match between the current view and a previous view, even if the camera is not precisely aimed at the same location. This allows each camera to work in concert with image recognition software to provide coverage of a much broader area, thereby reducing the number of cameras needed and increasing the viewing range of the system.
A second method that allows the use of fewer cameras in this preferred embodiment is a method of counting vehicles using user input to supplement the system's vehicle count estimate generated by the image recognition software. Using this method, the user enters the total number of vehicles observed in each camera view during the system set-up phase. Once in operation, the system relies on this data during conditions when automatic counting techniques are unreliable. For example, during high-telephoto views vehicles become compressed, and image recognition software is normally incapable of separating the image into individual vehicles. This renders data from such views inaccurate. Without a method for automatically switching to manually counted data, operation of the system with high telephoto cameras would be impractical under most circumstances unless cameras were mounted on exceedingly tall structures.
Data from the system may provide users with information about rate of flow of traffic in each and all lanes, the total number of cars waiting, and the estimated wait time for vehicles in line or arriving at the end of the line. Other data may include information about the vehicles observed, including the color, length, width of each vehicle and averages of all vehicles.
One embodiment uses a third camera to gather data about vehicle profiles to allow for a vehicle identification system that employs both profile and overhead video cameras to provide width, length and height. Additionally, the data may include the direction of each vehicle, whether the vehicle was moving in the predicted direction, and the average velocity vector of vehicles. This information provides “Opposite Direction Of Travel” (ODOT) alerts, which are specifically used to alert border agents, port police, or other officials of vehicles attempting to enter a secure area via an outbound lane. All of this data may be used immediately or it may be archived. The data may be sent to servers in other locations, distributed to other interested users for display to email or cell phone messaging or to highway alert signs. The system may also capture still images and video clips of target events or objects, and this recorded data may be stored or sent to a user.
In a first preferred embodiment, the system is designed specifically to monitor traffic conditions at a border crossing or port of entry. In such a system, data is generated about traffic conditions including: rate of flow of traffic, number of vehicles waiting in line, and total wait time. Data is also generated that includes the number of vehicles arriving over time, ODOT alerts, speed, color, size, and type of vehicles waiting in line. This data is provided to users in the form of text and/or graphical information, including still images and video clips. Some data provides user alerts. For example, data about a vehicle traveling in the opposite direction of normal traffic flow will generate an ODOT alert. If one lane of traffic is flowing at an unusually high rate, another alert may be triggered to notify management of the condition, for example, if a border crossing agent is not thoroughly processing or inspecting vehicles. Based on the user's set-up values, alerts may be text, a still image capture, video clip of the event, email, or other messaging method. Data about a specific color or size of cars may fit a requested “search image” and may also trigger a user alert that will employ the sending of a text message, still image capture, video clip, or other messaging method.
Real time images may be delivered from video cameras to a computer via hardwire, fiber optic or wireless connection. The images may be from cameras that are positioned to determine the vehicle passage rate, or the images may be from cameras that are positioned to find the number of cars or end of the line. Images are in the form of analog or digital video and may be from visible light, infrared or thermal imaging cameras. The images are imported into a computer and are processed by image recognition software. This software may use several methods to determine whether cars are present in the field of view. The software may also use several methods to determine the direction/vector of travel of any moving vehicles. The methods used by the image processing software include, alone or in combination, texture, edges, image blurring, color, and target shapes/line segmentation. Other methods also may include image registration to assure accurate camera positioning, and image subtraction to find new and moving targets. In the preferred embodiment cameras used in the system are visible light, however infrared or thermal imaging cameras may also be used.
Besides monitoring border crossings, the system can be used to monitor traffic intersections to provide data about traffic wait times, number of vehicles waiting, flow rates, and other data about individual vehicles waiting in line. The system may also be used to monitor and provide data about boat traffic in harbors or at toll-crossing road stops.
In a preferred embodiment, the software also provides for a method to help the system's ability to focus on relevant image areas by the employing “target area painting” software tools that allow a user to define specific areas of each camera view that are target or non-target areas. The system may also use software that allows the computer to “learn” traffic conditions over time by storing and then comparing image recognition data including the following methods, alone or in combination: texture, edges, image blurring, color, target shapes and image subtraction. The process may be used to assist the image processing system in accurately identifying current traffic conditions.
In one embodiment, the software may display a user interface that includes the flow rate of traffic in each lane, the total flow rate for all lanes, the number of cars waiting in all lanes, and the total wait time. Other data displayed in the user interface may include ODOT alerts, color and sizes of vehicles, whether or not the vehicle fits a desired profile, charts that display a histogram of traffic flow/wait time/total number of cars, camera views, current camera, and current camera position. In another embodiment, the software may be used without a user interface to display real-time data. In this embodiment, the software delivers data to a separate program which makes use of the data, and this program may reside on the same or a different computer than the image processing system. In both embodiments, when a user interface is used and when one is not used, the software may include networking capabilities to provide for data export to other users or servers. Still images captured during alert conditions may also be displayed or sent to other users. Video clips of alert conditions may also be displayed or sent to other users.
System Components. Referring to
Traffic Flow Component. To determine the flow rate of traffic, a preferred embodiment of the system uses of one or more cameras (visible light, infrared or thermal imaging) aimed at the “gating” area—the area where vehicles either enter or exit the checkpoint.
Set Up. In a preferred embodiment, a user sets up the area of the camera's view to be monitored by painting the area 25 of the lane with a marker color 26 to tell the computer program exactly which area is lane number 1, lane 2, lane 3, etc. Each painted area 25 can be labeled to identify the area. Similarly, the user can define each lane using a different color. Once this area is established, the user “clicks” on the colored area to bring up a text box in which the user can then enter a lane number. When a vehicle is detected moving across one of these painted areas 25, the program records activity in that lane. In another embodiment, lane markers 28 are defined by the computer as shown in
The user also may define “object length” information to the program. For example, the user clicks on the display screen of the computer and “pulls” a line across an object of any known length. The user then defines the length of the line by entering a line-length value into a text box, as shown in
In a preferred embodiment, the user may also define whether the computer should capture still images or video clips of target conditions including: Specific vehicles (based on vehicle color and/or size), vehicles that were stopped beyond or below user-set thresholds of time, and ODTO alerts.
Detection/image recognition-Image Processing. The invention will preferably use a computer with software to detect the presence or absence of vehicles. There are a plurality of methods for detecting images and the present invention may employ a variety of standard, commonly known techniques as described herein. In the preferred embodiment, the computer program employs one or a combination of the following image processing methods to determine the absence, presence, and relative location of vehicles in the field of view: texture, edges, line segments, shapes, and color (see description of these below). These may be combined with image subtraction techniques (motion) to determine whether vehicles are absent or present and, if present, their relative position in the field of view.
For example, the computer program determines when a vehicle is present in the target area by determining the contrast between an empty stretch of roadway and one with a vehicle one it. This method simply employs texture as defined below (see Texture Evaluation below). Because a roadway is relatively flat and has a generally low texture value, the change to a high textured area-as when a vehicle is present-can be used as a trigger event. The threshold for the trigger can be set manually, or by a combination of manual and automatic means. The automatic level can be based on an observed contrast, texture, or line segments over time. The automatic level may also consider these levels at the same time in previous days, to account for daily changes such as shadows. In addition to looking for shadows, the software may also look for edges, shapes, or colors. When using infrared or thermal imaging cameras, the system may also look for heat areas associated with each vehicle. Heat areas are treated in a manner similar to other image processing data where texture, edges and shapes assessments may be applied.
In another embodiment, the present invention detects motion by itself or in combination with other methods to determine when a vehicle is present (see Motion Tracking below). The system looks for significant change within a target area and correlates such a change with the presence of a vehicle.
In a third method, the software uses “image subtraction” (See Image Subtraction below) to determine whether vehicles are present. The program may also employ the use of texture, edge detection, color, line segment and motion path information (see description of these methods below) to determine the presence of and information about vehicles in the field of view.
The system may use a combination of one or more of the above listed methods to determine the absence or presence and location of vehicles. To anyone familiar with current image processing techniques it should be apparent that a variety of methods may be used to find vehicles.
To find the width and length of a vehicle, the system may use one or a combination of several of the above methods to “find” a vehicle and then calculate its width and length based on the calibration data provided by the user during program set up as described below.
Operation. In the preferred embodiment, operation of the traffic flow component is continuous while the program is active. All cars are monitored and data is made available to the program for displaying and storing flow rates for the lanes. The data is also made available to the system for calculating the total wait time (by using this data in conjunction with the total number of cars), and the data may be sent to a web page or other programs for use by other programs or display on the Internet.
In a first preferred embodiment, the normal operational loop for this component (the traffic flow primary loop) is shown in
Referring now to
End of Line Component. The end-of-line component provides information on the number of vehicles that are waiting. The term “waiting” implies vehicles that are moving at a rate defined by the user at startup with a default of approximately 5 miles per hour. This component uses information gathered from one or more cameras, movable, fixed or a combination of both, that provide video data to a computer that processes the video. In a preferred embodiment as described below, this component uses a variety of image processing techniques to determine where vehicles are located, the vehicle's trajectory and speed. Once vehicles are identified, the computer calculates the total number of vehicles that are considered to be “waiting”. This number is divided by the flow rate to produce a “total wait time” value. Additionally, data gathered from moving vehicles is provided to the computer to generator “ODOT alerts” if vehicles are moving in a non-standard manner. ODOT alerts are triggered when a target object does not move in the direction previously defined by the user during the program set-up phase as described below. ODOT alerts can be calculated by monitoring each target object and determining whether it is moving across the screen in the predefined direction. For example, if the user defined “expected traffic flow” to be from the upper left of the screen to the lower right, then the computer would not trigger ODTO alters for any target object moving in this direction. However, if an object is moving across the screen in a direction significantly different from the predefined direction (the threshold for this is defined by the user on the preference page of the program prior to running the program), an ODOT alert will be generated.
Set UP. In a preferred embodiment, setting up the program requires a user to specify key information. The information needed by the computer to accurately provide traffic information includes instructions for where to position movable cameras, camera view order, defining traffic lanes for each camera view, defining the number of vehicles in each camera view (and in each lane within the camera view), and the camera view order, and defining the default “wait time” for each new camera position.
Defining Camera Positions. The user moves one or several cameras to desired positions and then stores these positions as presets. The object is to provide camera coverage of the entire area in which traffic monitoring is desired, from the “gating/border crossing area” to the furthest part of the roadway where traffic may be backed up to.
Defining Camera View Order. A user may also define the order in which the computer calls each camera position and each camera (if more than one camera is present). For example, the user may send a command to move the first camera to a view of the road nearest the “gated or border crossing” area. The user may then define this as camera position number one. The user may then command the camera to move to a position where the next section of roadway can be viewed, and the user may then assign this position and camera as position number two. If the user then desires to use a view from another camera to view the next section of roadway, the user would then send a command to switch to the new camera and the new camera position. The user would then define this camera and its current view as position number three, and so on.
For example, this can be achieved by using multiple movable cameras to monitor a length of road of several hundred yards to many thousands of yards long. The cameras may be moved by using a camera control page on the user interface, and specific position for the cameras are set and stored. The position includes pan, tilt and zoom. Also, fixed cameras may be positioned in a manner that allows multiple cameras to view a desired roadway. In either case, a user interface allows a user to select the “view order” which defines a camera order from nearest to farthest. This information is used by the computer to progressively scan a line of vehicles and to know which camera view is the next view for the computer to switch to as data is gathered about a long line of vehicles.
Defining Lanes. Once the camera views are set and ordered, the actual lane areas may be defined. The lane areas can be defined either by a user painting the image or by the computer automatically by entering set-up mode and allowing the computer to track motion to automatically define lane areas.
Within each camera view, the area of each frame that the system should search for vehicles can be defined. This is the “traffic area” and by defining this area, the user may reduce the total number of pixels the computer needs to review for each camera position. To define the traffic area, a user may use one of several methods. A user may obtain a still image of each camera view and then paint the area where traffic is expected. The user may paint the area with a bright, highly saturated color to identify a lane as a “target area” as shown in
The “target area” may be defined automatically by the computer. In this method, a user instructs the computer to monitor motion within the camera view. The area in which motion occurs within the area during the set up time is automatically defined by the computer as “target areas.” While the computer gathers data on motion areas, it also gathers vector data-defining the direction that motion should be expected in each field of view.
Defining the number of vehicles. The user may also provide the computer with data about the total number of vehicles associated with each camera view. This may be accomplished by having the user assign a “vehicle waiting” value to any location within each field of view by entering a wait time data in a text box. For example, the first camera view (the one closest to the border crossing) might be labeled with a “0” near the crossing portion of the field of view and a “16 on the opposite site of the view area. This would mean that if vehicles are detected filling half of the field of view, that approximately 8 vehicles are waiting in line. If vehicles are found to fill none of the frame, then 0 vehicles will be reported. In this case the computer would automatically switch to the next previous of view (unless the current view is the first view) because this field of view shows that the end of line is somewhere before this camera view. The system may also determine the number of vehicles by “counting” the total number of vehicles it finds in each camera view. To automatically generate the total number of vehicles, the computer “counts” the vehicles present at the first camera view, and then it adds this number to the count at the second camera view, and so on. Once a camera view is reached in which the lane is not completely full of vehicles, the computer may end the addition process. The vehicles can be counted by the computer by a method described below.
Defining default “waits time.” In a preferred embodiment, the user may also define a default wait time for each new camera view. By defining this time, the user instructs the computer how long to “wait” at each new camera position while it collects image processing data. A typical time range is 20 to 40 seconds, and such time may be useful to help the computer confirm the presence or absence of vehicles. This is especially useful during times when the computer relies on motion information. If traffic is completely stopped, a longer “wait time” defined by the user will allow the computer more time to view motion in the frame and confirm presence of vehicles.
The user may also define the vector deviation tolerance. This value is entered on the set-up page and provides the system with the allowable degree of offset between the defined path of motion and the current path of motion before an alert is triggered. This feature allows users to adjust whether an alert is triggered for a vehicle that simply changes lanes or one that is driving the wrong way (180 degrees opposed to traffic) in a lane.
The user may also define the relative motion that a vehicle/target may move before it is considered to be “waiting”. In a preferred embodiment, this value provides the user with a method for adjusting the total wait time based, in part, on the speed of the waiting vehicles. This value may be used by the motion tracking method of the image processing function.
Image Processing. In a preferred embodiment of the present invention, a computer program employs one or more of the following image processing methods to determine the absence, presence, and relative location of vehicles in the field of view (from visible-light, infrared or thermal-imaging cameras): Texture, edges, line segments, shapes, color, and motion (see description of these below). These may or may not be combined with image subtraction techniques to determine whether vehicles are absent or present and if present, their relative position in the field of view.
Operation of End of Line Component. In a preferred embodiment, the end of line component of the system operates in two states: Startup and Normal operation mode.
Start Up Mode. During the startup mode, the computer program sends a command to selects the camera view of the roadway closest to the gated or border crossing area. The computer then processes the incoming video to determine if the current view if full of vehicles, partially full, or empty. The computer does this by “searching” for vehicles as described in “target finding”. Based on the determination, the computer may select the next camera view, a previous view, or it may keep the current image select. In the case of a partially filled lane, the camera may not be commanded to move to the next/previous position. If the view at the first position reveals an empty or partially filled lane, the camera will stay at the current position. While the computer is determining the lane's fullness, the computer may gather data on the motion of the vehicles, the number of vehicles, and other unique features of the vehicles (or other objects like humans, dogs, etc.) in the view. The computer stays in Set-Up mode until the computer detects that the current view is either empty or partially full of vehicles. Under these conditions, the end of the line has been reached, and data on the length of the line may be reported to the user or other peripheral devices. When the Set-Up mode is exited, the program enters the Normal Operation mode.
Normal Operation Mode. While in operation mode, the computer monitors the current view and moves the camera from one position to the next. The computer assesses the images from the current view and determines whether the current view is empty, partially full, or full of vehicles. For example, if the computer determines that the current lane view if full of vehicles, it may send a request to attached hardware (cameras and/or video switcher) to view the next section of roadway. This request can be either a request to move the current camera to its next position or to select a different camera. The computer will then evaluate the new camera position as outlined above and either move to the next position, stay at the current position or return to the previous position. If the computer recognizes that no vehicles are present in the new position, it will return to the previous view and it will extend the “wait time” at this position. By extending the wait time, the computer suppresses the possibility of an endless oscillation between two camera positions. The wait time may be returned to the default value when the computer detects a lane that is partially filled with vehicles.
There are two end-of-line loops running on separate threads of the program. These include the Main loop for camera control/data assessment and the image processing loop.
If the system perceives no vehicles in the lane, it then goes back to the previous camera position. If the current camera position is position number 1, it then reports that no vehicles are waiting and goes back to the beginning of the loop and starts again.
While this loop executes, data is continually collected by the traffic flow component. Also, the program will respond to any user requests while all of these loops are executing. These include requests for any information or any changes in operating parameters.
Image processing loop. Referring now to the flow chart of
4. The program uses image subtraction (see Image Subtraction below) to compare the current image to the previous image as described above. If no new objects are found, the current “Lane State” (LS) is set to “Empty”. If objects are found to fill a portion of the frame then the LS is set to “Partial”, and if objects are found throughout the frame, the LS is set to “Full”. The LS is reported upon request from the Main Loop, which may request the data when the “wait time” has expired. To confirm results from image subtraction or during conditions in which image subtraction can not provide high confidence data, as determined by observing erratic variations in both traffic areas and non-traffic areas of the scene, the system may rely on one or more of the following methods to confirm the objects are vehicles. In all of the cases below, the program may use pixel values from areas in the frame defined during user set up as “target areas” or lanes of traffic:
5. Track the motion of each object. Compare the motion path of the object over time with the expected trajectory path set up by the user. If the observed trajectory path is greater than the default amount, send a trajectory alert to the user. If an object is tracked moving through the entire frame within a default amount of time, the image processing component will send an immediate message to the Main Loop that the current LS is “EMPTY”. The current image subtraction, texture, color and motion data will also be stored as a body of data that signifies conditions where there is an empty lane. This becomes useful under conditions described below.
If the image processing component determines that no vehicles are present based on the above, the system may use other data collected to “learn”. For example, if image subtraction and texture assessments indicate that many vehicles are present but a target is tracked moving through the entire camera view, then the program may “learn” that current image subtraction and texture information provide unique conditions that should be assessed more carefully. Such assessment may include extending the “wait time” to allow more time to collect data under such conditions, and comparing current conditions to conditions during the same time on the previous day (in the case of shadows appearing or lights coming on/off). If, for example, a lane is empty but a flashing police light reflects off of roadway puddles, the computer may interpret this information as the presence of many “moving contiguous, high color saturation objects” which may suggest that the lane is full of vehicles. However, data provided by one vehicle moving quickly through the frame will confirm that no traffic is present since the car could not have driven through all of these vehicles. The system learns from these conflicting conditions and uses a history of previous conditions during its processing loop.
If the image recognition process determines that the lane is partially filled with vehicles, and the camera is not at a high telephoto lens condition, then the computer may report the number of vehicles in the current view (added to the total number from all previous camera positions). If the camera is at high telephoto, then the number that may be reported will be based on the number entered by the user during set up. In this case, the number of vehicles reported is proportional to the percentage of the screen currently full of vehicles. For example, if the user set up the current camera view as having 120 vehicles at a minimum and 160 at maximum (when all lanes in the current view are full), then the program will interpolate and report that 140 vehicles are present if the current camera view is half full.
Use of Data. In a preferred embodiment, the data collected from each of the camera views along with the data collected by the traffic flow portion of the system may be stored on the local computer or sent to anther computer. Data may include, among other things, the total number of vehicles, average speed, lanes most often full/empty, and the percentage of vehicles moving at each trajectory, the colors of vehicles, the average length of vehicles, the individual length of each vehicle, the average width of vehicles, and the width of each vehicle.
Auto Image Registration. In a preferred embodiment, the program compares pixel data currently received to the pixel data from the “set up” position or another previous position. The system finds the “registration error” created by slight or major inaccuracies in camera positioning due to mechanical error-especially in high telephoto conditions, and applies an error correction value to all subsequent processes. To find the correct registration, the system may compare the position of the pixels from each of the two images and “slide” one of the images to the left, right, up, down, up/left, up/right, down/left down/right, and find the position that best matches, or “registers”. The number of pixels and the direction of pixels that the registration is “off” is the correction data used throughout the image processing routines. If the registration error is beyond acceptable bounds, the program will report a “camera position error” to flag a request for user input. This system protection provides users with alerts of problems with camera mechanical devices, camera fogging, and other conditions that render the imaging system incapable of monitoring a predefined area. This is a useful component to assuring accurate data.
Image Subtraction. In a preferred embodiment, the system may use image subtraction to detect the presence of target objects such as vehicles. It does this by employing standard image subtraction techniques to determine when “new objects” are present. To do this, the computer looks at the pixel value of each or a select group of pixels in the video frame, where each of the pixels that represents an image of video has a pixel value characterizing the pixel intensity, and/or color. Data from initial frames are grabbed at times when no vehicles are present or when vehicles are present in the scene and then this data is compared with data collected throughout the operation of the program. At any given time, the program may compare the current video image to a previous image. By subtracting the current image from the initial image, an array of pixel value differences is derived. In the case where nothing has changed in a scene, the array/image resulting from subtraction of two sequential images is filled with zeros-a black image. If something has moved in the scene, subtraction produces a nonzero result at the location of the movement. The computer is then directed to begin working with the data of the nonzero pixel groups-which are considered to be candidates for being considered “target objects” or “vehicles”.
One key to this portion of the system may be to arrive at an array of pixel values that represent what the current scene would look like if no vehicles are present. To do this, the computer relies on a number of parameters to find a time when there is very high confidence that no cars are present and that the current image represents a “no vehicle” view. These can include one or more of the following methods: Waiting for conditions in which there is no movement for a predetermined amount of time, waiting for conditions in which the number of highly saturated pixels, i.e., pixels with high color value, are present, waiting for conditions in which edges and shapes that represent vehicles or vehicle shadows are not present, and waiting for conditions that closely relate to previously recorded scenes when the computer had a high confidence that no vehicles were present. In another embodiment, the program does not rely on waiting for an empty frame as the basis for image subtraction, but instead uses the difference, over time, to determine if motion is occurring. At the beginning of the wait time, the program may gather a series of images and then average these to get an “averaged view” of the lane. If vehicles are present and stopped, they will appear as solid or blurred objects. If vehicles are rushing through the lane, because there is no stopped cars, the rushing vehicles will be “averaged out” of the initial image and the remaining image will perceived as “road with no cars”. This image may be compared with a similar image created at the end of the wait time, and these two images may be compared against each other using image subtraction and other techniques to determine if vehicles are present or not. The system then finds a vehicle by evaluating whether the nonzero pixels are in contiguous groups large enough to be a vehicle. The threshold for such a determination may be made by defining vehicle size during set up for each camera position. Once a “target” is found, the computer may then monitor its movement. Here contrast, edges, or color may be used for evaluation of the scene. In either case, movement may also be tracked. Data from the movement can be used to alert users that a vehicle is moving in the correct/incorrect direction, and this may be used to alert users if a vehicle is attempting to enter a border crossing area on an outbound lane—thus triggering an ODOT alert to users.
Texture Evaluation. In a preferred embodiment, the pixels in the defined “target” areas are reviewed by the system and are searched for areas with strong intensity contrast. Since edges often occur at image locations representing object boundaries, this method may divide the image into areas corresponding to different objects. For example, this technique can provide data that the area under review is not an empty road, but a more complex, textured, object. This method allows the program to represent the target area of the image by its edges and therefore reduce the amount of data while retaining most of the image information-thus allowing other processing to be accomplished without hampering processor time. Since edges consist of mainly high frequencies, the program may also detect edges by applying a high pass frequency filter in the Fourier domain or by convolving the image with an appropriate kernel in the spatial domain.
Color Evaluation. The program may employ techniques of color evaluation to determine whether an object exists in an array of pixels and if that object may be a vehicle or other target object or not. The color evaluation process looks at each pixel's RGB, CMYK, or other color, hue or saturation value to determine, by processing, the values of the pixels' “color plane” values. The program then classifies, or aids in classifying fully or partially, the object as “target” or non-target. For example, the program may employ color evaluation by taking a video frame as input and then return as output how many objects are found, the coordinates of the center of each object, and the “color” of each object as classified into the classes “red”, “green”, “blue”, or ‘other’. The program may then use this information to conclude that a target with highly saturated red, for example, may only be a vehicle. This may be used to assist the program in finding and/or confirming previous data assessments. Practically, this is used to “confirm” the assessments of other functions in the program, and the program may not rely entirely on color data.
Line Segment Evaluation. The program may use line segments to determine or assist with finding the presence or absence of target objects. In this process, the system may use edge detection methods and evaluate the output of the edge detection process for attributes of length, direction, and (x,y) values for the center or for the endpoints of a found segment. Parallel lines found in defined target areas reflect with high confidence the presence of a vehicle. This data may be combined with motion path data to confirm the presence and movement of a target object. Because line segments are relatively invariant between consecutive frames and between ambient changes in lighting and color, this process can yield high-confidence data to the program.
Edge Detection and Evaluation. Conventional edge detection is used in the image processing component of the program to identify the boundaries of objects in a scene. The edge detection process used in the program takes one or more “pixels” in combination as input and by processing the values of the pixels' “color plane” values, classifies every pixel as part of an edge with a degree of certainty. The system may use edge detection to isolate a line in space by the grouping of contiguous edge pixels together as higher level objects, and using hysteresis to prevent streaking of the resulting line segments. These methods are standard image processing methods.
Motion Path Evaluation. In a preferred embodiment, the system's program may use data about the motion path of a group of pixels to determine if the group is an object, and to classify the object as a target or non-target object. To determine the motion path of a group of pixels or of a target or non-target, the program compares motion vectors as derived from the image subtraction process and calculated from one or more video features (using edge, color, segment data, etc.) to one or more previously specified path vectors to determine the motion of the video features. For example, this may be done by using the product derived between the two normalized vectors. If the product is 1 then the video feature is moving along the specified path. If the dot product is −1 then the video feature is moving against the specified path. If the product is zero then the video feature is moving perpendicular to the specified path. Under some conditions (as when a target is tracked through an entire frame along the predicted vector that a vehicle is expected to move) the motion path data provides high-confidence data of the state of traffic. This data may be used to trigger a learning event in the computer, and it may be used to confirm that the roadway is not full of stopped vehicles, among other things.
Vehicle Counting. In one embodiment, the program uses one or more of the following methods to find individual vehicles: contiguous areas of texture; line segments (for vehicles that are viewed from above or from near-top-down angle, two parallel horizontal lines filled in by a contiguous-color area may signify a single windshield—and this may be used for vehicles that are aimed straight at the camera, and tires may be viewed when the camera is placed at a near-perpendicular angle to the direction of moving vehicles; and continuous areas of color that are deemed by the computer to be large enough to be a vehicle or part of a vehicle such as a flat surface like hood, roof, etc. For texture, line segments and color, the determination of whether a target area is large enough is based on the vehicle size information provided in the camera set-up data. Because the size of a vehicle changes depending on the zoom value and distance between the vehicle and the camera, the user's initial set-up data can provide calibration data that the system can access whether line segments, texture or color areas are large enough to be considered targets. Once a target is considered to meet the criterion of texture, color, and/or line segment, the system may count the target as an additional vehicle and thus automatically add to the sum of all waiting vehicles.
Target Finding. Target finding includes the use of the above mentioned methods by themselves or in combination. These include: texture, line segments, color, and motion properties of the images. As discussed above, the program may use a variety of commonly known processing techniques to determine whether vehicles are present, where they are, and what their motion rates and vectors are. In one embodiment, the computer program is instructed to find areas of high texture within the targeting areas defined by the user. If a high texture area is found that matches the relative size of a vehicle (as defined by the user during set up), then the computer may apply a second test to the high texture area. This test may include looking for highly saturated colors which, if positive then the high texture area has very high probability of being a vehicle, or looking for line segments that match those of a vehicle. For example, if two parallel lines are found on a camera view that looks straight at an incoming vehicle, then the parallel lines will indicate high probability that the texture area is a vehicle. Next, the computer looks for motion. If the targeted objects are moving rapidly through the screen, then the computer can, with high probability of accuracy, asses the scene as containing no stopped traffic.
The End of Line Component may exist on the same computer as the Traffic Flow program or on anther computer. The programs may communicate with each other and other programs using a variety of communications protocols including Winsock ports. In a preferred embodiment, the program uses two computers, one for measuring traffic flow and another for the end of line component. Also, the Traffic Flow and End of Line components may each include at least one video card or other hardware method to allow video to be delivered to and processed by the computer.
User Interface. In a preferred embodiment, the user interface (see
The basic idea is to use multiple cameras and multiple preset positions for each camera, and user-specified settings such as where the road is on the screen and how far away various places are on the screen, to determine:
In any given preset position, in order to tell how long the line is, the system relies on two pieces of information:
For part 1 the program finds all pixels in the image that satisfy the following constraints:
Then the program finds all pixels that are within a given distance to these already chosen pixels. These pixels are then partitioned into groups of contiguous pixels such that there is a path from any pixel in a group to any other pixel in a group where every pixel on that path is selected. The program then attempts to match each pixel group to a previously existing pixel group. If no suitable match exists then if the group is sufficiently large it is categorized as a new car that has entered the scene. If it is not sufficiently large then it is ignored. Finally, the characteristics of each car is reported back to the program at the end of each frame. The attributes reported include the entrance distance of the car and the current distance of the car. The method of using this information to determine the length of the line is preferably in the program code. For example, a visual basic code could see that a car was picked up by the program when it was 100 car lengths away and that the car is now 25 car lengths away. From this information it could be assumed that the length of the line is at most 25 car lengths since there is a car that drove that up to that point without stopping. This is effective because if the car stops then the program may lose track of it since there is no motion.
The road region. The road region for each camera preset position is supplied by the user in the form of a bitmap file of the view of a given camera preset position the user has modified using an image processing application. The necessary modification is the selection of the region that could possibly contain cars (the road) by painting it a bright color. The program detects a large brightly colored area of the bitmap file and saves the position of these pixels for later use. The eventual use of these regions is to reduce the amount of processing needed per frame by ignoring regions where it is known that there aren't cars. An additional use is to reduce the number of false alarms that would occur when something that is not on the road exhibits characteristics that would be mistaken as car-like.
The pixel classifications. Every pixel on the screen at any moment in time is classified as having either high or low motion and as having either high or low texture. Each pixel has the opportunity to change state when each frame comes in. If the pixel is classified as high motion and is judged to have motion greater than the default minimum then the motion is reclassified as low motion. If the pixel is classified as low motion and is judged to have motion greater than the maximum then the motion is reclassified as high motion. The process is analogous with the texture state of each pixel. Notice since max>min that there is some hysteresis available so that the pixel states don't oscillate between low and high due to noise in the image.
The selection of neighboring pixels. If a pixel is next to an existing ‘car’ pixel then in the next iteration this pixel will be included as a ‘car’ pixel. This iteration is performed two or three times to expand the range of the ‘car’ pixels. The reason for this is that sometimes a car will be represented by several very close pixel groups that this operation combines into a single pixel group.
The division into contiguous pixel groups. A list of pixel groups is kept. Each pixel group is a list of pixels such that there is a path from any pixel in a group to any other pixel in a group such that every pixel on that path is also in the group. First the list of pixels groups is initialized to be empty and each pixel is marked as not being part of any group. Then iterate the following two steps until no suitable groups are found:
2) If a pixel was found then use this pixel as a seed for a flood fill algorithm that adds all the filled pixels to the group list and also marks all the filled pixels as belonging to a group.
Finding a corresponding pixel group from a previous frame. An old list of pixel groups is maintained. The old group list keeps track of the x,y position of the center of each group. There is a need to determine how the new groups correspond to the old groups. So to do this where something looks like a car in an old frame and something looks like a car in the current frame and that if the centers of these objects are close enough then they probably represent the same car. The system matches the groups that are closest to each other first, up to a given distance apart. If a group from the current frame isn't matched to any group from the previous frame then if the group has enough pixels in it, it will be marked as a new group and its entrance position will be noted. If a group from the current frame is matched to a group from the previous frame then the most recent location of the old group will be updated to reflect the current position. The result is that cars can be followed from frame to frame and their position and entrance times and positions are known. These characteristics are useful for calculating line length as previously mentioned.
Now that the end of the line is on the screen, how many car lengths is the vehicle line?
Determining the distance of the group in car-lengths. As part of the setup of the program, the user must provide a way for the program to know how long the line is at various points on the road for each preset camera position. This is done by the user by adding ‘milestones’ to the picture (by clicking the mouse on it) and associating a distance (in cars) with each milestone. If the program wants to know the line length at a given point on the screen, it will find the point on the line between milestones closest to the given point and estimate the distance by interpolating between the values of the milestones on either side of it.
The processing rate of the cars in each lane. The program processes video from a fixed-view camera that can see all four of the lanes, for example. The job of the program is to notify a visual basic program when a car goes through a processing area. The visual basic program then effectively divides the number of cars seen by the time period giving the rate of car passage through each lane.
Implementation. A user provides location input to the program by painting red on a screenshot of the car passage area. The user paints one red spot for each lane. The red spot is painted to be smaller than half the width of a car and is at a place where motion represents a car passing through the area. At any given time the program is in the ‘car’ or ‘not car’ state for each of the four lanes. Each of the four lanes starts in a ‘not car’ state. For each frame, if the motion in a given lane spot is greater than a maximum amount then the state is set to ‘car’. If the motion in a given lane spot is less than a minimum amount then the state for that lane is set to ‘not car’. The motion is calculated as the sum of the absolute values of pixel differences between successive frames of video in each lane spot. When the state changes from the ‘not car’ to ‘car’ for a given lane, the program notifies the visual basic program that a car was found at that lane.
Operation of the current embodiment uses two C++ programs to collect data. One C++ program processes image data from the traffic flow camera while the other processes end of line program. Data from both of these is sent to a single Visual Basic program which processes the data. Of course, as described above, there are a variety of embodiments which may be used to provide wait-time data at a boarder crossing. This particular embodiment, as reduced to function works like this:
C++ traffic Flow
As detailed below, VB does the following.
If Do_loop=true then ‘Do_loop is false first time through
If Last_Camera_Position < >CURRENT_CAM_POSITION Then
‘when Goto_here is called, it set's waiting for move=true. This allows the goto_here to send the camera to a position and because it knows how long it should take for the camera to get there, it stays in this sub until the camera has theoretically reached its target. As we exit the goto here sub, Waiting_for_move is set to false. This releases the following while loop. This means that this routine holds up here until the camera move is complete.
‘Data is arriving from the Find_End_Of_Line C++ program. While Allow_Data=false, this arriving data is ignored. This is due largely to the fact that the camera is probably moving or otherwise not ready to provide usable data. Here, we set Allow_Data=true so that we can start to build target data.
‘if HGP is >70 then there were stopped cars moving in the furthest part of the lane. This indicates that the lanes in the current view are full. To help confirm this, we want to make sure that no car arrived from the furthest part of the frame (LBP) and drove lower that 80% of the frame. If these conditions have occurred, then the lane viewed by the current camera view is most likely full of cars. The likelihood is so high, that we should not only report that the lane is full, but also take a sampling of the current texture value at this time and store it for future reference.
‘This is a little confusing, but that makes sense because we will enter this condition if the data is not consistent with our expectations. In the following, we check to see if there is blue that is lower than green. If this is the case, then a car drove in from the top (meaning the lane is at least not completely full). Set the current lane fullness value based on this. But, if green is higher then blue, then a few possibilities exist. If green is really high, then the lane is probably full of cars. If green is higher than blue and Green is still NOT very high (below 70%), then the lane might be full (but no cars have moved), or it might be empty and no cars have driven in from the top. In either case, we should rely on the latest texture information (neural) to help decide the current condition.
‘if there was high green
Else ‘maybe traffic is locked up and so we didn't get much good data and we should wait.
‘The following is designed to allow for longer wait time to asses conditions if current data is inconclusive. Waited for action is modified within the following loops and checked here. If we have waited more than 3 times, then that's enough. Go forward and make the decision.
‘This is the function that is called each time the C++ program sends data to the VB about end of lane conditions. This is called every 100 ms. Data is only used if Allow_Data is true. This is set in the main loop.
‘This sub gets data from the C++ program, parses it, and assigns data to variables.
‘this sub uses the data from the parsed C++ data to decide on blue and green percents. All subs are important, but this one is really important. Here, green and blue values are set following each delivery of data from the C++.
‘if the tracked object has moved an amount great enough (based on current zoom level) and it has done this moving in the allotted time, then we will consider it valid data. Movement that occurs over a long period of time can be “ripple texture” which occurs from transference from one vehicle to the next even if the vehicles are not really moving.
‘this is the final step for data after it arrives from the C++—except that it will be used after ALLOW Data=false. Then this all stops and the main loop uses this data.
‘While all of the above is happening, the other C++ program is collecting data on the flow rate of cars through the crossing. This sub is called each time the C++ gets data about flow.
‘if you click NO to virtually, then DO NOT DISPLAY. Data arrives here, but displays are not calculated here. Calculations and updates occur by calls from the timer . . . see below
‘flip data to account for lanes . . . this can be done in a set up box by assigning each lane a specific number. Here is hard coded, but in another embodiment it might not be. This would allow the program to output the same “lane numbers” as the boarder agents know the lanes as.
‘If the lane was not tripped in one second, then don't use the data. This protects from one car triggering multiple hits. We assume that if the C++ reads several cars in one second, then the threshold is off, and we assume that only one car went by. Since cars usually take 30 seconds or more, we can afford this buffer to protect from bad data.
‘Also display happens here for Total CPM . . .
‘but can be expanded here. We are building the CPS array here.
‘set Cars Per Second array to 0 for events that are more than 60 seconds old ‘this will make the array ready for providing information on the amount of cars ‘in the last minute.
‘we also call to display the data in both the lane text boxes and the totals.
‘this is where we calculate and display the total traffic flow.
‘ This is the sub that actually sends commands to camera and switcher to select a new camera position through camera movement and camera selection. This sends out an 8 byte string that is specific for one type of Pan/Tilt/Zoom camera. The program could be used with any PTZ camera by simply changing the commands sent out. Also, the system does not need a PTZ camera. Multiple fixed cameras can also be used.
SendOut1′ this function send bytes 0 through 8 to the comm. Port. This could also send to a winsock port so that another machine could send data out from another location.
This is the sub routine:
‘if the lable is green, then it is active. If it is active AND its flow rate is significantly more that 0 then assume the lane is active. In this case, reduce the value of this lane from the total.