Specifications
9
The first step is to detect occlusions. It is assumed that each object that enters the scene also exits 
the scene. If this assumption holds, any object track that either ends or begins within the object 
frame rather than at the edges of the frame can be assumed to have been occluded. Thus, the 
position of the first and last object of each object track is examined. If the first object of the track 
is near the image edge, the track is classified as having entered the scene. Otherwise, the track is 
classified as having been occluded. If the last object of the track is near the image edge, the track 
is classified as having left the scene. Otherwise, the track is classified as having been occluded 
before leaving the scene. 
Once the occluded object tracks have been identified, the ends of the object tracks are examined 
for potential matches across occlusions. Due to the occlusion, spatial position is not a reliable 
indicator of object identity. Therefore, the visual features of the object are used as a matching 
measure. For each end of the object track that is occluded, a comparison is made between the 
visual features of the object and the visual features of objects within λ frames. 
The SIFT algorithm (Lowe 1999) is employed to match visual features from object to object. 
This matching algorithm provides robust matching performance when the compared objects 
differ in position, rotation, scaling, and global changes in color and intensity. Although matching 
the visual features of objects is quite slow, it is more reliable than using position alone. Once 
links between object tracks have been established, the object tracks are reorganized to create a 
master object track list where each track uniquely describes a single scene object. 
4.5 Noise Filtering 
In order to minimize the effects of noise, temporal averaging is applied to the object tracks. A 
noncausal boxcar filter of size τ is run across the object center and bounding box coordinates 
separately. The filter size τ should be chosen based on the properties of the camera. It should be 
noted that a large filter can result in an underestimation of minimum and maximum velocities. 
5. DEPLOYMENT OF DIGITAL VIDEO ANALYSIS SYSTEM AT NON-SIGNALIZED 
INTERSECTIONS 
To obtain video that could be used to test the video analysis algorithms, the system was deployed 
to a total of five intersections in Ames, Iowa. Four non-signalized intersections on Bissell Road 
were chosen. The location of these intersections is shown on the map in Figure 5.1. The position 
and direction of the camera in the standard camera–intersection configuration is shown with a 
blue box and arrow. A video frame taken from each intersection is also shown. 
A single, high-speed rural intersection at U.S. 69 and 190th
was also chosen. A collision diagram 
and video frame from the recording are shown in Figure 5.2. 










