Vehicle Segmentation and Tracking in the Presence of Occlusions
|
|
|
- Eric Washington
- 9 years ago
- Views:
Transcription
1 Word count Table + 4 Figures TRB Paper Number: Vehicle Segmentation and Tracking in the Presence of Occlusions Neeraj K. Kanhere Department of Electrical and Computer Engineering 207-A Riggs Hall Clemson University, Clemson, SC Phone: (864) , FAX: (864) [email protected] Stanley T. Birchfield Department of Electrical and Computer Engineering 207-A Riggs Hall Clemson University, Clemson, SC Phone: (864) , FAX: (864) [email protected] Wayne A. Sarasua Department of Civil Engineering 312 Lowry Hall, Box Clemson University, Clemson, SC Phone: (864) , FAX: (864) [email protected] Nov 14, 2005
2 ABSTRACT A novel method is presented for automatically visually monitoring a highway when the camera is relatively low to the ground and on the side of the road. In such a case, occlusion and the perspective effects due to the heights of the vehicles cannot be ignored. Using a single camera, the system automatically detects and tracks feature points throughout the image sequence, estimates the 3D world coordinates of the points on the vehicles, and groups those points together in order to segment and track the individual vehicles. Experimental results on different highways demonstrate the ability of the system to segment and track vehicles even in the presence of severe occlusion and significant perspective changes. By handling perspective effects, the approach overcomes a limitation of commercially available machine vision-based traffic monitoring systems that are used in many intelligent transportation systems (ITS) applications. The researchers are targeting this system as a step toward a next generation ITS sensor for automated traffic analysis. INTRODUCTION Traffic counts, speed and vehicle classification are fundamental data for a variety of transportation projects ranging from transportation planning to modern intelligent transportation systems (ITS). Most ITS applications are designed using readily available technology (e.g., sensors and communication), such as the inductive loop detector. Other sensing technologies include radar, infrared (IR), lasers, ultrasonic sensors and magnetometers. In (1) and (2), state of the art traffic sensing technologies are discussed along with the algorithms used to support traffic management functions. Since the late 1980s, video imaging detection systems have been marketed in the U.S. and elsewhere. One of the most popular video based traffic counting systems is Autoscope, which is currently distributed by Econolite. Autoscope uses high-angle cameras to count traffic by detecting vehicles passing digital sensors. As a pattern passes over the digital detector, the change is recognized and a vehicle is counted. The length of time that this change takes place can be translated into speed estimates. Autoscope includes significant built-in heuristics to differentiate between shadows and vehicles in various weather conditions. The accuracy of Autoscope and other commercially based video detection systems is compromised if the cameras are mounted too low or have poor perspective views of traffic. If the camera is not mounted high enough, a vehicle s image will spill over onto neighboring lanes, resulting in double counting. A more promising approach to traffic monitoring is to track vehicles over time throughout an image sequence. This approach yields the trajectories of the vehicles, which are necessary for applications such as traffic flow modeling and counting turn movements. In this paper we present an automatic technique for detecting and tracking vehicles at a low angle, even in the presence of severe occlusion, to obtain those trajectories. Although the problem of tracking is much more difficult at a low angle because of perspective effects, a solution to this problem would be greatly beneficial for data-collecting applications in which the permanent infrastructure necessary for high-angle cameras is not feasible. Research in Vehicle Tracking Over the years researchers in computer vision have proposed various solutions to the automated tracking problem. These approaches can be classified as follows:
3 Blob Tracking. In this approach, a background model is generated for the scene. For each input image frame, the absolute difference between the input image and the background image is processed to extract foreground blobs corresponding to the vehicles on the road. Variations of this approach have been proposed in (3, 4, 5). Gupte et al. (3) perform vehicle tracking at two levels: the region level and the vehicle level, and they formulate the association problem between regions in consecutive frames as the problem of finding a maximally weighted graph. These algorithms have difficulty handling shadows, occlusions, and large vehicles (e.g., trucks, and trailers), all of which cause multiple vehicles to appear as a single region. Active Contour Tracking. A closely related approach to blob tracking is based on tracking active contours (also known as snakes) representing the boundary of an object. Vehicle tracking using active contour models has been reported by Koller et al. (6), in which the contour is initialized using a background difference image and tracked using intensity and motion boundaries. Tracking is achieved using two Kalman filters, one for estimating the affine motion parameters, and the other for estimating the shape of the contour. An explicit occlusion detection step is performed by intersecting the depth ordered regions associated with the objects. The intersection is excluded in the shape and motion estimation. As with the previous technique, results are shown on image sequences without shadows or severe occlusions, and the algorithm is limited to tracking cars. 3D-Model Based Tracking. Tracking vehicles using three-dimensional models has been studied by several research groups (7, 8, 9, 10). Some of these approaches assume an aerial view of the scene which virtually eliminates all occlusions (10) and match the three-dimensional wireframe models for different types of vehicles to edges detected in the image. In (9), a single vehicle is successfully tracked through a partial occlusion, but its applicability to congested traffic scenes has not been demonstrated. Markov Random Field Tracking. An algorithm for segmenting and tracking vehicles in lowangle frontal sequences has been proposed by Kamijo et al. (11). In their work, the image is divided into pixel blocks, and a spatiotemporal Markov random field (ST-MRF) is used to update an object map using the current and previous image. One drawback of the algorithm is that it does not yield 3D information about vehicle trajectories in the world coordinate system. In addition, in order to achieve accurate results the images in the sequence are processed in reverse order to ensure that vehicles recede from the camera. The accuracy decreases by a factor of two when the sequence is not processed in reverse, thus making the algorithm unsuitable for on-line processing when time-critical results are required. Feature Tracking. In this approach, instead of tracking a whole object, feature points on an object are tracked. The method is useful in situations of partial occlusions, where only a portion of an object is visible. The task of tracking multiple objects then becomes the task of grouping the tracked features based on one or more similarity criteria. Beymer et al.(12,13) have proposed a feature tracking based approach for traffic monitoring applications. In their approach, point features are tracked throughout the detection zone specified in the image. Feature points which are tracked successfully from the entry region to the exit region are considered in the process of grouping. Grouping is done by constructing a graph over time, with vertices representing subfeature tracks and edges representing the grouping relationships between tracks. The algorithm 3
4 was implemented on multi-processor digital signal processing (DSP) board for real-time performance. Results were reported for day and night sequences with varying levels of traffic congestion. Color and Pattern-Based Tracking. Chachich et al. (14) use color signatures in quantized RGB space for tracking vehicles. In this work, vehicle detections are associated with each other by combining color information with driver behavior characteristics and arrival likelihood. In addition to tracking vehicles from a stationary camera, a pattern-recognition based approach to on-road vehicle detection has been studied in (15). The camera is placed inside a vehicle looking straight ahead, and vehicle detection is treated as a pattern classification problem using support vector machines (SVMs). APPROACH In comparison with the research just described, the novelty of the proposed approach lies in the estimation of the 3D coordinates of feature points in order to track vehicles at low angles, when a single homography is insufficient. This research contains several extensions over our previous work (16): a technique for associating tracks is proposed to facilitate long-term tracking, a method we call incremental normalized cuts is introduced to improve the quality of the segmentation, and a single perspective mapping is used instead of a multi-level homography, which increases robustness by taking advantage of the inherent constraints of the data. The sequence is assumed to be taken from a single grayscale camera pointing at the road from the side. The task of segmenting and tracking vehicles in cluttered scenes is formulated as a feature tracking and grouping problem. Feature points are tracked in the image sequence, followed by estimation of the 3D world coordinates for those points, which are then grouped using a segmentation algorithm. The novelty of this work lies primarily in the technique for estimating the 3D coordinates from a single camera. Offline Calibration Calibration is required to estimate the 3D world coordinates for corresponding 2D points in the image. It should be emphasized that the calibration process described below is for a single camera and does not require knowledge about the camera specifications such as focal length or sensor dimensions, which makes it possible to process pre-recorded sequences captured from unknown cameras. The only information that is needed is six or more point correspondences. We assume a pinhole camera model exhibiting perspective projection. The general relationship between an object point measured with respect to a user-selected world coordinate system and its image plane point is denoted by a 3 x 4 homogeneous transformation matrix (17). This matrix will be referred to as the camera calibration matrix C. These equations define a mapping from the world coordinates (x,y,z) to the image coordinates (u,v) as described in (18). [1] [2] 4
5 Calibration Process The camera calibration matrix C can be computed from the correspondence of 2D image points with the 3D coordinates of the associated world points. Each correspondence yields two equations of the form [1] and [2]. Six or more correspondences from a non-degenerate configuration lead to an over-determined system which can be solved using a standard least squares technique. The offline calibration process depends upon the user-specified point correspondences for the calibration process. Although it would be ideal to have known markers placed at known locations in the scene, in practice this is often not feasible (e.g., on prerecorded data). In addition, as we will show, sufficient accuracy can be obtained simply by using standard specifications such as the width of a lane and the length of a truck, because the algorithm is not sensitive to slight errors in calibration. 1 We have developed a calibration application that is straightforward, simple to learn and use, and provides adequate results. Our calibration tool is similar to that developed by Gupte et al. (3), except that their system finds a planar mapping between the points on the road and the image points, whereas ours estimates a full perspective mapping, leading to 3D coordinates. An example of the calibration process is shown in Figure 1. First, the user places a marker across the width of the road and perpendicular to the lane markings as shown in Figure 1 (a). With the marker position unchanged, the sequence is advanced till the rear end of the truck appears to align with the marker position on the ground. A new marker is placed to align with the height of the truck (b). In the same frame a marker is placed on the ground to align with the front end of the truck (c). Once again, the sequence is advanced till the marker placed on the ground in (c) appears to align with the rear end of the truck. This is shown in (d). For the same frame, the marker is realigned with the front end of the truck as shown in (e). A new marker is placed across the width of the road (f). One more time, the sequence is advanced for the new marker to appear aligning with the rear end of the truck. An additional marker is placed as shown in (g) in such a way that it appears to be aligned with the height of the truck. The result looks as shown in (h). Using the dimensions of a known type of vehicle, lane width (e.g. 12 feet on an interstate), and number of lanes yields an approximate method for estimating the world coordinates of the control points. The calibration process is simple and takes only a couple of minutes to complete. Backprojections The imaging process maps a point in three dimensional space into a two dimensional image plane. The loss of dimension results in a non-invertible mapping. The calibration parameters for the camera and the image coordinates of a single point determine a ray in space passing through the optical center and the unknown point in the world. Rearranging equations [1] and [2] yields equations for two planes in 3D space. [3] [4] 1. Three untrained users were asked to perform the offline calibration as described. For the standard deviation (in random direction) of three pixels for each control point, the segmentation accuracy was observed to fall by 5 %. 5
6 FIGURE 1 Offline calibration process using our calibration application. (a) (b) (c) (d) (e) (f) (g) (h) (a) (h) show steps in offline calibration as described in the text. 6
7 The intersection of these two planes is the ray in 3D passing through the point in the world that is projected onto the image plane. Because there are two equations and three unknowns, the problem is under-constrained. If we know either x, y, or z we can solve for the other two world coordinates using the image coordinates and C. In the sections to follow, we present a simple technique to overcome the under-constrained nature of the problem, enabling us to solve for all three world coordinates of a point. Processing a Block of Frames In the algorithm proposed by Beymer et al. (12) and Malik et al. (13), the point features tracked successfully from the entry region to the exit region are considered in the grouping step, which does not pose a problem when the camera is placed at a high vantage point looking down on the road. In our scenario, however, frequent occlusions and appearance changes (as vehicles approach the camera) result in the loss of a large number of features. As a result, the number of features that are tracked for the whole extent of the detection zone is not enough to achieve useful results. This problem is overcome by processing a block of consecutive image frames (typically twenty frames per block) in order to segment the vehicles in each block, and then to associate the segmented vehicles between the successive blocks. Features are tracked throughout a block of F image frames, overlapping with the previous block by N frames. The length of a block is determined by the average speed of the vehicles and the placement of the camera with respect to the road. If the number of frames in a block is too small, then although a large number of features will be tracked successfully throughout the frames in the block, the motion information will be insufficient for effective segmentation. On the other hand, using more frames in a frame-block will yield more reliable motion information at the expense of losing important features. The proposed algorithm relies on human judgment to balance between these tradeoffs. The steps described in the following sections are performed on the features tracked over a single block. Tracking Features Feature points are automatically selected and tracked using the Kanade-Lucas-Tomasi (KLT) feature tracker (19) based on the algorithm proposed in (20). Background Subtraction Background subtraction is a simple and effective technique for extracting foreground objects from a scene. The process of background subtraction involves initializing and maintaining a background model of the scene off-line. At run time, the estimated background image is subtracted from the image frame being processed, followed by thresholding the absolute value of the difference image, along with morphological processing to reduce the effects of noise, to yield foreground blobs. A review of several background modeling techniques is presented in (21). For the scope of this research, the median filtering technique was chosen for its simplicity and effectiveness. The median filter belongs to a general class of rank filters. It is frequently used in image processing for removing noise in an image. For background modeling, we perform one dimensional median filtering in the time domain. For each pixel in the background image, the median value is selected from the set of values observed at the same pixel location in the previous n frames. 7
8 FIGURE 2 Road projection. (a) (b) (c) Projecting a feature on the road surface in the image for estimating its height. (a) the first frame of a frame-block, with a feature point on a truck; (b) the foreground mask obtained by background subtraction, with the projection of the feature down to the bottom of the blob; (c) the corresponding 3D illustration, in which p and g are image points corresponding to P and G respectively, and O is the optical center of the camera. 8
9 Selecting Stable Features It was shown earlier that the 3D coordinates of a world point can be estimated using its corresponding image coordinates, the calibration parameters, and at least one component of the world coordinates. A simple technique to achieve the same is presented here which involves finding the vertical projection of a point on the road surface in the image. The foreground mask generated in the previous step is used to find the projection as shown in Figure 2. P=[x y z] is a vector of world coordinates corresponding to the point p=[u, v] in the image. O is the optical center of the camera. G=[x y z g ] is a 3 x 1 vector containing world coordinates of ground projection of P. From equations [3] and [4] it follows that [5] Since G lies on the ground (or at least sufficiently close), we can compute its 3D coordinates by assuming z g = 0 (corresponding to the road plane) in the above equation. P and G have the same (x, y) coordinates. Now, we know the image coordinates p of the world point P along with its (x, y) coordinates, and the camera calibration matrix C. Substituting these values into equations [1] and [2], we solve for z: [6] For this technique to work, a simple box-model for the vehicles is assumed. A vehicle is modeled using five orthogonal, rectangular surfaces as shown in Figure 2 (c). Two such models have been used to represent cars and heavy vehicles. Dimensions of the corresponding models are computed using the calibration information (in proportion to the lane width). In addition, the technique assumes that the world point G is directly below P. In practice these assumptions generally hold, because our threshold for the maximum height of a stable feature is 0.8 m, below which vehicles are well-modeled as boxes with vertical sides. Shadows tend not to cause a problem because features are rarely detected on the road itself due to lack of texture; as a result, shadows usually cause the height prediction of a feature to exceed the threshold, causing the 9
10 feature to be ignored. In a similar manner, although occlusions cause the point G to be estimated incorrectly, the estimated height will be above the threshold, since the problem only occurs when the feature is higher than the hood (or trunk) of the occluding vehicle. In addition, features rarely belong to the top surface, primarily due to insufficient texture and a relatively small projection in the image. Even when the assumptions are violated, the segmentation error is just 7% when the image height of G is perturbed by an error with standard deviation of 5 pixels, which we computed using simulations on equations [5] and [6]. Moreover, it is important to keep in mind that our technique only requires that at least one feature on a vehicle satisfy the assumption of lying on one of these four surfaces, so that multiple features with erroneous values do not pose a problem. After estimating height of all the features using this technique, features which are close to the road surface are selected as stable features. In our previous work (16), stable features were selected based on an additional condition of low variance in height estimation for each frame of the block, but we have since found that removing this criterion reduces the number of computations without any noticeable degradation in the segmentation results. World Coordinates from Multiple Frames Factors like occlusion and shadows introduce significant error in the height estimates of many of the feature points obtained using the technique presented in the previous section, but as long as an accurate height is obtained for at least one feature per vehicle, the coordinates of that point can be used to estimate the world coordinates of the rest of the features on the vehicle. To accomplish this task, we employ rigidity constraints and a translational motion model. A line in 3D can be represented in a parametric form as: P = P R + [P H - P R ] where P H and P R are 3 x 1 vectors representing any two points on the line, and is a scalar which defines location of a point along the line. The above representation simplifies the mathematical analysis to follow. As shown in Figure 3, we consider two points, P and Q which undergo a translational motion from P 0, Q 0 in initial frame to P t, Q t after duration t. If Q is one of the stable features, then its real world coordinates are known. As derived in (18), can be solved as [7] where represents translational motion for the respective point. 10
11 FIGURE 3 Estimating world coordinates using rigid motion. Coordinates of P are unknown. Q is a stable feature point with known world coordinates. 11
12 Estimates for the world coordinates of non-stable feature points can be computed with respect to each stable feature point using the two equations above. Among all the estimates, we select the estimate that minimizes the weighted sum of Euclidean distance in (x, y) and squared trajectory error over all s frames in the block. [8] Affinity Matrix and Normalized Cuts We form the affinity matrix composed of three components, namely, the 3D Euclidean distance in world coordinates, the difference in trajectory and the background content measure. Euclidean distance and background content are measured using the coordinates of feature points in the first frame of the block. The details of the affinity matrix formulation can be found in (18). In Shi et al.(22, 23), it is mentioned that for the normalized cut algorithm to be computationally efficient, the affinity matrix (also called the weight matrix) should be sparse. Experiments were performed using sparse affinity matrices as well, i.e. using only local edge connections for a feature, but it was observed that using full matrices produced better results without a significant increase in the computing time. 2 Grouping With Incremental Normalized Cuts Based on Shi and Malik s normalized cuts, we have developed a grouping procedure that we call incremental normalized cuts for segmenting a set of features into meaningful groups. The process involves applying normalized cuts to the affinity matrix with increasing number of cuts until a valid group is found. The corresponding entries for the features in the detected group are removed from the affinity matrix and the process is repeated. The key part of this step is to use the calibration information to accept or reject a feature group based on following three criteria: The group has a minimum number of required features. The centroid (in 3D coordinates) lies inside the detection zone. Dimensions of the group are within a valid range. Using incremental normalized cuts avoids explicitly specifying the number of cuts required for a good segmentation, which depends upon quantities such as the number of features and the number of vehicles. Further details about incremental normalized cuts can be found in (15). 2. For sparse affinity matrices using 70, 60 and 50 pixel neighborhood, the computational time saved was found to be 3.8%, 4.6%, 5.8% with the drop in accuracy of 14%, 22% and 31% respectively. 12
13 Correspondence Between Frame Blocks In the previous sections, we looked at how to track feature points through a block of F frames, estimate the corresponding world coordinates, and group features using incremental normalized cuts. This procedure is applied to all the blocks of frames (overlapping by F-1 frames), using the same set of parameters. Long-term tracking requires us to compute the correspondence between the vehicles detected in consecutive frame blocks. In this section we describe our approach for correspondence. Consider two consecutive frame-blocks A and B overlapping by F-1 frames. Let A denote the feature groups segmented in a frame-block A, and let B denote the feature groups segmented in frame-block B. An undirected graph is formed with the segmented feature groups in both frame blocks as nodes and the number of common feature points shared by a pair of groups as the weight of an edge connecting the respective nodes. If a group in the previous block shares features with only a single group in the current block, then we call this a one-to-one unique correspondence. A group in A sharing features with more than one group from B indicates splitting. Similarly, two or more groups in A sharing common features with a group in B indicates merging. A group in A having no association is considered a missing event, and a group in B having no association with any of the groups in the previous block is considered as a new detection. If a group is associated with a one-to-one correspondence over consecutive blocks, it is labeled as a reliable group. If a group is missing for consecutive blocks, it is labeled as inactive. During initialization, each group in the first frame-block is assigned a unique label. For each consecutive frame-block, a graph is constructed as mentioned above. To neglect minor segmentation errors, all the edges having weights are removed. This is followed by searching for the unique one-to-one correspondences between the groups of previous and current frameblocks. Groups of the current block having unique correspondences are assigned the labels of respective groups in the previous block. After processing all the unique associations, the graph is searched for splits. For a split event, the edge with maximum weight is used for correspondence and the remaining edges are removed. Merge events are handled the same way. Groups in A which are no longer connected to any of the groups in B and are labeled as reliable, are declared missing. Groups in B which are not connected with any of the groups in A are declared as new detections. Each group that is declared as a new detection is matched with all the active missing groups to find a possible correspondence. If a correspondence with missing groups is not found, the group is assigned a new label. EXPERIMENTAL RESULTS The algorithm was tested on four grayscale image sequences, each containing 1200 frames captured at 30 frames per second. The camera was placed on an approximately 10 m pole on the side of the road. The sequences were digitized at 640 x 480 resolution. No preprocessing was done to suppress shadows or to stabilize occasional camera jitter. For each sequence, offline camera calibration was performed once, as explained earlier. The first sequence was captured on a clear day. Vehicles are traveling in three lanes and there are moderate moving shadows. The second sequence shows a four-lane highway with the last lane blocked for maintenance work. The lane closure results in slow moving traffic with vehicles traveling close to each other. The sequence was captured during data collection for studying the effect of a workzone on freeway traffic (24). The third sequence was found to be even more challenging because the vehicles cast long shadows, making the process of segmentation based of size constraints harder. In Figure 4(b) the small vehicle traveling next to a trailer is correctly 13
14 segmented and tracked demonstrating the system s ability to handle severe partial occlusions. One simple method was tested for detecting and removing groups that belong to shadows as shown in Figure 4 (f). If the height of a group is below a threshold value, it is classified as a shadow group and is discarded. Having zero as the threshold (which is theoretically correct) does not yield the desired results, since the estimation process is based on the approximate calibration along with simple assumption for the shape of vehicles resulting in height estimation error. If the threshold is set higher, more shadow-groups are detected and discarded at the cost occasionally detecting a small vehicle (e.g. a compact sports car) as a shadow group. The fourth and the last sequence was also captured for the workzone study, and the vehicles are traveling close to each other at low speeds. Because of the presence of fog, the images in this sequence are noisy compared with those in the other sequences. A quantitative assessment of the results on all the sequences is presented in Table 1. The algorithm is able to detect all of the cars and trucks in Sequence 4, and nearly all of them in Sequence 1. Sequences 2 and 3 proved to be more challenging, but the system nevertheless detects over 90% of the cars, and over 85% of the trucks. The detection rate for occluded vehicles ranges from 55% to 84%, showing that significant progress over previous systems has been made, while pointing to the need for further research. The false positive rate is generally 5 to 10% of the total number of vehicles and is usually caused by shadows. Video sequences demonstrating the performance of the system can be found at TABLE 1 Accuracy of the algorithm on the test sequences Sequence C T O DC DT DO FP (98%) 9 (100%) 16 (84%) (96%) 7 (88%) 11 (65%) (93%) 6 (86%) 6 (55%) (100%) 3 (100%) 6 (67%) 2 The columns show the sequence, number of cars (C), number of trucks (T), number of vehicles among the cars and trucks that were significantly occluded (O), number of cars tracked (DC), number of trucks tracked (DT), number of occluded vehicles detected and tracked (DO) and number of false detections (FP) respectively. 14
15 FIGURE 4 Experimental Results. (a) Sequence 1, frame 35 (b) Sequence 1, frame 592 (c) Sequence 2, frame 330 (d) Sequence 3, frame 311 (e) Sequence 4, frame 415 (f) removing shadow groups 15
16 CONCLUSIONS AND RECOMMENDATIONS Most approaches to segmenting and tracking vehicles from a stationary camera assume that the camera is high above the ground, thus simplifying the problem. A technique has been presented in this paper that works when the camera is relatively at a low angle with respect to the ground and/or is on the side of the road, in which case occlusions are more frequent. In such a situation, the planar motion assumption for vehicles is violated, especially in case of heavy vehicles like trailers. The approach proposed is based upon grouping tracked features using a segmentation algorithm called incremental normalized cuts, which is a slight variation of the popular normalized cuts algorithm. A novel part of the technique is the estimation of the 3D world coordinates of features using a combination of background subtraction, offline camera calibration (for a single camera), and rigidity constraints under translational motion. Experimental results on real sequences show the ability of the algorithm to handle the low-angle situation, including severe occlusion. This is a significant achievement in automated vehicle detection over commercially available systems. With further development, the proposed approach may lead to a next generation ITS sensor as well as an automated turn movement counter for use in conducting various traffic studies. In this later scenario, a camera mounted on a tripod set back from an intersection can collect video that can be digitally post-processed to determine turn-movement counts. Some of the aspects of the proposed algorithm need further analysis and improvement. A simple approach has been adopted for associating the results between the frame-blocks which is based solely upon the number of common features. Using the spatial proximity, color, and motion information will help in making a more robust association. In addition, future research should be aimed at handling low-angle image sequences taken at intersections, where the resulting vehicle trajectories are more complicated than those on highways. 16
17 References 1. Lawrence A. Klein, Sensor Technologies and Data Requirements for ITS, Boston: Artech House, Dan Middleton, Deepak Gopalakrishna, and Mala Raman. Advances in traffic data collection and management (white paper), January (Accessed on June 23, 2005) 3. S. Gupte, O. Masoud, R. F. K. Martin, and N. P. Papanikolopoulos. Detection and classification of vehicles. IEEE Transactions on Intelligent Transportation Systems, 3(1):37 47, March D. Magee. Tracking multiple vehicles using foreground, background and motion models. In Proceedings of ECCV Workshop on Statistical Methods in Video Processing, (2002). 5. D. Daily, F.W. Cathy, and S. Pumrin. An algorithm to estimate mean traffic speed using uncalibrated cameras. In IEEE Conference for Intelligent Transportation Systems, pages , D. Koller, J. Weber, and J. Malik. Robust multiple car tracking with occlusion reasoning. In European Conference on Computer Vision, pages , D. Koller, K Dandilis, and H. H. Nagel. Model based object tracking in monocular image sequences of road traffic scenes. International Journal of Computer Vision, 10(3): , M. Haag and H. Nagel. Combination of edge element and optical flow estimate for 3Dmodel-based vehicle tracking in traffic image sequences. International Journal of Computer Vision, 35(3): , Dec J. M. Ferryman, A. D. Worrall, and S. J. Maybank. Learning enhanced 3d models for vehicle tracking. In British Machine Vision Conference, pages , C. Schlosser, J. Reitberger, and S. Hinz. Automatic car detection in high resolution urban scenes based on an adaptive 3D-model. In EEE/ISPRS Joint Workshop on Remote Sensing and Data Fusion over Urban Areas, Berlin, pages , S. Kamijo, K. Ikeuchi, and M. Sakauchi. Vehicle tracking in low-angle and front view images based on spatio-temporal markov random fields. In Proceedings of the 8th World Congress on Intelligent Transportation Systems (ITS), D. Beymer, P. McLauchlan, B. Coifman, and J. Malik. A real time computer vision system for measuring traffic parameters. In IEEE Conference on Computer Vision and Pattern Recognition, pages ,
18 13. Jitendra Malik and Stuart Russel. Final Report for Traffic Surveillance and Detection Technology Development New Traffic Sensor Technology, December 19, (Accessed on October 21, 2005). 14. Chachich, A. A. Pau, A. Barber, K. Kennedy, E. Oleiniczak, J. Hackney, Q. Sun, E. Mireles, "Traffic sensor using a color vision method," Proceedings of the International Society for Optical Engineering, Vol. 2902, January, 1997, pp Zehang Sun, George Bebis and Ronald Miller. Improving the Performance of On-Road Vehicle Detection by Combining Gabor and Wavelet Features. In Proceedings of the 2002 IEEE International Conference on Intelligent Transportation Systems, September Neeraj K. Kanhere, Shrinivas J. Pundlik, and Stanley T. Birchfield. Vehicle segmentation and tracking from a low-angle off-axis camera. In IEEE Conference on Computer Vision and Pattern Recognition, pages , Robert J. Schalkoff. Digital Image Processing and Computer Vision. New York: John Wiley & Sons, Inc., first edition, Neeraj K. Kanhere, Vehicle Segmentation and tracking from a low-angle off-axis camera, Master s thesis, August 2005, Clemson University. (Accessed on July 20, 2005) 19. S. Birchfield. KLT: An implementation of the Kanade-Lucas-Tomasi feature tracker. stb/klt/ (Accessed on January 15, 2004) 20. Carlo Tomasi and Takeo Kanade. Detection and tracking of point features. Technical Report CMU-CS , Carnegie Mellon University, Apr S. C. Cheung and Chandrika Kamath. Robust techniques for background subtraction in urban traffic video. In Proceedings of Electronic Imaging: Visual Communications and Image Processing, J. Shi and J. Malik. Normalized cuts and image segmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 22(8): , Aug Jianbo Shi and Jitendra Malik. Motion segmentation and tracking using normalized cuts. In IEEE International Conference on Computer Vision, pages , Wayne Sarasua. Traffic Impacts of Short Term Interstate Work Zone Lane Closures: The South Carolina Experience. (Accessed on June 23, 2005) 18
Vision based Vehicle Tracking using a high angle camera
Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu [email protected] [email protected] Abstract A vehicle tracking and grouping algorithm is presented in this work
A feature-based tracking algorithm for vehicles in intersections
A feature-based tracking algorithm for vehicles in intersections Nicolas Saunier and Tarek Sayed Departement of Civil Engineering, University of British Columbia 6250 Applied Science Lane, Vancouver BC
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA N. Zarrinpanjeh a, F. Dadrassjavan b, H. Fattahi c * a Islamic Azad University of Qazvin - [email protected]
EFFICIENT VEHICLE TRACKING AND CLASSIFICATION FOR AN AUTOMATED TRAFFIC SURVEILLANCE SYSTEM
EFFICIENT VEHICLE TRACKING AND CLASSIFICATION FOR AN AUTOMATED TRAFFIC SURVEILLANCE SYSTEM Amol Ambardekar, Mircea Nicolescu, and George Bebis Department of Computer Science and Engineering University
Neural Network based Vehicle Classification for Intelligent Traffic Control
Neural Network based Vehicle Classification for Intelligent Traffic Control Saeid Fazli 1, Shahram Mohammadi 2, Morteza Rahmani 3 1,2,3 Electrical Engineering Department, Zanjan University, Zanjan, IRAN
ROBUST VEHICLE TRACKING IN VIDEO IMAGES BEING TAKEN FROM A HELICOPTER
ROBUST VEHICLE TRACKING IN VIDEO IMAGES BEING TAKEN FROM A HELICOPTER Fatemeh Karimi Nejadasl, Ben G.H. Gorte, and Serge P. Hoogendoorn Institute of Earth Observation and Space System, Delft University
An Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network
Proceedings of the 8th WSEAS Int. Conf. on ARTIFICIAL INTELLIGENCE, KNOWLEDGE ENGINEERING & DATA BASES (AIKED '9) ISSN: 179-519 435 ISBN: 978-96-474-51-2 An Energy-Based Vehicle Tracking System using Principal
3D Vehicle Extraction and Tracking from Multiple Viewpoints for Traffic Monitoring by using Probability Fusion Map
Electronic Letters on Computer Vision and Image Analysis 7(2):110-119, 2008 3D Vehicle Extraction and Tracking from Multiple Viewpoints for Traffic Monitoring by using Probability Fusion Map Zhencheng
VISION-BASED DETECTION, TRACKING AND CLASSIFICATION OF VEHICLES USING STABLE FEATURES WITH AUTOMATIC CAMERA CALIBRATION
VISION-BASED DETECTION, TRACKING AND CLASSIFICATION OF VEHICLES USING STABLE FEATURES WITH AUTOMATIC CAMERA CALIBRATION A Dissertation Presented to the Graduate School of Clemson University In Partial
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING
REAL TIME TRAFFIC LIGHT CONTROL USING IMAGE PROCESSING Ms.PALLAVI CHOUDEKAR Ajay Kumar Garg Engineering College, Department of electrical and electronics Ms.SAYANTI BANERJEE Ajay Kumar Garg Engineering
Automatic Traffic Estimation Using Image Processing
Automatic Traffic Estimation Using Image Processing Pejman Niksaz Science &Research Branch, Azad University of Yazd, Iran [email protected] Abstract As we know the population of city and number of
Automatic Labeling of Lane Markings for Autonomous Vehicles
Automatic Labeling of Lane Markings for Autonomous Vehicles Jeffrey Kiske Stanford University 450 Serra Mall, Stanford, CA 94305 [email protected] 1. Introduction As autonomous vehicles become more popular,
Human Behavior Analysis in Intelligent Retail Environments
Human Behavior Analysis in Intelligent Retail Environments Andrea Ascani, Emanuele Frontoni, Adriano Mancini, Primo Zingaretti 1 D.I.I.G.A., Università Politecnica delle Marche, Ancona - Italy, {ascani,
Tracking Moving Objects In Video Sequences Yiwei Wang, Robert E. Van Dyck, and John F. Doherty Department of Electrical Engineering The Pennsylvania State University University Park, PA16802 Abstract{Object
Traffic Monitoring Systems. Technology and sensors
Traffic Monitoring Systems Technology and sensors Technology Inductive loops Cameras Lidar/Ladar and laser Radar GPS etc Inductive loops Inductive loops signals Inductive loop sensor The inductance signal
Detection and Recognition of Mixed Traffic for Driver Assistance System
Detection and Recognition of Mixed Traffic for Driver Assistance System Pradnya Meshram 1, Prof. S.S. Wankhede 2 1 Scholar, Department of Electronics Engineering, G.H.Raisoni College of Engineering, Digdoh
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS Norbert Buch 1, Mark Cracknell 2, James Orwell 1 and Sergio A. Velastin 1 1. Kingston University, Penrhyn Road, Kingston upon Thames, KT1 2EE,
False alarm in outdoor environments
Accepted 1.0 Savantic letter 1(6) False alarm in outdoor environments Accepted 1.0 Savantic letter 2(6) Table of contents Revision history 3 References 3 1 Introduction 4 2 Pre-processing 4 3 Detection,
FLEXSYS Motion-based Traffic Analysis and Incident Detection
FLEXSYS Motion-based Traffic Analysis and Incident Detection Authors: Lixin Yang and Hichem Sahli, IBBT/VUB-ETRO Contents.1 Introduction......................................... 1.2 Traffic flow modelling
Multi-view Intelligent Vehicle Surveillance System
Multi-view Intelligent Vehicle Surveillance System S. Denman, C. Fookes, J. Cook, C. Davoren, A. Mamic, G. Farquharson, D. Chen, B. Chen and S. Sridharan Image and Video Research Laboratory Queensland
Edge tracking for motion segmentation and depth ordering
Edge tracking for motion segmentation and depth ordering P. Smith, T. Drummond and R. Cipolla Department of Engineering University of Cambridge Cambridge CB2 1PZ,UK {pas1001 twd20 cipolla}@eng.cam.ac.uk
Object tracking in video scenes
A Seminar On Object tracking in video scenes Presented by Alok K. Watve M.Tech. IT 1st year Indian Institue of Technology, Kharagpur Under the guidance of Dr. Shamik Sural Assistant Professor School of
Method for Traffic Flow Estimation using Ondashboard
Method for Traffic Flow Estimation using Ondashboard Camera Image Kohei Arai Graduate School of Science and Engineering Saga University Saga, Japan Steven Ray Sentinuwo Department of Electrical Engineering
Vehicle Tracking System Robust to Changes in Environmental Conditions
INORMATION & COMMUNICATIONS Vehicle Tracking System Robust to Changes in Environmental Conditions Yasuo OGIUCHI*, Masakatsu HIGASHIKUBO, Kenji NISHIDA and Takio KURITA Driving Safety Support Systems (DSSS)
CCTV - Video Analytics for Traffic Management
CCTV - Video Analytics for Traffic Management Index Purpose Description Relevance for Large Scale Events Technologies Impacts Integration potential Implementation Best Cases and Examples 1 of 12 Purpose
Colorado School of Mines Computer Vision Professor William Hoff
Professor William Hoff Dept of Electrical Engineering &Computer Science http://inside.mines.edu/~whoff/ 1 Introduction to 2 What is? A process that produces from images of the external world a description
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
Practical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006
Practical Tour of Visual tracking David Fleet and Allan Jepson January, 2006 Designing a Visual Tracker: What is the state? pose and motion (position, velocity, acceleration, ) shape (size, deformation,
Real time vehicle detection and tracking on multiple lanes
Real time vehicle detection and tracking on multiple lanes Kristian Kovačić Edouard Ivanjko Hrvoje Gold Department of Intelligent Transportation Systems Faculty of Transport and Traffic Sciences University
Static Environment Recognition Using Omni-camera from a Moving Vehicle
Static Environment Recognition Using Omni-camera from a Moving Vehicle Teruko Yata, Chuck Thorpe Frank Dellaert The Robotics Institute Carnegie Mellon University Pittsburgh, PA 15213 USA College of Computing
Speed Performance Improvement of Vehicle Blob Tracking System
Speed Performance Improvement of Vehicle Blob Tracking System Sung Chun Lee and Ram Nevatia University of Southern California, Los Angeles, CA 90089, USA [email protected], [email protected] Abstract. A speed
Traffic Flow Monitoring in Crowded Cities
Traffic Flow Monitoring in Crowded Cities John A. Quinn and Rose Nakibuule Faculty of Computing & I.T. Makerere University P.O. Box 7062, Kampala, Uganda {jquinn,rnakibuule}@cit.mak.ac.ug Abstract Traffic
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
Mean-Shift Tracking with Random Sampling
1 Mean-Shift Tracking with Random Sampling Alex Po Leung, Shaogang Gong Department of Computer Science Queen Mary, University of London, London, E1 4NS Abstract In this work, boosting the efficiency of
3D Vision An enabling Technology for Advanced Driver Assistance and Autonomous Offroad Driving
3D Vision An enabling Technology for Advanced Driver Assistance and Autonomous Offroad Driving AIT Austrian Institute of Technology Safety & Security Department Christian Zinner Safe and Autonomous Systems
A Reliability Point and Kalman Filter-based Vehicle Tracking Technique
A Reliability Point and Kalman Filter-based Vehicle Tracing Technique Soo Siang Teoh and Thomas Bräunl Abstract This paper introduces a technique for tracing the movement of vehicles in consecutive video
Wii Remote Calibration Using the Sensor Bar
Wii Remote Calibration Using the Sensor Bar Alparslan Yildiz Abdullah Akay Yusuf Sinan Akgul GIT Vision Lab - http://vision.gyte.edu.tr Gebze Institute of Technology Kocaeli, Turkey {yildiz, akay, akgul}@bilmuh.gyte.edu.tr
THE CONTROL OF A ROBOT END-EFFECTOR USING PHOTOGRAMMETRY
THE CONTROL OF A ROBOT END-EFFECTOR USING PHOTOGRAMMETRY Dr. T. Clarke & Dr. X. Wang Optical Metrology Centre, City University, Northampton Square, London, EC1V 0HB, UK [email protected], [email protected]
Real-time Traffic Congestion Detection Based on Video Analysis
Journal of Information & Computational Science 9: 10 (2012) 2907 2914 Available at http://www.joics.com Real-time Traffic Congestion Detection Based on Video Analysis Shan Hu a,, Jiansheng Wu a, Ling Xu
Vision-Based Blind Spot Detection Using Optical Flow
Vision-Based Blind Spot Detection Using Optical Flow M.A. Sotelo 1, J. Barriga 1, D. Fernández 1, I. Parra 1, J.E. Naranjo 2, M. Marrón 1, S. Alvarez 1, and M. Gavilán 1 1 Department of Electronics, University
VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS
VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS Aswin C Sankaranayanan, Qinfen Zheng, Rama Chellappa University of Maryland College Park, MD - 277 {aswch, qinfen, rama}@cfar.umd.edu Volkan Cevher, James
Building an Advanced Invariant Real-Time Human Tracking System
UDC 004.41 Building an Advanced Invariant Real-Time Human Tracking System Fayez Idris 1, Mazen Abu_Zaher 2, Rashad J. Rasras 3, and Ibrahiem M. M. El Emary 4 1 School of Informatics and Computing, German-Jordanian
Real-Time Tracking of Pedestrians and Vehicles
Real-Time Tracking of Pedestrians and Vehicles N.T. Siebel and S.J. Maybank. Computational Vision Group Department of Computer Science The University of Reading Reading RG6 6AY, England Abstract We present
Feature Tracking and Optical Flow
02/09/12 Feature Tracking and Optical Flow Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem Many slides adapted from Lana Lazebnik, Silvio Saverse, who in turn adapted slides from Steve
Removing Moving Objects from Point Cloud Scenes
1 Removing Moving Objects from Point Cloud Scenes Krystof Litomisky [email protected] Abstract. Three-dimensional simultaneous localization and mapping is a topic of significant interest in the research
Object tracking & Motion detection in video sequences
Introduction Object tracking & Motion detection in video sequences Recomended link: http://cmp.felk.cvut.cz/~hlavac/teachpresen/17compvision3d/41imagemotion.pdf 1 2 DYNAMIC SCENE ANALYSIS The input to
CS231M Project Report - Automated Real-Time Face Tracking and Blending
CS231M Project Report - Automated Real-Time Face Tracking and Blending Steven Lee, [email protected] June 6, 2015 1 Introduction Summary statement: The goal of this project is to create an Android
SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING
AAS 07-228 SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING INTRODUCTION James G. Miller * Two historical uncorrelated track (UCT) processing approaches have been employed using general perturbations
Tracking Groups of Pedestrians in Video Sequences
Tracking Groups of Pedestrians in Video Sequences Jorge S. Marques Pedro M. Jorge Arnaldo J. Abrantes J. M. Lemos IST / ISR ISEL / IST ISEL INESC-ID / IST Lisbon, Portugal Lisbon, Portugal Lisbon, Portugal
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
T-REDSPEED White paper
T-REDSPEED White paper Index Index...2 Introduction...3 Specifications...4 Innovation...6 Technology added values...7 Introduction T-REDSPEED is an international patent pending technology for traffic violation
3D Model based Object Class Detection in An Arbitrary View
3D Model based Object Class Detection in An Arbitrary View Pingkun Yan, Saad M. Khan, Mubarak Shah School of Electrical Engineering and Computer Science University of Central Florida http://www.eecs.ucf.edu/
Tracking and Recognition in Sports Videos
Tracking and Recognition in Sports Videos Mustafa Teke a, Masoud Sattari b a Graduate School of Informatics, Middle East Technical University, Ankara, Turkey [email protected] b Department of Computer
How To Fuse A Point Cloud With A Laser And Image Data From A Pointcloud
REAL TIME 3D FUSION OF IMAGERY AND MOBILE LIDAR Paul Mrstik, Vice President Technology Kresimir Kusevic, R&D Engineer Terrapoint Inc. 140-1 Antares Dr. Ottawa, Ontario K2E 8C4 Canada [email protected]
OBJECT TRACKING USING LOG-POLAR TRANSFORMATION
OBJECT TRACKING USING LOG-POLAR TRANSFORMATION A Thesis Submitted to the Gradual Faculty of the Louisiana State University and Agricultural and Mechanical College in partial fulfillment of the requirements
A Vision-Based Tracking System for a Street-Crossing Robot
Submitted to ICRA-04 A Vision-Based Tracking System for a Street-Crossing Robot Michael Baker Computer Science Department University of Massachusetts Lowell Lowell, MA [email protected] Holly A. Yanco
ASSESSMENT OF VISUALIZATION SOFTWARE FOR SUPPORT OF CONSTRUCTION SITE INSPECTION TASKS USING DATA COLLECTED FROM REALITY CAPTURE TECHNOLOGIES
ASSESSMENT OF VISUALIZATION SOFTWARE FOR SUPPORT OF CONSTRUCTION SITE INSPECTION TASKS USING DATA COLLECTED FROM REALITY CAPTURE TECHNOLOGIES ABSTRACT Chris Gordon 1, Burcu Akinci 2, Frank Boukamp 3, and
A SPECIAL APPLICATION OF A VIDEO-IMAGE SYSTEM FOR VEHICLE TRACKING AND SPEED MEASUREMENT
A SPECIAL APPLICATION OF A VIDEO-IMAGE SYSTEM FOR VEHICLE TRACKING AND SPEED MEASUREMENT Zong Tian (Corresponding Author) Texas Transportation Institute The Texas A&M University System College Station,
VEHICLE TRACKING AND SPEED ESTIMATION SYSTEM CHAN CHIA YIK. Report submitted in partial fulfillment of the requirements
VEHICLE TRACKING AND SPEED ESTIMATION SYSTEM CHAN CHIA YIK Report submitted in partial fulfillment of the requirements for the award of the degree of Bachelor of Computer System & Software Engineering
AN EXPERT SYSTEM TO ANALYZE HOMOGENEITY IN FUEL ELEMENT PLATES FOR RESEARCH REACTORS
AN EXPERT SYSTEM TO ANALYZE HOMOGENEITY IN FUEL ELEMENT PLATES FOR RESEARCH REACTORS Cativa Tolosa, S. and Marajofsky, A. Comisión Nacional de Energía Atómica Abstract In the manufacturing control of Fuel
ECE 533 Project Report Ashish Dhawan Aditi R. Ganesan
Handwritten Signature Verification ECE 533 Project Report by Ashish Dhawan Aditi R. Ganesan Contents 1. Abstract 3. 2. Introduction 4. 3. Approach 6. 4. Pre-processing 8. 5. Feature Extraction 9. 6. Verification
Social Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
Solving Simultaneous Equations and Matrices
Solving Simultaneous Equations and Matrices The following represents a systematic investigation for the steps used to solve two simultaneous linear equations in two unknowns. The motivation for considering
Segmentation & Clustering
EECS 442 Computer vision Segmentation & Clustering Segmentation in human vision K-mean clustering Mean-shift Graph-cut Reading: Chapters 14 [FP] Some slides of this lectures are courtesy of prof F. Li,
Optical Tracking Using Projective Invariant Marker Pattern Properties
Optical Tracking Using Projective Invariant Marker Pattern Properties Robert van Liere, Jurriaan D. Mulder Department of Information Systems Center for Mathematics and Computer Science Amsterdam, the Netherlands
Environmental Remote Sensing GEOG 2021
Environmental Remote Sensing GEOG 2021 Lecture 4 Image classification 2 Purpose categorising data data abstraction / simplification data interpretation mapping for land cover mapping use land cover class
ESE498. Intruder Detection System
0 Washington University in St. Louis School of Engineering and Applied Science Electrical and Systems Engineering Department ESE498 Intruder Detection System By Allen Chiang, Jonathan Chu, Siwei Su Supervisor
Tracking of Small Unmanned Aerial Vehicles
Tracking of Small Unmanned Aerial Vehicles Steven Krukowski Adrien Perkins Aeronautics and Astronautics Stanford University Stanford, CA 94305 Email: [email protected] Aeronautics and Astronautics Stanford
DYNAMIC RANGE IMPROVEMENT THROUGH MULTIPLE EXPOSURES. Mark A. Robertson, Sean Borman, and Robert L. Stevenson
c 1999 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or
How To Analyze Ball Blur On A Ball Image
Single Image 3D Reconstruction of Ball Motion and Spin From Motion Blur An Experiment in Motion from Blur Giacomo Boracchi, Vincenzo Caglioti, Alessandro Giusti Objective From a single image, reconstruct:
VECTORAL IMAGING THE NEW DIRECTION IN AUTOMATED OPTICAL INSPECTION
VECTORAL IMAGING THE NEW DIRECTION IN AUTOMATED OPTICAL INSPECTION Mark J. Norris Vision Inspection Technology, LLC Haverhill, MA [email protected] ABSTRACT Traditional methods of identifying and
Efficient Background Subtraction and Shadow Removal Technique for Multiple Human object Tracking
ISSN: 2321-7782 (Online) Volume 1, Issue 7, December 2013 International Journal of Advance Research in Computer Science and Management Studies Research Paper Available online at: www.ijarcsms.com Efficient
Tracking performance evaluation on PETS 2015 Challenge datasets
Tracking performance evaluation on PETS 2015 Challenge datasets Tahir Nawaz, Jonathan Boyle, Longzhen Li and James Ferryman Computational Vision Group, School of Systems Engineering University of Reading,
Data Cleansing for Remote Battery System Monitoring
Data Cleansing for Remote Battery System Monitoring Gregory W. Ratcliff Randall Wald Taghi M. Khoshgoftaar Director, Life Cycle Management Senior Research Associate Director, Data Mining and Emerson Network
V-PITS : VIDEO BASED PHONOMICROSURGERY INSTRUMENT TRACKING SYSTEM. Ketan Surender
V-PITS : VIDEO BASED PHONOMICROSURGERY INSTRUMENT TRACKING SYSTEM by Ketan Surender A thesis submitted in partial fulfillment of the requirements for the degree of Master of Science (Electrical Engineering)
How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm
IJSTE - International Journal of Science Technology & Engineering Volume 1 Issue 10 April 2015 ISSN (online): 2349-784X Image Estimation Algorithm for Out of Focus and Blur Images to Retrieve the Barcode
Vehicle Tracking Using On-Line Fusion of Color and Shape Features
Vehicle Tracking Using On-Line Fusion of Color and Shape Features Kai She 1, George Bebis 1, Haisong Gu 1, and Ronald Miller 2 1 Computer Vision Laboratory, University of Nevada, Reno, NV 2 Vehicle Design
Automatic Calibration of an In-vehicle Gaze Tracking System Using Driver s Typical Gaze Behavior
Automatic Calibration of an In-vehicle Gaze Tracking System Using Driver s Typical Gaze Behavior Kenji Yamashiro, Daisuke Deguchi, Tomokazu Takahashi,2, Ichiro Ide, Hiroshi Murase, Kazunori Higuchi 3,
Cloud tracking with optical flow for short-term solar forecasting
Cloud tracking with optical flow for short-term solar forecasting Philip Wood-Bradley, José Zapata, John Pye Solar Thermal Group, Australian National University, Canberra, Australia Corresponding author:
Mouse Control using a Web Camera based on Colour Detection
Mouse Control using a Web Camera based on Colour Detection Abhik Banerjee 1, Abhirup Ghosh 2, Koustuvmoni Bharadwaj 3, Hemanta Saikia 4 1, 2, 3, 4 Department of Electronics & Communication Engineering,
Detecting and Tracking Moving Objects for Video Surveillance
IEEE Proc. Computer Vision and Pattern Recognition Jun. 3-5, 1999. Fort Collins CO Detecting and Tracking Moving Objects for Video Surveillance Isaac Cohen Gérard Medioni University of Southern California
VOLUMNECT - Measuring Volumes with Kinect T M
VOLUMNECT - Measuring Volumes with Kinect T M Beatriz Quintino Ferreira a, Miguel Griné a, Duarte Gameiro a, João Paulo Costeira a,b and Beatriz Sousa Santos c,d a DEEC, Instituto Superior Técnico, Lisboa,
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY V. Knyaz a, *, Yu. Visilter, S. Zheltov a State Research Institute for Aviation System (GosNIIAS), 7, Victorenko str., Moscow, Russia
Real Time Target Tracking with Pan Tilt Zoom Camera
2009 Digital Image Computing: Techniques and Applications Real Time Target Tracking with Pan Tilt Zoom Camera Pankaj Kumar, Anthony Dick School of Computer Science The University of Adelaide Adelaide,
ROBOTRACKER A SYSTEM FOR TRACKING MULTIPLE ROBOTS IN REAL TIME. by Alex Sirota, [email protected]
ROBOTRACKER A SYSTEM FOR TRACKING MULTIPLE ROBOTS IN REAL TIME by Alex Sirota, [email protected] Project in intelligent systems Computer Science Department Technion Israel Institute of Technology Under the
Human behavior analysis from videos using optical flow
L a b o r a t o i r e I n f o r m a t i q u e F o n d a m e n t a l e d e L i l l e Human behavior analysis from videos using optical flow Yassine Benabbas Directeur de thèse : Chabane Djeraba Multitel
Face Model Fitting on Low Resolution Images
Face Model Fitting on Low Resolution Images Xiaoming Liu Peter H. Tu Frederick W. Wheeler Visualization and Computer Vision Lab General Electric Global Research Center Niskayuna, NY, 1239, USA {liux,tu,wheeler}@research.ge.com
Face detection is a process of localizing and extracting the face region from the
Chapter 4 FACE NORMALIZATION 4.1 INTRODUCTION Face detection is a process of localizing and extracting the face region from the background. The detected face varies in rotation, brightness, size, etc.
A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow
, pp.233-237 http://dx.doi.org/10.14257/astl.2014.51.53 A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow Giwoo Kim 1, Hye-Youn Lim 1 and Dae-Seong Kang 1, 1 Department of electronices
