Head and Facial Animation Tracking using Appearance-Adaptive Models and Particle Filters

Size: px
Start display at page:

Download "Head and Facial Animation Tracking using Appearance-Adaptive Models and Particle Filters"

Transcription

1 Head and Facial Animation Tracking using Appearance-Adaptive Models and Particle Filters F. Dornaika and F. Davoine CNRS HEUDIASYC Compiègne University of Technology 625 Compiègne Cedex, FRANCE {dornaika, Abstract This paper introduces two frameworks for head and facial animation tracking. The first framework introduces a particle-filter tracker capable of tracking the 3D head pose using a statistical facial texture model. The second framework introduces an appearance-adaptive tracker capable of tracking the 3D head pose and the facial animations in real-time. This framework has the merits of both deterministic and stochastic approaches. It consists of an online adaptive observation model of the face texture together with an adaptive transition motion model. The latter is based on a registration technique between the appearance model and the incoming observation. The second framework extends the concept of Online Appearance Models to the case of tracking 3D non-rigid face motion (3D head pose and facial animations). Tracking long video sequences demonstrated the effectiveness of the developed methods. Accurate tracking was obtained even in the presence of perturbing factors such as illumination changes, significant head pose and facial expression variations as well as occlusions. 1. Introduction Object tracking is required by many vision applications, especially in video technology and visual interface systems. The ability to track facial motion is useful in applications such as face-based biometric person authentication, expression analysis, and human computer interaction (HCI). Visual tracking of faces can be utilized for recognizing basic activities from videos. Detecting and tracking faces in video sequences is a challenging task because faces are non-rigid and their images have a high degree of variability. A huge research effort has already been devoted to detecting and tracking of facial features in 2D and 3D (e.g., [3, 4, 6, 9]). Visual tracking can be performed by feature-based or appearance-based approaches. Moreover, tracking approaches can be divided into two groups: deterministic and stochastic. Probabilistic video analysis has recently gained significant attention in the computer vision community through the use of stochastic sampling techniques. Visual tracking based on probabilistic analysis is formulated from a Bayesian perspective as a problem of estimating some degree of belief in the state of an object at the current time given a sequence of observations. Particle filtering methods [2] (also known as Sequential Monte Carlo (SMC) methods) were independently used and proposed by several research groups. These algorithms provide flexible tracking frameworks as they are neither limited to linear systems nor require the noise to be Gaussian and proved to be more robust to distracting clutter as the randomly sampled particles allow to maintain several competing hypotheses of the hidden state. These algorithms have gained prevalence in the video tracking literature due in part to the CONDENSATION algorithm [8]. Classical 3D vision techniques provide tools for computing the 3D pose and the possible facial animations from images. However, such trackers very often suffer from the drifting problem (error accumulation) since facial features do not have enough stable local appearance due to many factors. Within deterministic approaches, appearance-based techniques have been widely used in object tracking. These techniques have the advantage that they are easy to implement and are generally more robust than feature-based methods. To overcome the problem of appearance changes, recent works on faces adopted statistical facial textures. For example, the Active Appearance Models have been proposed as a powerful tool for analyzing facial images [5, 11]. In [1], the author proposed a framework that tracks face and facial features using Active Appearance Models. While statistical appearance-based tracking methods are promis-

2 ing with respect to some aspects, they are depending on the imaging conditions under which the learning is performed. Thus, by changing these conditions, one should repeat the whole learning process, which can be very tedious. Recently, 2D tracking approaches have adopted online appearance models (OAMs) [1, 16]. In [16], a deterministic and stochastic approach was developed to track the 2D motion of faces using an affine transform. The developed approach adopts adaptive observation models since the object appearance is learned during the tracking. Thus, unlike tracking approaches using statistical texture modelling OAMs are expected to offer a lot of flexibility. This paper has two main contributions. First, we develop a particle filter-based approach for tracking the 3D head pose using a statistical facial texture model. Second, we propose a framework for tracking the 3D head pose and the facial animations in real-time using an online appearance model where both the observation and transition models are adaptive. The second framework extends the concept of OAMs to the case of tracking 3D non-rigid face motion (3D head pose and facial animation). The rest of the paper is organized as follows. Section 2 describes the deformable 3D face model. Section 3 states the tracking problem. Section 4 describes the particle filter-based 3D head tracking. Section 5 describes the head and facial animation tracking using appearance-adaptive models and particle filters. Section 6 presents some experimental results. 2. Modelling faces 2.1. A deformable 3D model Building a generic 3D face model is a challenging task. Indeed, such a model should account for the differences between different specific human faces as well as between different facial expressions. This modelling was explored in the computer graphics, computer vision, and model-based image coding communities. In our study, we use the 3D face model Candide. This 3D deformable wireframe model was first developed for the purpose of model-based image coding and computer animation. The 3D shape of this deformable 3D wireframe model is directly recorded in coordinate form. The 3D face model is given by the 3D coordinates of the vertices P i, i = 1,..., n where n is the number of vertices. Thus, the shape up to a global scale can be fully described by the 3n-vector g the concatenation of the 3D coordinates of all vertices P i. The vector g can be written as: g = g + S τ s + A τ a (1) where g is the standard shape of the model, and the columns of S and A are the Shape and Animation Units, respectively. A Shape Unit provides a way to deform the 3D wireframe such as to adapt the eye width, the head width, the eye separation distance etc. Thus, the term S τ s accounts for shape variability (inter-person variability) while the term A τ a accounts for the facial animation (intra-person variability). The shape and animation variabilities can be approximated well enough for practical purposes by this linear relation. Also, we assume that the two kinds of variability are independent. In this study, we use 12 modes for the Shape Units matrix and six modes for the Animation Units matrix. Without loss of generality, we have chosen the following Action Units: 1) Jaw drop, 2) Lip stretcher, 3) Lip corner depressor, 4) Upper lip raiser, 5) Eyebrow lowerer, 6) Outer eyebrow raiser. These Action Units are enough to cover most common facial actions (mouth and eyebrow movements). In Equation (1), the 3D shape is expressed in a local coordinate system. However, one should relate the 3D coordinates to the image coordinate system. To this end, we adopt the weak perspective projection model. We neglect the perspective effects since the depth variation of the face can be considered as small compared to its absolute depth. Therefore, the mapping between the 3D face model and the image is given by a 2 4 matrix, M, encapsulating both the 3D head pose and the camera parameters. Thus, a 3D vertex P i = (X i, Y i, Z i ) T g will be projected onto the image point p i = (u i, v i ) T given by: (u i, v i ) T = M (X i, Y i, Z i, 1) T (2) For a given person, τ s is constant. Estimating τ s can be carried out using either feature-based or featureless approaches. Thus, the state of the 3D model is given by the 3D head pose (three rotations and three translations) and the control vector τ a. This is given by the state vector b: b = [θ x, θ y, θ z, t x, t y, t z, τ a T ] T (3) 2.2. Shape-free facial images A face texture is represented as a shape-free texture (geometrically normalized image). The geometry of this image is obtained by projecting the standard shape g (wireframe) using a standard 3D pose (frontal view) onto an image with a given resolution. The texture of this geometrically normalized image is obtained by texture mapping from the triangular 2D mesh in the input image using a piece-wise affine transform, W. Mathematically, the warping process applied to an input image

3 y is denoted by: x(b) = W(y, b) (4) where x denotes the shape-free texture and b denotes the geometrical parameters. Here images are represented by one-dimensional vectors. Two resolution levels have been used for the shape-free textures, encoded by 131 and 5392 pixels. Figure 1 illustrates the warping process applied to two input images. In this example, the resolution of the shape-free images is 5392 pixels. Regarding photometric transformations, a zeromean-unit-variance normalization is used to partially compensate for contrast variations. The complete image transformation is implemented as follows: (i) transfer the texture y using the piece-wise affine transform associated with the geometric parameters b = [θ x, θ y, θ z, t x, t y, t z, τ a T ] T, and (ii) perform zero-meanunit-variance normalization on the obtained patch. 4. 3D head tracking with a particle filter In this Section, we are interested in tracking the 3D head pose. Therefore, the state vector b is given by b = [θ x, θ y, θ z, t x, t y, t z ] T. In this particular case, the animation parameters τ a can be set to zero. Almost all previous works using particle filters have focused on tracking 2D objects or motions such as 2D contours, 2D blobs, 2D affine transforms [18], and 2D ellipses [13, 15]. Here (in this Section), we propose a particle filtering method that tracks the six degrees of freedom associated with the head motion where the face model is given the Candide model. Particle filtering (or Sequential Monte Carlo method) is an inference process which can be considered as a generalization of the Kalman filter. It aims at estimating the unknown state b t from a set of noisy observations (images), y 1:t = {y 1,, y t } arriving in a sequential fashion. Two important components of this approach are the state transition and observation models whose most general forms can be given by: State transition model b t = F t (b t 1, U t ) (5) Observation model y t = G t (b t, V t ) (6) (a) (b) Figure 1: (a) Two input images with correct adaptation. (b) The corresponding shape-free facial images. 3. The tracking problem Given a video sequence depicting a moving face, the tracking consists in estimating (for each frame) the 3D head pose as well as the facial animations encoded by the control vector τ a. In other words, one would like to estimate the vector b (Eq.(3)) for each frame t. In a tracking context (deterministic and stochastic trackers), the model parameters associated with the current frame will be handed over to the next frame. where U t is the system noise, F t is the kinematics, V t is the observation noise, and G t models the observer. The particle filter approximates the posterior distribution p(b t y 1:t ) by a set of weighted particles {b (j) t, w (j) t } J j=1. Each element b(j) t represents the hypothetical state of the object and w (j) t is the corresponding discrete probability. Then, the state estimate can be set to the minimum mean square error (MMSE) estimate or the maximum a posteriori (MAP). We use the following simple state transition model: b t = b t 1 + U t (7) In this model, U t is a random vector having a centred normal distribution, N(, Σ). The covariance matrix Σ is learned off-line from the state vector differences b t b t 1 associated with previously tracked video sequences. Since image data y are represented as shape-free textures x (the warped texture), we can set the observation likelihood p(y t b t ) to p(x t b t ). The observation likelihood p(x t b t ) quantifies the consistence of the texture x(b t ) with the statistical texture model represented by texture modes (eigenvectors). For this purpose, we use a likelihood measure such the

4 one proposed in [12]: ( p(x t b t ) = c exp 1 2 M ξ 2 i λ 2 i=1 i ) ( exp e ) 2ρ (8) where e is the reconstruction error, λ i s are the eigenvalues associated with the first M eigenvectors, and ρ is the arithmetic average of the remaining eigenvalues. This likelihood measure takes into account two distances (i) the DFFS (distance-from-feature-space), and (ii) the DIFS (distance-in-feature-space). Maximizing this likelihood is equivalent to minimizing the Mahalanobis distance over the original textures. A particle filter algorithm proceeds as follows: Initialize a sample set S = {b (j), 1}J j=1 according to some prior distribution p(b ). For t = 1, 2, For j = 1, 2,, J Resample S t 1 (b (j) t 1, 1). Predict the sample b (j) t to obtain a new sample from b (j) t 1 by drawing U (j) t and computing according to Eq. (7). Compute the geometrically normalized texture x(b (j) t ) according to Eq. (4). Update the weight using w (j) t according to Eq. (8). End Normalize the weights w (j) t = p(x t b (j) t ) = w (j) t / J i=1 w(i) t ; j = 1, 2,..., J End During filtering, due to the resampling step, samples with a high weight may be chosen several times while others with relatively low weights may not be chosen at all. Note that the initial distribution p(b ) can be either a Dirac or Gaussian distribution centred on a solution provided by a detector algorithm or manually specified. 5. Appearance-adaptive models and particle filters In this Section, we consider the 3D head pose as well as the facial animations, that is, the state vector b is given by b = [θ x, θ y, θ z, t x, t y, t z, τ a T ] T Motivations The efficiency of the stochastic tracking algorithm presented above depends on many factors. However, the main factor which limits the efficiency of stochastic tracking algorithms is the lack of suitable state transition models. Indeed, there are two ways for handling the transition model. (i) The first is to learn state transition models directly from training videos. For example, in [14] the authors use Expectation-Maximization based on the CONDENSATION algorithm to learn multiclass dynamics associated with a juggled ball. However, such models may not necessarily succeed when presented with testing videos featuring different types of motions. (ii) The second is to use a fixed model with fixed noise variance for simplicity, that is, the predicted state is simply the previous state (or a shifted version of it) to which a random noise with fixed variance is added (this methodology was adopted in Section 4). If the variance is very small, it is very hard to model rapid movements; if the variance is large, it is computationally inefficient since many more particles are needed to accommodate large noise variance. In addition to the problems associated with the state transition model, the observation model has its own limitations. For example, if the observation model (observation likelihood) is built upon a statistical texture model, any significant change in the imaging conditions will make the corresponding learned observation model useless and one should build a new observation model based on a new statistical texture model. For all these factors, we develop a new tracking framework capable of coping with the limitations mentioned above. Our approach is to make both observation and state transition models adaptive in the framework of a particle filter, with provisions for handling outliers embedded. The main features of the developed approach are: Adaptive observation model. We adopt an appearance-based approach. Our approach adopts the concept of online appearance model (OAM) where the appearance is learned on-line from the tracked video sequence [16, 17, 1]. However, in our case, we extend this concept to the case of tracking the 3D non-rigid face motion (3D head pose and facial animation). Therefore, the observation model is adaptive as the appearance of the texture. Adaptive state transition model. Instead of using a fixed state transition model, we use an adaptivevelocity model, where the adaptive motion velocity is predicted using a registration technique between the incoming observation and the current appearance configuration. We also use an adaptive noise component whose magnitude is a function of the registration error. We vary the number of particles based on the noise component.

5 Handling occlusion. Occlusion and large image variations are handled using robust statistics. We robustify the likelihood measurement and the adaptive velocity estimate by downweighting the outlier pixels Adaptive observation model Now we consider the shape-free texture x t. Remember that x t is a transformed version of the observation y t (see Eq. (4)). The appearance model at time t, A t, is a time-varying one that models the appearances present in all observations x up to time t 1. For each frame, the observation is simply the warped texture associated with the computed geometric parameters b t. We use the hat symbol for the tracked parameters and textures. For a given frame t, ˆb t represents the computed geometric parameters and ˆx t the corresponding texture patch, that is, ˆx t = x(ˆb t ) = W(y t, ˆb t ) The appearance model A t obeys a Gaussian with a center µ and a variance σ. Notice that µ and σ are vectors consisting of d pixels (d is the size of x) that are assumed to be independent of each other. In summary, the observation likelihood is written as p(y t b t ) = p(x t b t ) = d N(x i ; µ i, σ i ) (9) i=1 where N(x; µ i, σ i ) is a normal density: [ ( )] x N(x; µ i, σ i ) = (2πσi 2 ) 1/2 µi exp ρ σ i (1) where ρ(x) is given by ρ(x) = 1 2 x2. We assume that A t summarizes the past observations under an exponential envelop with a forgetting factor α. When the appearance is tracked for the current input image, i.e. the texture ˆx t is available, we can compute the updated appearance and use it to track in the next frame. It can be shown that the appearance model parameters, i.e, µ and σ can be updated using the following equations (see [1] for more details on OAMs): µ t+1 = α µ t + (1 α) ˆx t (11) σ 2 t+1 = α σ 2 t + (1 α) (ˆx t µ t ) 2 (12) In the above equations, all µ s and σ 2 s are vectorized and the operation is element-wise. This technique, also called recursive filtering, is simple, time-efficient and therefore, suitable for real-time applications. Note that µ is initialized with the first patch x. However, Eq.(12) is not used until the number of frames reaches a certain value (e.g., the first 4 frames). For these frames, the classical variance is used, that is, Eq.(12) is utilized with α being set to 1 1 t Adaptive transition model Instead of using a fixed function F to predict the transition state from time (t 1) to time t, we use the following adaptive transition model: b t = ˆb t 1 + b t + U t (13) where b t is the shift in the geometric parameters and U t is the random noise. Our basic idea allowing to recover the solution b t or equivalently the deterministic part of Eq.(13) is to use region-based registration techniques. In other words, the current input image y t is registered with the current appearance model A t. For this purpose, we minimize the Mahalanobis distance between the warped texture and the current appearance mean, d ( ) 2 xi µ i min e(b t ) = min D(x(b t ), µ t ) = b t b t σ i=1 i (14) Note the appearance parameters µ t and σ t are known. The above criterion can be minimized using iterative first-order linear approximation. Gradient-descent registration We assume that there exists a b t = ˆb t 1 + b t such that the warped texture will be very close to the appearance mean, i.e, W(y t, b t ) µ t Approximating W(y t, b t ) via a first-order Taylor series expansion around ˆb t 1 yields W(y t, b t ) W(y t, ˆb t 1 ) + G t (b t ˆb t 1 ) where G t is the gradient matrix. above two equations we have: µ t = W(y t, ˆb t 1 ) + G t (b t ˆb t 1 ) By combining the Therefore, the shift in the parameter space is given by: b t = b t ˆb t 1 = G t (W(y t, ˆb t 1 ) µ t ) (15) In practice, the solution b t (or equivalently the shift b t ) is estimated by running several iterations until the error cannot be improved. We proceed as follows. Starting from b = ˆb t 1, we compute the error vector (W(y t, ˆb t 1 ) µ t ) and the corresponding Mahalanobis

6 distance e(b) (given by Eq.(14)). We find a shift b by multiplying the error vector with the negative pseudoinverse of the gradient matrix using Eq.(15). The vector b gives a displacement in the search space for which the error, e, can be minimized. We compute a new parameter vector and a new error: b = b + ρ b (16) e = e(b ) where ρ is a positive real. If e < e, we update b according to Eq.( 16) and the process is iterated until convergence. If e e, we try smaller update steps having the same direction (i.e., smaller ρ is used). Convergence is declared when the error cannot be improved anymore. Computing the gradient matrix is given by: G = W(y t, b t) b The gradient matrix = x t b It is approximated by numerical differences. Once the solution ˆb t becomes available for a given frame, it is possible to compute the gradient matrix from the associated input image. We use the following: The j th column of G (j = 1,..., dim(b)): G j = W(y t, b t) b j can be estimated using differences G j W(y t, b t) W(y t, b t + δ q j ) δ where δ is a suitable step size and q j is a vector with all elements zero except the j th element that equals one. To gain more accuracy, the j th column of G is estimated using several steps around the current value b j, and then averaging over all these, we get our final G j as G j = 1 K K/2 k= K/2,k W(y t, b t ) W(y t, b t + k δ j q j ) k δ j where δ j is the smallest perturbation associated with the parameter b j and K is the number of steps (in our experiments, K is set to 8). Notice that the computation of the gradient matrix G t at time t is carried out using the estimated geometric parameters ˆb t 1 and the associated input image y t 1 since the adaptation for the time t has not been computed. It is worthwhile noting that the gradient matrix is computed for each time step. The advantage is twofold. First, a varying gradient matrix is able to accommodate appearance changes. Second, it will be closer to the exact gradient matrix since it is computed for the current geometric configuration (3D head pose and facial animations) whereas a fixed gradient matrix can be a source of errors for some kinds of motions such as out-of-plane motions Handling outliers and occlusions We assume that occlusion and large image differences can be treated as outliers. Outlier pixels cannot be explained by the underlying process (the current appearance model A t ) and their influences on the estimation process should be reduced. Robust statistics provide such mechanisms [7]. The mechanism will have impact on three items: (i) the likelihood measure, (ii) the gradient descent method, and (iii) the update of the online appearance model A t. We use the ˆρ function defined as follows: { 1 ˆρ(x) = 2 x2 if x c c x 1 2 c2 if x > c where x is the value of a pixel in the patch x, normalized by the mean and the variance of the appearance at the same pixel, i.e. µ i and σ i. The constant c controls the outlier rate. In our experiment, we take c = 3 based on experimental experience. If x > c is satisfied, we declare the corresponding pixel an outlier. Likelihood measure To make the likelihood measure robust, we replace the one-dimensional normal density N(x; µ i, σ i ) by [ ( )] x ˆN(x; µ i, σ i ) = (2πσi 2 ) 1/2 µi exp ˆρ Gradient method To downweight the influence of the outlier pixels in the registration technique, we introduce a d d diagonal matrix L t with its i th diagonal element being L t (i) = η(x i ) where x i is the i th element of the difference image (W(y t, ˆb t 1 ) µ t ) normalized by the corresponding variance σ i and η(x) = 1 { dˆρ(x) 1 if x c x dx = c x if x > c Therefore, the shift used in the gradient-descent registration becomes b t = b t ˆb t 1 = G t L t (W(y t, ˆb t 1 ) µ t ) (17) Note that the quadratic error function (Eq.(15)) can also be replaced by the robust function ˆρ. σ i

7 Appearance update Once the solution b t is ready, the corresponding patch x will be used to update the appearance. For non-outlier pixels the update equations are given by Eqs. (11) and (12); for outlier pixels the corresponding means and variances are not updated. This mechanism is very useful for preventing occlusions from deteriorating the online appearance model. Occlusions Occlusion is declared if the percentage of outliers exceeds a certain threshold A tracking algorithm Tracking the 3D head pose and the facial animations is performed as follows. Starting from the solution, ˆb t 1, associated with the previous frame, we predict the state using Eq.(13) in which the deterministic part of the prediction, i.e. ˆb t 1 + b t, is computed by the registration technique and the noise variance was set as a monotically increasing function of the registration error obtained at convergence. Once a set of particles is obtained, the MAP of Eq.(9) is again chosen to be the solution of the current frame. As can be seen, unlike the classical particle filtering, the propagation concerns the MAP solution only and not the whole particle set. It is worthwhile noting that although the solutions provided by the deterministic and stochastic parts of Eq.(13) have utilized the same observation model, there are some differences. The deterministic solution is obtained by a directed continuous search starting from the solution associated with the previous frame. The stochastic solution is obtained by diffusing the deterministic solution in order to obtain possible refinement. 6. Experimental results D head tracking with a particle filter Figure 2 displays the tracking results associated with several frames of a long test sequence. It corresponds to the particle-filter-based tracking algorithm using a statistical texture model. The number of particles is set to 3. For each frame in this figure, only the MAP solution is displayed. The statistical facial texture is built with 33 training face images and the number of the principal eigenvectors is 2. The bottom plot displays the weights associated with the bottom-right image Head and facial animation tracking with OAMs Figure 3 displays the head and facial animation tracking results associated with a 8-frame-long (only four Particles Figure 2: Particle-filter-based 3D head tracking with a statistical facial texture model. The bottom plot displays the weights associated with the bottom-right image. frames are shown). These results correspond to the realtime tracker based on appearance-adaptive model (described in Section 5). The sequence features quite large head pose variations as well as large facial animations. The sequence is of resolution 64x48 pixels. As can be seen with the very little prior information, the 3D motion of the face as well as the facial actions associated with the mouth and the eyebrows are accurately recovered. The upper left corner shows the current appearance (µ t ) and the current shape-free texture (ˆx t ). Figure 4 displays the estimated values of the 3D head pose parameters (the three rotations and the three translations) as well as the lower lip depressor, and the inner brow lowerer as a function of the frames of the same sequence. Figure 5 displays the face and facial animation results associated with a 62-frame-long sequence. The left column displays the tracking results of the realtime tracker using the appearance-adaptive model with a

8 fixed gradient matrix computed at the first video frame. The right column displays the tracking results when a time-varying gradient matrix has been used. We have noticed, for this sequence, that whenever the face performs an out-of-plane motion the tracker with a timevarying gradient is more accurate than the one using a fixed gradient. Moreover, in other experiments, the tracker using a fixed gradient matrix has totally lost the track. Figure 6 displays the face and facial animation results associated with another 42-frame-long sequence featuring two occlusions (only the second occlusion is displayed). The two occlusions are caused by putting the hand in front of the face. The frames 218 and 265 shows the start and the end of the second occlusion, respectively. On a 2.2GHz PC, a non-optimized C code of the developed appearance-adaptive algorithm runs around 14 frames-per-second assuming that the patch resolution is 131 pixels, K is 8, and the average number of particles is 6. By dropping the stochastic diffusion stage, the algorithm runs 2 frames-per-second. 7 Conclusion In this paper, we have proposed two tracking methods. The first method is fully stochastic and uses a particle filter with an observation likelihood based on statistical facial textures. This method has been utilized for 3D head tracking. The second method combines the merits of both stochastic and deterministic methods and is capable of tracking the head and facial animation. It employs an Online Appearance Model where both the observation and transition models are adaptive. The deterministic part exploits a directed continuous search aiming at minimizing the discrepancy between the upcoming observation and the current appearance model. Tracking long video sequences demonstrated the effectiveness of the developed methods. Accurate tracking was obtained even in the presence of perturbing factors such as illumination changes, significant head pose and facial expression variations as well as occlusions. Currently, we are investigating the recognition of facial expressions and gestures from the tracked parameters. Figure 3: Our real-time framework for tracking the 3D head pose and the facial animations with an appearanceadaptive model. The sequence length is 8 frames.

9 Pitch Yaw Deg. Deg Roll Scale Deg X Translation 1 Y Translation Pixels Pixels Lower lip depressor Inner brow lowerer Figure 5: 94, 21, 376, and 569 (from top to bottom). The left column displays the head and facial animation tracking using the appearance-adaptive framework with a fixed gradient. The right column displays the tracking results when a time-varying gradient has been used (see text) Figure 4: The tracked parameters as a function of time associated with the 8-frame-long sequence. The first six plots display the six degrees of freedom of the 3D head pose. The two bottom plots display the lower lip depressor and inner brow parameters, respectively. Frame 218 Frame 265 Frame 392 Figure 6: Tracking another test sequence featuring two occlusions. 218, 265, and 392 are displayed. Only the second occlusion is shown. The frames 218 and 265 shows the start and the end of the second occlusion, respectively. As can be seen, the head is accurately tracked even when a sustained occlusion occurs.

10 References [1] J. Ahlberg. An active model for facial feature tracking. EURASIP Journal on Applied Signal Processing, 22(6): , June 22. [2] S. Arulampalam, S.R. Maskell, N.J. Gordon, and T. Clapp. A tutorial on particle filters for on-line nonlinear/non-gaussian Bayesian tracking. IEEE Trans. on Signal Processing, 5(2): , 22. [3] S. Birchfield. Elliptical head tracking using intensity gradients and color histograms. In Proc. IEEE Conference on Computer Vision and Pattern Recognition, [4] M.L. Cascia, S. Sclaroff, and V. Athitsos. Fast, reliable head tracking under varying illumination: An approach based on registration of texturemapped 3D models. IEEE Transactions on Pattern Analysis and Machine Intelligence, 22(4): , 2. [5] T.F. Cootes, G.J. Edwards, and C.J. Taylor. Active appearance models. IEEE Transactions on Pattern Analysis and Machine Intelligence, 23(6): , 21. [6] S.B. Gokturk, J.Y. Bouguet, and R. Grzeszczuk. A data-driven model for monocular face tracking. In Proc. IEEE International Conference on Computer Vision, 21. [7] P.J. Huber. Robust Statistics. Wiley, 23. Transactions on Pattern Analysis and Machine Intelligence, 19(7):696 71, [13] H. Nait-Cherif and S.J. McKenna. Head tracking and action recognition in a smart meeting room. In IEEE International Workshop on Performance Evaluation of Tracking and Surveillance, March 23. [14] B. North, A. Blake, M. Isard, and J. Rittscher. Learning and classification of complex dynamics. IEEE Transactions on Pattern Analysis and Machine Intelligence, 22(9): , 2. [15] K. Nummiaro, E. Koller-Meier, and L. Van Gool. Object tracking with an adaptive color-based particle filter. In Symposium for Pattern Recognition of the DAGM, September 22. [16] S. Zhou, R. Chellappa, and B. Mogghaddam. Adaptive visual tracking and recognition using particle filters. In IEEE International Conference on Multimedia and Expo, 23. [17] S. Zhou, R. Chellappa, and B. Mogghaddam. Adaptive visual tracking and recognition using appearance-adaptive models in particle filters. IEEE Transactions on Image Processing, 24. [18] S. Zhou, V. Krueger, and R. Chellappa. Probabilistic recognition of human faces from video. Computer Vision and Image Understinding, 91(1-2): , 23. [8] M. Isard and A. Black. Contour tracking by stochastic propogation of conditioanl density. In Proc. European Conference on Computer Vision, [9] T.S. Jebara and A. Pentland. Parameterized structure from motion for 3D adaptative feedback tracking of faces. In Proc. IEEE Conference on Computer Vision and Pattern Recognition, [1] A.D. Jepson, D.J. Fleet, and T.F. El-Maraghi. Robust online appearance models for visual tracking. IEEE Transactions on Pattern Analysis and Machine Intelligence, 25(1): , 23. [11] I. Matthews and S. Baker. Active appearance models revisited. Technical Report CMU-RI-TR-3-2, The Robotics Institute, Carnegie Mellon University, 22. [12] B. Moghaddam and A. Pentland. Probabilistic visual learning for object representation. IEEE

Mean-Shift Tracking with Random Sampling

Mean-Shift Tracking with Random Sampling 1 Mean-Shift Tracking with Random Sampling Alex Po Leung, Shaogang Gong Department of Computer Science Queen Mary, University of London, London, E1 4NS Abstract In this work, boosting the efficiency of

More information

Practical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006

Practical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006 Practical Tour of Visual tracking David Fleet and Allan Jepson January, 2006 Designing a Visual Tracker: What is the state? pose and motion (position, velocity, acceleration, ) shape (size, deformation,

More information

Subspace Analysis and Optimization for AAM Based Face Alignment

Subspace Analysis and Optimization for AAM Based Face Alignment Subspace Analysis and Optimization for AAM Based Face Alignment Ming Zhao Chun Chen College of Computer Science Zhejiang University Hangzhou, 310027, P.R.China zhaoming1999@zju.edu.cn Stan Z. Li Microsoft

More information

Deterministic Sampling-based Switching Kalman Filtering for Vehicle Tracking

Deterministic Sampling-based Switching Kalman Filtering for Vehicle Tracking Proceedings of the IEEE ITSC 2006 2006 IEEE Intelligent Transportation Systems Conference Toronto, Canada, September 17-20, 2006 WA4.1 Deterministic Sampling-based Switching Kalman Filtering for Vehicle

More information

Vision based Vehicle Tracking using a high angle camera

Vision based Vehicle Tracking using a high angle camera Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu gramos@clemson.edu dshu@clemson.edu Abstract A vehicle tracking and grouping algorithm is presented in this work

More information

PASSIVE DRIVER GAZE TRACKING WITH ACTIVE APPEARANCE MODELS

PASSIVE DRIVER GAZE TRACKING WITH ACTIVE APPEARANCE MODELS PASSIVE DRIVER GAZE TRACKING WITH ACTIVE APPEARANCE MODELS Takahiro Ishikawa Research Laboratories, DENSO CORPORATION Nisshin, Aichi, Japan Tel: +81 (561) 75-1616, Fax: +81 (561) 75-1193 Email: tishika@rlab.denso.co.jp

More information

Template-based Eye and Mouth Detection for 3D Video Conferencing

Template-based Eye and Mouth Detection for 3D Video Conferencing Template-based Eye and Mouth Detection for 3D Video Conferencing Jürgen Rurainsky and Peter Eisert Fraunhofer Institute for Telecommunications - Heinrich-Hertz-Institute, Image Processing Department, Einsteinufer

More information

A Robust Multiple Object Tracking for Sport Applications 1) Thomas Mauthner, Horst Bischof

A Robust Multiple Object Tracking for Sport Applications 1) Thomas Mauthner, Horst Bischof A Robust Multiple Object Tracking for Sport Applications 1) Thomas Mauthner, Horst Bischof Institute for Computer Graphics and Vision Graz University of Technology, Austria {mauthner,bischof}@icg.tu-graz.ac.at

More information

Face Model Fitting on Low Resolution Images

Face Model Fitting on Low Resolution Images Face Model Fitting on Low Resolution Images Xiaoming Liu Peter H. Tu Frederick W. Wheeler Visualization and Computer Vision Lab General Electric Global Research Center Niskayuna, NY, 1239, USA {liux,tu,wheeler}@research.ge.com

More information

Robust Visual Tracking and Vehicle Classification via Sparse Representation

Robust Visual Tracking and Vehicle Classification via Sparse Representation Robust Visual Tracking and Vehicle Classification via Sparse Representation Xue Mei and Haibin Ling, Member, IEEE Abstract In this paper, we propose a robust visual tracking method by casting tracking

More information

Face detection is a process of localizing and extracting the face region from the

Face detection is a process of localizing and extracting the face region from the Chapter 4 FACE NORMALIZATION 4.1 INTRODUCTION Face detection is a process of localizing and extracting the face region from the background. The detected face varies in rotation, brightness, size, etc.

More information

Robust Infrared Vehicle Tracking across Target Pose Change using L 1 Regularization

Robust Infrared Vehicle Tracking across Target Pose Change using L 1 Regularization Robust Infrared Vehicle Tracking across Target Pose Change using L 1 Regularization Haibin Ling 1, Li Bai, Erik Blasch 3, and Xue Mei 4 1 Computer and Information Science Department, Temple University,

More information

Efficient Generic Face Model Fitting to Images and Videos

Efficient Generic Face Model Fitting to Images and Videos Efficient Generic Face Model Fitting to Images and Videos Luis Unzueta a,, Waldir Pimenta b, Jon Goenetxea a, Luís Paulo Santos b, Fadi Dornaika c,d a Vicomtech-IK4, Paseo Mikeletegi, 57, Parque Tecnológico,

More information

Bayesian Image Super-Resolution

Bayesian Image Super-Resolution Bayesian Image Super-Resolution Michael E. Tipping and Christopher M. Bishop Microsoft Research, Cambridge, U.K..................................................................... Published as: Bayesian

More information

Real-time Visual Tracker by Stream Processing

Real-time Visual Tracker by Stream Processing Real-time Visual Tracker by Stream Processing Simultaneous and Fast 3D Tracking of Multiple Faces in Video Sequences by Using a Particle Filter Oscar Mateo Lozano & Kuzahiro Otsuka presented by Piotr Rudol

More information

Least-Squares Intersection of Lines

Least-Squares Intersection of Lines Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a

More information

Linear Threshold Units

Linear Threshold Units Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear

More information

AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION

AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION Saurabh Asija 1, Rakesh Singh 2 1 Research Scholar (Computer Engineering Department), Punjabi University, Patiala. 2 Asst.

More information

Understanding Purposeful Human Motion

Understanding Purposeful Human Motion M.I.T Media Laboratory Perceptual Computing Section Technical Report No. 85 Appears in Fourth IEEE International Conference on Automatic Face and Gesture Recognition Understanding Purposeful Human Motion

More information

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical

More information

A Comparative Study between SIFT- Particle and SURF-Particle Video Tracking Algorithms

A Comparative Study between SIFT- Particle and SURF-Particle Video Tracking Algorithms A Comparative Study between SIFT- Particle and SURF-Particle Video Tracking Algorithms H. Kandil and A. Atwan Information Technology Department, Faculty of Computer and Information Sciences, Mansoura University,El-Gomhoria

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes

More information

Lecture 3: Linear methods for classification

Lecture 3: Linear methods for classification Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,

More information

Real-Time Camera Tracking Using a Particle Filter

Real-Time Camera Tracking Using a Particle Filter Real-Time Camera Tracking Using a Particle Filter Mark Pupilli and Andrew Calway Department of Computer Science University of Bristol, UK {pupilli,andrew}@cs.bris.ac.uk Abstract We describe a particle

More information

A Learning Based Method for Super-Resolution of Low Resolution Images

A Learning Based Method for Super-Resolution of Low Resolution Images A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 emre.ugur@ceng.metu.edu.tr Abstract The main objective of this project is the study of a learning based method

More information

Understanding and Applying Kalman Filtering

Understanding and Applying Kalman Filtering Understanding and Applying Kalman Filtering Lindsay Kleeman Department of Electrical and Computer Systems Engineering Monash University, Clayton 1 Introduction Objectives: 1. Provide a basic understanding

More information

An Experimental Comparison of Online Object Tracking Algorithms

An Experimental Comparison of Online Object Tracking Algorithms An Experimental Comparison of Online Object Tracking Algorithms Qing Wang a, Feng Chen a, Wenli Xu a, and Ming-Hsuan Yang b a Tsinghua University, Beijing, China b University of California at Merced, Calfironia,

More information

EFFICIENT VEHICLE TRACKING AND CLASSIFICATION FOR AN AUTOMATED TRAFFIC SURVEILLANCE SYSTEM

EFFICIENT VEHICLE TRACKING AND CLASSIFICATION FOR AN AUTOMATED TRAFFIC SURVEILLANCE SYSTEM EFFICIENT VEHICLE TRACKING AND CLASSIFICATION FOR AN AUTOMATED TRAFFIC SURVEILLANCE SYSTEM Amol Ambardekar, Mircea Nicolescu, and George Bebis Department of Computer Science and Engineering University

More information

ROBUST REAL-TIME ON-BOARD VEHICLE TRACKING SYSTEM USING PARTICLES FILTER. Ecole des Mines de Paris, Paris, France

ROBUST REAL-TIME ON-BOARD VEHICLE TRACKING SYSTEM USING PARTICLES FILTER. Ecole des Mines de Paris, Paris, France ROBUST REAL-TIME ON-BOARD VEHICLE TRACKING SYSTEM USING PARTICLES FILTER Bruno Steux Yotam Abramson Ecole des Mines de Paris, Paris, France Abstract: We describe a system for detection and tracking of

More information

Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization

Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization Introduction to Mobile Robotics Bayes Filter Particle Filter and Monte Carlo Localization Wolfram Burgard, Maren Bennewitz, Diego Tipaldi, Luciano Spinello 1 Motivation Recall: Discrete filter Discretize

More information

These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop

These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop Music and Machine Learning (IFT6080 Winter 08) Prof. Douglas Eck, Université de Montréal These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher

More information

Component Ordering in Independent Component Analysis Based on Data Power

Component Ordering in Independent Component Analysis Based on Data Power Component Ordering in Independent Component Analysis Based on Data Power Anne Hendrikse Raymond Veldhuis University of Twente University of Twente Fac. EEMCS, Signals and Systems Group Fac. EEMCS, Signals

More information

Statistical Machine Learning from Data

Statistical Machine Learning from Data Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Gaussian Mixture Models Samy Bengio IDIAP Research Institute, Martigny, Switzerland, and Ecole Polytechnique

More information

3D Template-Based Single Camera Multiple Object Tracking

3D Template-Based Single Camera Multiple Object Tracking Computer Vision Winter Workshop 2006, Ondřej Chum, Vojtěch Franc (eds.) Telč, Czech Republic, February 6 8 Czech Pattern Recognition Society 3D Template-Based Single Camera Multiple Object Tracking Michal

More information

A Reliability Point and Kalman Filter-based Vehicle Tracking Technique

A Reliability Point and Kalman Filter-based Vehicle Tracking Technique A Reliability Point and Kalman Filter-based Vehicle Tracing Technique Soo Siang Teoh and Thomas Bräunl Abstract This paper introduces a technique for tracing the movement of vehicles in consecutive video

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct

More information

An Introduction to Machine Learning

An Introduction to Machine Learning An Introduction to Machine Learning L5: Novelty Detection and Regression Alexander J. Smola Statistical Machine Learning Program Canberra, ACT 0200 Australia Alex.Smola@nicta.com.au Tata Institute, Pune,

More information

Visual Tracking. Frédéric Jurie LASMEA. CNRS / Université Blaise Pascal. France. jurie@lasmea.univ-bpclermoaddressnt.fr

Visual Tracking. Frédéric Jurie LASMEA. CNRS / Université Blaise Pascal. France. jurie@lasmea.univ-bpclermoaddressnt.fr Visual Tracking Frédéric Jurie LASMEA CNRS / Université Blaise Pascal France jurie@lasmea.univ-bpclermoaddressnt.fr What is it? Visual tracking is the problem of following moving targets through an image

More information

OBJECT TRACKING USING LOG-POLAR TRANSFORMATION

OBJECT TRACKING USING LOG-POLAR TRANSFORMATION OBJECT TRACKING USING LOG-POLAR TRANSFORMATION A Thesis Submitted to the Gradual Faculty of the Louisiana State University and Agricultural and Mechanical College in partial fulfillment of the requirements

More information

Fall detection in the elderly by head tracking

Fall detection in the elderly by head tracking Loughborough University Institutional Repository Fall detection in the elderly by head tracking This item was submitted to Loughborough University's Institutional Repository by the/an author. Citation:

More information

VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS

VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS VEHICLE TRACKING USING ACOUSTIC AND VIDEO SENSORS Aswin C Sankaranayanan, Qinfen Zheng, Rama Chellappa University of Maryland College Park, MD - 277 {aswch, qinfen, rama}@cfar.umd.edu Volkan Cevher, James

More information

Tracking in flussi video 3D. Ing. Samuele Salti

Tracking in flussi video 3D. Ing. Samuele Salti Seminari XXIII ciclo Tracking in flussi video 3D Ing. Tutors: Prof. Tullio Salmon Cinotti Prof. Luigi Di Stefano The Tracking problem Detection Object model, Track initiation, Track termination, Tracking

More information

A Movement Tracking Management Model with Kalman Filtering Global Optimization Techniques and Mahalanobis Distance

A Movement Tracking Management Model with Kalman Filtering Global Optimization Techniques and Mahalanobis Distance Loutraki, 21 26 October 2005 A Movement Tracking Management Model with ing Global Optimization Techniques and Raquel Ramos Pinho, João Manuel R. S. Tavares, Miguel Velhote Correia Laboratório de Óptica

More information

Feature Tracking and Optical Flow

Feature Tracking and Optical Flow 02/09/12 Feature Tracking and Optical Flow Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem Many slides adapted from Lana Lazebnik, Silvio Saverse, who in turn adapted slides from Steve

More information

Epipolar Geometry. Readings: See Sections 10.1 and 15.6 of Forsyth and Ponce. Right Image. Left Image. e(p ) Epipolar Lines. e(q ) q R.

Epipolar Geometry. Readings: See Sections 10.1 and 15.6 of Forsyth and Ponce. Right Image. Left Image. e(p ) Epipolar Lines. e(q ) q R. Epipolar Geometry We consider two perspective images of a scene as taken from a stereo pair of cameras (or equivalently, assume the scene is rigid and imaged with a single camera from two different locations).

More information

HANDS-FREE PC CONTROL CONTROLLING OF MOUSE CURSOR USING EYE MOVEMENT

HANDS-FREE PC CONTROL CONTROLLING OF MOUSE CURSOR USING EYE MOVEMENT International Journal of Scientific and Research Publications, Volume 2, Issue 4, April 2012 1 HANDS-FREE PC CONTROL CONTROLLING OF MOUSE CURSOR USING EYE MOVEMENT Akhil Gupta, Akash Rathi, Dr. Y. Radhika

More information

2. Colour based particle filter for 3-D tracking

2. Colour based particle filter for 3-D tracking MULTI-CAMERA 3-D TRACKING USING PARTICLE FILTER Pablo Barrera, José M. Cañas, Vicente Matellán, and Francisco Martín Grupo de Robótica, Universidad Rey Juan Carlos Móstoles, Madrid (Spain) {barrera,jmplaza,vmo,fmartin}@gsyc.escet.urjc.es

More information

Accurate and robust image superresolution by neural processing of local image representations

Accurate and robust image superresolution by neural processing of local image representations Accurate and robust image superresolution by neural processing of local image representations Carlos Miravet 1,2 and Francisco B. Rodríguez 1 1 Grupo de Neurocomputación Biológica (GNB), Escuela Politécnica

More information

Low-resolution Character Recognition by Video-based Super-resolution

Low-resolution Character Recognition by Video-based Super-resolution 2009 10th International Conference on Document Analysis and Recognition Low-resolution Character Recognition by Video-based Super-resolution Ataru Ohkura 1, Daisuke Deguchi 1, Tomokazu Takahashi 2, Ichiro

More information

Static Environment Recognition Using Omni-camera from a Moving Vehicle

Static Environment Recognition Using Omni-camera from a Moving Vehicle Static Environment Recognition Using Omni-camera from a Moving Vehicle Teruko Yata, Chuck Thorpe Frank Dellaert The Robotics Institute Carnegie Mellon University Pittsburgh, PA 15213 USA College of Computing

More information

SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING

SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING AAS 07-228 SPECIAL PERTURBATIONS UNCORRELATED TRACK PROCESSING INTRODUCTION James G. Miller * Two historical uncorrelated track (UCT) processing approaches have been employed using general perturbations

More information

Active Voodoo Dolls: A Vision Based Input Device for Nonrigid Control

Active Voodoo Dolls: A Vision Based Input Device for Nonrigid Control BU CS TR98-006. To appear in Proc. Computer Animation, Philadelphia, PA, June 1998. Active Voodoo Dolls: A Vision Based Input Device for Nonrigid Control John Isidoro and Stan Sclaroff Computer Science

More information

A Short Introduction to Computer Graphics

A Short Introduction to Computer Graphics A Short Introduction to Computer Graphics Frédo Durand MIT Laboratory for Computer Science 1 Introduction Chapter I: Basics Although computer graphics is a vast field that encompasses almost any graphical

More information

Building an Advanced Invariant Real-Time Human Tracking System

Building an Advanced Invariant Real-Time Human Tracking System UDC 004.41 Building an Advanced Invariant Real-Time Human Tracking System Fayez Idris 1, Mazen Abu_Zaher 2, Rashad J. Rasras 3, and Ibrahiem M. M. El Emary 4 1 School of Informatics and Computing, German-Jordanian

More information

Analyzing Facial Expressions for Virtual Conferencing

Analyzing Facial Expressions for Virtual Conferencing IEEE Computer Graphics & Applications, pp. 70-78, September 1998. Analyzing Facial Expressions for Virtual Conferencing Peter Eisert and Bernd Girod Telecommunications Laboratory, University of Erlangen,

More information

Object tracking & Motion detection in video sequences

Object tracking & Motion detection in video sequences Introduction Object tracking & Motion detection in video sequences Recomended link: http://cmp.felk.cvut.cz/~hlavac/teachpresen/17compvision3d/41imagemotion.pdf 1 2 DYNAMIC SCENE ANALYSIS The input to

More information

Probabilistic Latent Semantic Analysis (plsa)

Probabilistic Latent Semantic Analysis (plsa) Probabilistic Latent Semantic Analysis (plsa) SS 2008 Bayesian Networks Multimedia Computing, Universität Augsburg Rainer.Lienhart@informatik.uni-augsburg.de www.multimedia-computing.{de,org} References

More information

Performance evaluation of multi-camera visual tracking

Performance evaluation of multi-camera visual tracking Performance evaluation of multi-camera visual tracking Lucio Marcenaro, Pietro Morerio, Mauricio Soto, Andrea Zunino, Carlo S. Regazzoni DITEN, University of Genova Via Opera Pia 11A 16145 Genoa - Italy

More information

Machine Learning and Pattern Recognition Logistic Regression

Machine Learning and Pattern Recognition Logistic Regression Machine Learning and Pattern Recognition Logistic Regression Course Lecturer:Amos J Storkey Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh Crichton Street,

More information

Tracking and Recognition in Sports Videos

Tracking and Recognition in Sports Videos Tracking and Recognition in Sports Videos Mustafa Teke a, Masoud Sattari b a Graduate School of Informatics, Middle East Technical University, Ankara, Turkey mustafa.teke@gmail.com b Department of Computer

More information

3D Model based Object Class Detection in An Arbitrary View

3D Model based Object Class Detection in An Arbitrary View 3D Model based Object Class Detection in An Arbitrary View Pingkun Yan, Saad M. Khan, Mubarak Shah School of Electrical Engineering and Computer Science University of Central Florida http://www.eecs.ucf.edu/

More information

UW CSE Technical Report 03-06-01 Probabilistic Bilinear Models for Appearance-Based Vision

UW CSE Technical Report 03-06-01 Probabilistic Bilinear Models for Appearance-Based Vision UW CSE Technical Report 03-06-01 Probabilistic Bilinear Models for Appearance-Based Vision D.B. Grimes A.P. Shon R.P.N. Rao Dept. of Computer Science and Engineering University of Washington Seattle, WA

More information

Human Appearance and Compound Indexing

Human Appearance and Compound Indexing Adaptive Learning of Statistical Appearance Models for 3D Human Tracking Timothy J. Roberts, Stephen J. McKenna, Ian W. Ricketts Department of Applied Computing University of Dundee, Scotland, DD1 4HN

More information

Making Machines Understand Facial Motion & Expressions Like Humans Do

Making Machines Understand Facial Motion & Expressions Like Humans Do Making Machines Understand Facial Motion & Expressions Like Humans Do Ana C. Andrés del Valle & Jean-Luc Dugelay Multimedia Communications Dpt. Institut Eurécom 2229 route des Crêtes. BP 193. Sophia Antipolis.

More information

Learning Enhanced 3D Models for Vehicle Tracking

Learning Enhanced 3D Models for Vehicle Tracking Learning Enhanced 3D Models for Vehicle Tracking J.M. Ferryman, A. D. Worrall and S. J. Maybank Computational Vision Group, Department of Computer Science The University of Reading RG6 6AY, UK J.M.Ferryman@reading.ac.uk

More information

Tracking Groups of Pedestrians in Video Sequences

Tracking Groups of Pedestrians in Video Sequences Tracking Groups of Pedestrians in Video Sequences Jorge S. Marques Pedro M. Jorge Arnaldo J. Abrantes J. M. Lemos IST / ISR ISEL / IST ISEL INESC-ID / IST Lisbon, Portugal Lisbon, Portugal Lisbon, Portugal

More information

15.062 Data Mining: Algorithms and Applications Matrix Math Review

15.062 Data Mining: Algorithms and Applications Matrix Math Review .6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop

More information

JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions

JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions Edith Cowan University Research Online ECU Publications Pre. JPEG compression of monochrome D-barcode images using DCT coefficient distributions Keng Teong Tan Hong Kong Baptist University Douglas Chai

More information

Least Squares Estimation

Least Squares Estimation Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David

More information

EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set

EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin

More information

Kristine L. Bell and Harry L. Van Trees. Center of Excellence in C 3 I George Mason University Fairfax, VA 22030-4444, USA kbell@gmu.edu, hlv@gmu.

Kristine L. Bell and Harry L. Van Trees. Center of Excellence in C 3 I George Mason University Fairfax, VA 22030-4444, USA kbell@gmu.edu, hlv@gmu. POSERIOR CRAMÉR-RAO BOUND FOR RACKING ARGE BEARING Kristine L. Bell and Harry L. Van rees Center of Excellence in C 3 I George Mason University Fairfax, VA 22030-4444, USA bell@gmu.edu, hlv@gmu.edu ABSRAC

More information

Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression

Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression Logistic Regression Department of Statistics The Pennsylvania State University Email: jiali@stat.psu.edu Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max

More information

RECOGNIZING SIMPLE HUMAN ACTIONS USING 3D HEAD MOVEMENT

RECOGNIZING SIMPLE HUMAN ACTIONS USING 3D HEAD MOVEMENT Computational Intelligence, Volume 23, Number 4, 2007 RECOGNIZING SIMPLE HUMAN ACTIONS USING 3D HEAD MOVEMENT JORGE USABIAGA, GEORGE BEBIS, ALI EROL, AND MIRCEA NICOLESCU Computer Vision Laboratory, University

More information

Tracking of Small Unmanned Aerial Vehicles

Tracking of Small Unmanned Aerial Vehicles Tracking of Small Unmanned Aerial Vehicles Steven Krukowski Adrien Perkins Aeronautics and Astronautics Stanford University Stanford, CA 94305 Email: spk170@stanford.edu Aeronautics and Astronautics Stanford

More information

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic

More information

Tutorial on Markov Chain Monte Carlo

Tutorial on Markov Chain Monte Carlo Tutorial on Markov Chain Monte Carlo Kenneth M. Hanson Los Alamos National Laboratory Presented at the 29 th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Technology,

More information

CS 688 Pattern Recognition Lecture 4. Linear Models for Classification

CS 688 Pattern Recognition Lecture 4. Linear Models for Classification CS 688 Pattern Recognition Lecture 4 Linear Models for Classification Probabilistic generative models Probabilistic discriminative models 1 Generative Approach ( x ) p C k p( C k ) Ck p ( ) ( x Ck ) p(

More information

Discrete Frobenius-Perron Tracking

Discrete Frobenius-Perron Tracking Discrete Frobenius-Perron Tracing Barend J. van Wy and Michaël A. van Wy French South-African Technical Institute in Electronics at the Tshwane University of Technology Staatsartillerie Road, Pretoria,

More information

Vehicle Tracking in Occlusion and Clutter

Vehicle Tracking in Occlusion and Clutter Vehicle Tracking in Occlusion and Clutter by KURTIS NORMAN MCBRIDE A thesis presented to the University of Waterloo in fulfilment of the thesis requirement for the degree of Master of Applied Science in

More information

Probabilistic Models for Big Data. Alex Davies and Roger Frigola University of Cambridge 13th February 2014

Probabilistic Models for Big Data. Alex Davies and Roger Frigola University of Cambridge 13th February 2014 Probabilistic Models for Big Data Alex Davies and Roger Frigola University of Cambridge 13th February 2014 The State of Big Data Why probabilistic models for Big Data? 1. If you don t have to worry about

More information

How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm

How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm IJSTE - International Journal of Science Technology & Engineering Volume 1 Issue 10 April 2015 ISSN (online): 2349-784X Image Estimation Algorithm for Out of Focus and Blur Images to Retrieve the Barcode

More information

Affine arithmetic based estimation of cue distributions in deformable model tracking

Affine arithmetic based estimation of cue distributions in deformable model tracking Affine arithmetic based estimation of cue distributions in deformable model tracking Siome Goldenstein Christian Vogler Dimitris Metaxas V.A.S.T. Laboratory - University of Pennsylvania 200 S rd Street,

More information

Simultaneous Gamma Correction and Registration in the Frequency Domain

Simultaneous Gamma Correction and Registration in the Frequency Domain Simultaneous Gamma Correction and Registration in the Frequency Domain Alexander Wong a28wong@uwaterloo.ca William Bishop wdbishop@uwaterloo.ca Department of Electrical and Computer Engineering University

More information

Interactive Offline Tracking for Color Objects

Interactive Offline Tracking for Color Objects Interactive Offline Tracking for Color Objects Yichen Wei Jian Sun Xiaoou Tang Heung-Yeung Shum Microsoft Research Asia, Beijing, China {yichenw,jiansun,xitang,hshum}@microsoft.com Abstract In this paper,

More information

An Introduction to Applied Mathematics: An Iterative Process

An Introduction to Applied Mathematics: An Iterative Process An Introduction to Applied Mathematics: An Iterative Process Applied mathematics seeks to make predictions about some topic such as weather prediction, future value of an investment, the speed of a falling

More information

Jiří Matas. Hough Transform

Jiří Matas. Hough Transform Hough Transform Jiří Matas Center for Machine Perception Department of Cybernetics, Faculty of Electrical Engineering Czech Technical University, Prague Many slides thanks to Kristen Grauman and Bastian

More information

Canny Edge Detection

Canny Edge Detection Canny Edge Detection 09gr820 March 23, 2009 1 Introduction The purpose of edge detection in general is to significantly reduce the amount of data in an image, while preserving the structural properties

More information

Towards robust automatic detection of vulnerable road users: monocular pedestrian tracking from a moving vehicle

Towards robust automatic detection of vulnerable road users: monocular pedestrian tracking from a moving vehicle Towards robust automatic detection of vulnerable road users: monocular pedestrian tracking from a moving vehicle Kristof Van Beeck 1, Floris De Smedt 1, Sander Beckers 1, Lars Struyf 1, Joost Vennekens

More information

CCNY. BME I5100: Biomedical Signal Processing. Linear Discrimination. Lucas C. Parra Biomedical Engineering Department City College of New York

CCNY. BME I5100: Biomedical Signal Processing. Linear Discrimination. Lucas C. Parra Biomedical Engineering Department City College of New York BME I5100: Biomedical Signal Processing Linear Discrimination Lucas C. Parra Biomedical Engineering Department CCNY 1 Schedule Week 1: Introduction Linear, stationary, normal - the stuff biology is not

More information

Tracking Moving Objects In Video Sequences Yiwei Wang, Robert E. Van Dyck, and John F. Doherty Department of Electrical Engineering The Pennsylvania State University University Park, PA16802 Abstract{Object

More information

Image Segmentation and Registration

Image Segmentation and Registration Image Segmentation and Registration Dr. Christine Tanner (tanner@vision.ee.ethz.ch) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation

More information

Model-based Avatar Head Animation for Interactive Virtual Reality Applications

Model-based Avatar Head Animation for Interactive Virtual Reality Applications Model-based Avatar Head Animation for Interactive Virtual Reality Applications Emil M. Petriu, Dr. Eng., FIEEE Marius D. Cordea, M.A.Sc. Michel Bondy, M.A.Sc. School of Information Technology and Engineering

More information

ILLUMINATION NORMALIZATION BASED ON SIMPLIFIED LOCAL BINARY PATTERNS FOR A FACE VERIFICATION SYSTEM. Qian Tao, Raymond Veldhuis

ILLUMINATION NORMALIZATION BASED ON SIMPLIFIED LOCAL BINARY PATTERNS FOR A FACE VERIFICATION SYSTEM. Qian Tao, Raymond Veldhuis ILLUMINATION NORMALIZATION BASED ON SIMPLIFIED LOCAL BINARY PATTERNS FOR A FACE VERIFICATION SYSTEM Qian Tao, Raymond Veldhuis Signals and Systems Group, Faculty of EEMCS University of Twente, the Netherlands

More information

Switching Gaussian Process Dynamic Models for Simultaneous Composite Motion Tracking and Recognition

Switching Gaussian Process Dynamic Models for Simultaneous Composite Motion Tracking and Recognition Switching Gaussian Process Dynamic Models for Simultaneous Composite Motion Tracking and Recognition Jixu Chen Minyoung Kim Yu Wang Qiang Ji Department of Electrical,Computer and System Engineering Rensselaer

More information

Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 269 Class Project Report

Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 269 Class Project Report Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 69 Class Project Report Junhua Mao and Lunbo Xu University of California, Los Angeles mjhustc@ucla.edu and lunbo

More information

Automatic Calibration of an In-vehicle Gaze Tracking System Using Driver s Typical Gaze Behavior

Automatic Calibration of an In-vehicle Gaze Tracking System Using Driver s Typical Gaze Behavior Automatic Calibration of an In-vehicle Gaze Tracking System Using Driver s Typical Gaze Behavior Kenji Yamashiro, Daisuke Deguchi, Tomokazu Takahashi,2, Ichiro Ide, Hiroshi Murase, Kazunori Higuchi 3,

More information

System Identification for Acoustic Comms.:

System Identification for Acoustic Comms.: System Identification for Acoustic Comms.: New Insights and Approaches for Tracking Sparse and Rapidly Fluctuating Channels Weichang Li and James Preisig Woods Hole Oceanographic Institution The demodulation

More information

Lecture 9: Introduction to Pattern Analysis

Lecture 9: Introduction to Pattern Analysis Lecture 9: Introduction to Pattern Analysis g Features, patterns and classifiers g Components of a PR system g An example g Probability definitions g Bayes Theorem g Gaussian densities Features, patterns

More information

Sampling via Moment Sharing: A New Framework for Distributed Bayesian Inference for Big Data

Sampling via Moment Sharing: A New Framework for Distributed Bayesian Inference for Big Data Sampling via Moment Sharing: A New Framework for Distributed Bayesian Inference for Big Data (Oxford) in collaboration with: Minjie Xu, Jun Zhu, Bo Zhang (Tsinghua) Balaji Lakshminarayanan (Gatsby) Bayesian

More information

A Tutorial on Particle Filters for Online Nonlinear/Non-Gaussian Bayesian Tracking

A Tutorial on Particle Filters for Online Nonlinear/Non-Gaussian Bayesian Tracking 174 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 50, NO. 2, FEBRUARY 2002 A Tutorial on Particle Filters for Online Nonlinear/Non-Gaussian Bayesian Tracking M. Sanjeev Arulampalam, Simon Maskell, Neil

More information