Incremental PCA: An Alternative Approach for Novelty Detection
|
|
|
- Charla McCormick
- 10 years ago
- Views:
Transcription
1 Incremental PCA: An Alternative Approach for Detection Hugo Vieira Neto and Ulrich Nehmzow Department of Computer Science University of Essex Wivenhoe Park Colchester CO4 3SQ {hvieir, Abstract Exploration and inspection of dynamic environments using mobile robots are applications that benefit immensely from novelty detection algorithms. In this paper we propose the use of a new approach for on-line novelty detection based on incremental Principal Component Analysis and compare its performance and functionality with a previously studied technique based on a GWR neural network. We have conducted a series of experiments using visual input from a mobile robot interacting with a controlled laboratory environment that show advantages and disadvantages for each method. 1. Introduction The ability to differentiate between normal and abnormal sensory perceptions is a desirable competence for mobile robots operating in dynamic environments usually uncommon features carry the most useful information and therefore deserve to be analysed in more detail. detection is of interest to any agent aiming at true autonomy, continuous operation and adaptability to new situations through on-line unsupervised learning. From an application point of view, reliable novelty detection systems would facilitate the implementation of automated inspection and surveillance. As for these tasks one commonly desires to detect any previously unknown feature as opposed to recognition tasks in which features of interest are already known the feasible approach to be followed is to learn a model of normality and use it to filter out abnormal perceptions. Previous work has demonstrated that the approach of learning a model of normality instead of abnormality is very effective in mobile robots that use sonar readings as perceptual input (Marsland et al., 2002). This work resulted in the development of the Grow-When-Required (GWR) neural network, a self-organising learning mechanism able to determine whether an input is novel or not through the use of a model of habituation. The GWR network has shown to work well with lowdimensional input data, such as a vector of sonar measurements, but in the past few years we have been interested in investigating the scalability of the GWR approach to another rather different sensor modality: vision. Our interest in using vision for novelty detection comes from the wide range of different information from the environment that it can provide the robot with. A major difficulty that comes with vision is how to select which aspects of the data are important to be encoded, as it is undesirable to process high-dimensional data directly. So far, we have successfully employed a mechanism of attention to select salient locations of the input image frame and perform some image encoding in their vicinity (Vieira Neto and Nehmzow, 2004, Nehmzow and Vieira Neto, 2004). The purpose of the image encoding stage is to reduce dimensionality of input data to the novelty filter while trying to preserve discriminability between different classes of features as much as possible. Designing the image encoding stage is a hard task as it is not always clear to the designer which aspects of the data are important to be encoded. It would be more desirable to select which aspects are the most relevant in a bottom-up approach and hence our interest in the Principal Component Analysis (PCA) algorithm. PCA consists in projecting the data onto principal axes the axes in which variance is maximised obtained from the data itself. The standard PCA algorithm is usually computed in batch mode and needs all the input data to be available at once, making it unsuitable for on-line learning systems. Recently, however, an algorithm for incremental computation of PCA was introduced by Artač et al. (Artač et al., 2002) and we have identified its potential for novelty detection. In this paper we implement an incremental PCA novelty filter and compare its performance with the already well-known GWR-based approach. Figure 1 shows the
2 block diagram that illustrates the framework in which both methods were tested. Environment Camera (1) Raw Image Attention Mechanism (2) Image patches Filter (3) Indication of Figure 1: The visual novelty detection framework: salient image patches are selected within the images acquired from the environment and then either of the novelty filters is used to indicate the presence of novelty. In order to generate the input for the novelty filters we have employed a visual attention mechanism to select smaller image patches within the input image frame. We briefly describe the attention mechanism in the next section, followed by both techniques for novelty detection. Finally, we present in detail the experimental setup in which we conducted our tests and discuss the results obtained. 2. Visual attention The high dimensionality of visual data imposes serious constraints to real-time processing, especially when using mobile robots with restricted computational resources. It is therefore prohibitive to process an entire image frame, even when using low resolution the use of some sort of dimensionality reduction scheme that preserves important and distinctive features of the visual data is needed before any higher level processing can be done. Moreover, images acquired from a moving platform are subject to several geometrical transformations resulting from changes in perspective. Therefore, the naïve approach of comparing entire image frames directly would most definitely not work properly, leading to the misclassification of known features as novel due to simple image transformations. In this paper we have used the saliency map (Itti et al., 1998) as a mechanism of visual attention to select distinctive regions in the input image frame ( pixels in size). The saliency map consists in the combination of several multi-scale feature maps in intensity, colour and orientation of visual features, allowing the detection of conspicuous locations within the image frame that are generally robust to geometric transformations. Details of our implementation of the saliency map can be found in (Vieira Neto and Nehmzow, 2004). We have used the nine highest values of the saliency map for each input frame to select the most interesting image regions so that image patches can be extracted from their vicinity. The image patches used in the experiments reported here have pixels in size. As we have used RGB colour images, the resulting input vectors for the two alternative novelty filters being compared have = 1728 elements. Input vectors were normalised before being fed to the novelty filters to even out lighting conditions. As in this work the similarity between image patches is computed in a pixel-by-pixel basis using the Euclidean distance, it is important to have the patches aligned as much as possible to minimise errors. The saliency map has an important role in this task as the location of salient points tend to be very stable and therefore minimises image patch misregistration. At the moment our current implementation only handles translation effects, but we are currently investigating algorithms to achieve invariance to scale and rotation (Lowe, 2004), and also invariance to affine transformations (Mikolajczyk and Schmid, 2002). In this paper, generalisation with respect to scaling, rotation and affine transformations is achieved by means of the acquisition of several vectors for geometrically transformed versions of the same visual features. 3. filters 3.1 The GWR network The Grow-When-Required (GWR) is an artificial neural network that was especially designed for the task of on-line novelty detection (Marsland et al., 2002). It combines a clustering mechanism and a model of habituation to decide if a determined input is novel and therefore needs to be incorporated to the current model. A summary of the operation of the GWR network is given in Algorithm 1. The network is initialised with two unconnected nodes, whose weights are set to the two first input vectors. As input vectors are presented, the algorithm is able to decide which represent novel features based on how well they match existing habituated nodes, building and exploiting a topological map in input space. We have previously used the GWR network for visual novelty detection (Vieira Neto and Nehmzow, 2004, Nehmzow and Vieira Neto, 2004), using local colour histograms as input vectors to achieve dimensionality reduction. In the experiments reported here, we used normalised raw image patches in order to fairly compare performance and functionality with an alternative novelty detection algorithm based on incremental PCA. In this work, the parameters used for the GWR net-
3 Algorithm 1: GWR network novelty detection Input: current set of nodes A, current set of connections C, new input vector x. Output: updated set of nodes A, updated set of connections C, novelty indication N. Find the best and second best matching nodes s and t: s = arg min x w i, i A (n) t = arg min x w i, where w i is the weight i A (n) /{s} vector of the node i. If there is a connection between s and t, set its age to zero, otherwise create it: C = C {(s, t)}). Compute the activity of the best matching node: a s = exp( x w s ). Test if the activity and habituation values of the best matching node characterise novelty: if a s < a T and h s < h T then Add a new node: A = A {r}. Set the weight vector of the new node: w r = (x + w s )/2. Create connections between the new node and the best and second best matching nodes: C = C {(r, s), (r, t)}. Remove the connection between the best and second best matching nodes: C = C/{(s, t)}. Indicate novelty detected: N = 1. end else Indicate no novelty detected: N = 0. Compute the activity of the best matching node s neighbours (nodes with connections to the best matching node): a n = exp( x w n ). Adapt the positions of the best matching node and its neighbours: w s = w s + ɛ(x w s ), w n = w n + ηan a s ɛ(x w n ). Age connections to the best matching node: age (s,n) = age (s,n) + 1. Habituate the best matching node and its neighbours: τ dhs(t) = α[h 0 h s (t)] S(t), a s ηa n τ dhn(t) dt dt = α[h 0 h n (t)] S(t). Remove any nodes without any neighbours. Remove any connections with age greater than age max. work were: a T = 0.9, h T = 0.3, η = 0.1, ɛ = 0.1, τ = 3.33, α = 1.05, h 0 = 1, S(t) = 1 and age max = Incremental PCA Principal Component Analysis (PCA) is a very useful tool for dimensionality reduction that allows optimal reconstruction of the original data. It consists on projecting the input data onto its principal axes and is usually computed off-line, as the standard algorithm requires that all data samples are available a priori, making it unsuitable for applications that demand on-line learning. However, a method for the incremental computation of PCA recently introduced by Artač et al. (Artač et al., 2002) makes simultaneous learning and recognition possible. Their technique allows the original data to be discarded immediately after the eigenspace is updated, storing only the (reduced dimension) data projected onto it. In this work we employ the method proposed by Artač et al. to perform on-line novelty detection, using the magnitude of the residual vector, i.e. the RMS error between original data and the reconstruction of its projection onto the current eigenspace, to classify the input as novel or not. Algorithm 2 summarises how this approach is implemented Algorithm 2: Incremental PCA novelty detection Input: current mean vector µ (n), current eigenvectors U (n), current projected vectors A (n), new input vector x, residual threshold r T. Output: updated mean vector µ (n+1), updated eigenvectors U (n+1), updated projected vectors A (n+1), novelty indication N. Compute the projection of the new input vector using the current basis: a = U (n)t (x µ (n) ). Compute the reconstruction of the new input vector: y = U (n) a + µ (n). Compute the residual vector (orthogonal to U (n) ): r = x y. Test if the magnitude of the residual vector is large enough to characterise novelty: if r > r T then Append residual vector as a new basis vector: [ ] U = U (n). r r Append projected vector: [ ] A A (n) a =. 0 r Perform batch PCA on A, obtaining its mean vector µ and eigenvectors U. Update projected vectors using the new basis: A (n+1) = U (A µ 1 1 n+1 ). Update eigenvectors: U (n+1) = U U. Update mean vector: µ (n+1) = µ (n) + U µ. Indicate novelty detected: N = 1. end else Indicate no novelty detected: N = 0. The algorithm is made completely incremental by initialising the eigenspace and projected vectors as follows: µ (1) = x (1), U (1) = 0 M 1 and A (1) = 0, where x (1) is
4 the first input vector and 0 M 1 denotes an M 1 matrix of zeros, M being the dimensionality of the input. In this approach, dimensionality reduction is achieved by exploiting the fact that the number of eigenvectors in the model are likely to be less in number than the dimensionality of the input vectors ( = 1728). If all eigenvectors are kept in the model, perfect reconstruction of the original data is achieved. This functionality allows the user to reconstruct the input image patches from the stored projected vectors and have a perfect notion of which aspects of the environment were learnt. Further dimensionality reduction can be achieved by keeping only the eigenvectors corresponding to the k < n largest eigenvalues in the model at the expense of losses in reconstruction (and possibly in the recognition rate of the system). The selection of eigenvectors can be done while computing the batch PCA during learning (step 7 in Algorithm 2). In this work we have set the threshold for the magnitude of the residual vector as r T = 0.2 and we kept only eigenvectors whose corresponding eigenvalues were larger than 1% of the largest eigenvalue. 4. Experiments In order to compare the performance of the aforementioned novelty detection mechanisms we have designed a series of experiments consisting of two different phases: an exploration phase in which the robot learns a visual model of normality for its operating environment; and an inspection phase in which the learnt model is used to highlight any abnormal visual feature that may appear in the environment. We have built a bounded square arena with cardboard boxes and then used our Magellan Pro mobile robot to collect images while navigating using a simple obstacle avoidance behaviour. Although both novelty filters being compared in this paper are able to run in real-time, the images were collected for off-line processing in order to make fair comparisons by using the same dataset for both methods. The performance of each method was evaluated by relating the actual presence of novelty and the response of the system in the form of a contingency table and later performing the χ 2 analysis, followed by the computation of Cramer s V and the uncertainty coefficient U. We have also compared the size of the models acquired by the two approaches. Ground truth for the presence of novelty in each frame was generated manually in the form of binary images with the novel features highlighted. Any image patch selected by the mechanism of visual attention containing at least 10% of highlighted pixels was considered as novel for the purposes of performance analysis. Every round of the exploration or inspection phases consisted of five consecutive loops around the arena. 4.1 Results The first round of experiments comprised the robot exploring the empty arena in order to build a model of normality for it. After that, the robot inspected the arena when two different objects were introduced in it: a very conspicuous orange football and a much less conspicuous grey box. The results obtained for both approaches are given in Table 1. Table 1: Experiment 1 - exploration of the empty arena and inspection of the arena containing novel objects (orange football and a grey box). GWR Network Not Cramer s V = 0.86 Uncertainty Coefficient U = 0.73 Incremental PCA Not Cramer s V = 0.88 Uncertainty Coefficient U = 0.76 Both systems were able to highlight the novel objects, yielding statistically significant results according to the χ 2 analysis (p < 0.05). It can be noticed from Table 1 that results obtained for both systems were very similar and that the contingency table entries add up to 4500 samples, which correspond to 9 salient regions per image frame in a total of 500 frames (50 image frames per loop in a total of 10 loops around the arena). Figure 2 visually depicts the results obtained for frames where the orange football and the grey box appeared, respectively. For the particular image frames shown in Figure 2, results were exactly the same for both approaches. In order to establish whether or not the novelty detection systems were able to highlight novel but not very salient features, a second round of experiments was conducted. This time the robot has explored the arena containing the orange football and inspected it with the inclusion of the grey box in two different situations: first in a different location and then in the same location as the ball. Both systems were able to successfully detect the novel object in each situation, regardless of the presence
5 (a) Figure 3: Results obtained with the visual novelty detection mechanism in experiment 2: the grey box is correctly detected as novel regardless of the presence of the much more salient but already known orange football. The output shown is the one provided by the incremental PCA approach, as the GWR-based approach only labelled region 2 as novel. (b) Figure 2: Results obtained with the visual novelty detection mechanism in experiment 1: (a) the orange football as new object; and (b) the grey box as new object. White circles designate image patches labelled as novel by the system, while numbers indicate salient locations within the image frame (lower numbers indicate an item of greater saliency). or not of the much more salient orange ball in the same image frame, as shown in Figure 3. The results obtained for this round of experiments are given in Table 2. Again statistically significant (χ 2 analysis, p < 0.05) for both systems, but slightly better for the incremental PCA. Finally, we conducted a third round of experiments in order to evaluate the ability of the systems to differentiate between visually similar objects. This time the robot explored the arena containing the grey box and inspected it with another grey box, identical in colour and texture, but larger in size. Two instances were analysed: the larger box in the same location as the previous smaller box was and also in a different location in the arena. Table 2: Experiment 2 - exploration of the arena containing the orange football and inspection of the arena containing a novel object (a grey box in the same location as the ball and in a different location as the ball). GWR Network Not Cramer s V = 0.58 Uncertainty Coefficient U = 0.31 Incremental PCA Not Cramer s V = 0.70 Uncertainty Coefficient U = 0.44 Both approaches were able to correctly detect the larger grey box as novel and the results obtained are given in Table 3. Once more, results were statistically significant (χ 2 analysis, p < 0.05) for both systems, but better for incremental PCA. Figure 4 shows an example of output obtained during the third round of experiments.
6 Table 3: Experiment 3 - exploration of the arena containing the grey box and inspection of the arena containing a novel object (a larger grey box visually similar to the original box in the same location and in a different location). GWR Network Not Cramer s V = 0.37 Uncertainty Coefficient U = 0.11 Incremental PCA Not Cramer s V = 0.53 Uncertainty Coefficient U = 0.23 Figure 4: Results obtained with the visual novelty detection mechanism in experiment 3: the larger grey box is correctly detected as novel. The output shown is the one provided by the incremental PCA approach, as the GWR-based approach only labelled regions 0 and 8 as novel. 4.2 Discussion The results given in Tables 1, 2 and 3 demonstrate the statistical significance in the association of the responses of both approaches and the actual novel features introduced in the robot s original environment. Although the strength of the association measured by Cramer s V and the uncertainty coefficient U can be thought as modest, we consider these values as very conservative for two reasons. First, if we take into account the consistency of novelties detected between successive image frames we are able to rule out most false positives (novelty detected but not present). And second, most false negatives (novelty present but not detected) can be eliminated by the fact that a single image patch within the new object labelled as new is enough to characterise the entire object as novel. Nevertheless, the values of V and U serve well to the purpose of comparing performances. Results given by both approaches are similar in performance, although the size of the models acquired by each one are very different for the set of parameters used: in experiment 1 the GWR acquired only five vectors while the incremental PCA acquired 35; in experiment 2 the GWR acquired a total of 18 vectors and the incremental PCA 45; finally, in experiment 3, 11 vectors were acquired by the GWR and 47 by the incremental PCA. The small amount of vectors learnt by the GWR had always the original input dimensionality (1728 elements), while the dimensionality of the vectors acquired by incremental PCA varied from 29 to 33 dimensions. However, every dimension of the projected vectors acquired by the incremental PCA approach corresponds to an eigenvector with 1728 elements. Dimensionality issues become important when we consider that the Euclidean metric was used to determine similarity between vectors. When the Euclidean distance is used, a small difference between two high-dimensional vectors tend to be large in value, making it difficult to establish thresholds of similarity for high-dimensional spaces, as it is the case with the vectors acquired by the GWR network. The incremental PCA approach offers a clear advantage as similarity between inputs is performed by the residual error in reconstruction from the projected space. Moreover, if a direct comparison of projected vectors is to be made, substitution of the Euclidean distance by the Mahalanobis distance can be easily implemented in the incremental PCA approach once the covariance matrix of the stored projected vectors is available as a sub-product of the method (step 7 in Algorithm 2). The Mahalanobis distance normalises the contribution of vector elements according to the covariance matrix of the data: d xy = (x y) T C 1 (x y) (1) where d xy is the Mahalanobis distance between the column vectors x and y and C is the covariance matrix of the data. Euclidean distance corresponds to the special case where C is the identity matrix. Another advantage of the incremental PCA approach is the ability to automatically reduce dimensionality allowing reconstruction of the original input image patch from the inverse transformation of the corresponding projected vector. Therefore, the user can evaluate which aspects of the environment were actually learnt by the
7 system. Reconstruction of the stored vectors in the GWR network resulted in averaged image patches resulting from the learning procedure (step 13 in Algorithm 1). The GWR network, however, has the advantage of building a topological map for the stored vectors, through connections between similar patterns. We have experimented increasing the number of stored vectors in the GWR approach by raising the activation threshold a T in order to acquire a number of vectors as close as possible to the number of vectors acquired by the incremental PCA. This resulted in better reconstruction of the stored vectors, but has also sensibly decreased the overall performance of the GWR-based system. As one would expect, the number of false negatives has decreased, but on the other hand the number of false positives has increased immensely. We attribute this effect to the use of Euclidean distance in a high-dimensional space. 5. Conclusion We have presented an alternative to perform on-line novelty detection using an incremental PCA approach and compared its performance against a previously studied approach based on the GWR network. The proposed incremental PCA approach provides slightly better performance, while offering advantages of embedded dimensionality reduction and good reconstruction ability, which is extremely useful to assess which aspects of the environment were actually learnt by the system. The inability of the GWR network of evaluating similarity between inputs in reduced dimensions normally forces the system designer to use an additional preprocessing stage for dimensionality reduction, such as the use of colour statistics (Vieira Neto and Nehmzow, 2004). On the other hand, the GWR approach offers the functionality of constructing a topological relationship between inputs. Future investigations aim at combining the dimensionality reduction feature of the incremental PCA with the topological construction algorithm of the GWR network using the Mahalanobis distance as a measure of similarity between patterns. Considering the overall system functionality, the attention mechanism plays an important role in generalisation by providing image patches that are robust to translations and therefore reducing the number of stored vectors. We are currently studying alternatives for the attention model that also offer invariance to scale and rotation (Lowe, 2004), and also affine transformations (Mikolajczyk and Schmid, 2002). Extensions to the incremental PCA algorithm to make it robust to occlusions (Skočaj and Leonardis, 2003) are also attractive for future work. A final contribution of this paper is the introduction of a method to evaluate performance of novelty detection systems based on contingency table analysis. A quantitative assessment can be made by the computation of Cramer s V and the uncertainty coefficient U, while the statistical significance of the association between system response and actual novelty status can be made by χ 2 analysis. Acknowledgements Hugo Vieira Neto is sponsored by the Brazilian Government through CAPES Foundation and CEFET-PR, whose support is gratefully acknowledged. References Artač, M., Jogan, M., and Leonardis, A. (2002). Incremental PCA for on-line visual learning and recognition. In Proceedings of the 16th International Conference on Pattern Recognition (ICPR 2002). Itti, L., Koch, C., and Niebur, E. (1998). A model of saliency-based visual attention for rapid scene analysis. IEEE Transactions on Pattern Analysis and Machine Intelligence, 20(11): Lowe, D. G. (2004). Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision, 60(2): Marsland, S., Nehmzow, U., and Shapiro, J. (2002). Environment-specific novelty detection. In From Animals to Animats: Proceedings of the 7th International Conference on the Simulation of Adaptive Behaviour (SAB 02), Edinburgh, UK. MIT Press. Mikolajczyk, K. and Schmid, C. (2002). An affine invariant interest point detector. In Computer Vision - ECCV 2002: 7th European Conference on Computer Vision (ECCV 2002), volume 1, pages , Copenhagen, Denmark. Nehmzow, U. and Vieira Neto, H. (2004). based visual inspection using mobile robots. In Towards Autonomous Robotic Systems: Proceedings of the 5th British Conference on Mobile Robotics (TAROS 04), Colchester, UK. Skočaj, D. and Leonardis, A. (2003). Weighted and robust incremental method for subspace learning. In Proceedings of the 9th International Conference on Computer Vision (ICCV 03), pages Vieira Neto, H. and Nehmzow, U. (2004). Visual novelty detection for inspection tasks using mobile robots. In Proceedings of the 8th Brazilian Symposium on Neural Networks (SBRN 2004), São Luís, Brazil.
Subspace Analysis and Optimization for AAM Based Face Alignment
Subspace Analysis and Optimization for AAM Based Face Alignment Ming Zhao Chun Chen College of Computer Science Zhejiang University Hangzhou, 310027, P.R.China [email protected] Stan Z. Li Microsoft
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Journal of Industrial Engineering Research. Adaptive sequence of Key Pose Detection for Human Action Recognition
IWNEST PUBLISHER Journal of Industrial Engineering Research (ISSN: 2077-4559) Journal home page: http://www.iwnest.com/aace/ Adaptive sequence of Key Pose Detection for Human Action Recognition 1 T. Sindhu
Machine Learning and Pattern Recognition Logistic Regression
Machine Learning and Pattern Recognition Logistic Regression Course Lecturer:Amos J Storkey Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh Crichton Street,
Accurate and robust image superresolution by neural processing of local image representations
Accurate and robust image superresolution by neural processing of local image representations Carlos Miravet 1,2 and Francisco B. Rodríguez 1 1 Grupo de Neurocomputación Biológica (GNB), Escuela Politécnica
Volume 2, Issue 9, September 2014 International Journal of Advance Research in Computer Science and Management Studies
Volume 2, Issue 9, September 2014 International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case Study Available online at: www.ijarcsms.com
3D Model based Object Class Detection in An Arbitrary View
3D Model based Object Class Detection in An Arbitrary View Pingkun Yan, Saad M. Khan, Mubarak Shah School of Electrical Engineering and Computer Science University of Central Florida http://www.eecs.ucf.edu/
Object Recognition and Template Matching
Object Recognition and Template Matching Template Matching A template is a small image (sub-image) The goal is to find occurrences of this template in a larger image That is, you want to find matches of
Novelty Detection in image recognition using IRF Neural Networks properties
Novelty Detection in image recognition using IRF Neural Networks properties Philippe Smagghe, Jean-Luc Buessler, Jean-Philippe Urban Université de Haute-Alsace MIPS 4, rue des Frères Lumière, 68093 Mulhouse,
Vision based Vehicle Tracking using a high angle camera
Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu [email protected] [email protected] Abstract A vehicle tracking and grouping algorithm is presented in this work
Tracking performance evaluation on PETS 2015 Challenge datasets
Tracking performance evaluation on PETS 2015 Challenge datasets Tahir Nawaz, Jonathan Boyle, Longzhen Li and James Ferryman Computational Vision Group, School of Systems Engineering University of Reading,
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
Probabilistic Latent Semantic Analysis (plsa)
Probabilistic Latent Semantic Analysis (plsa) SS 2008 Bayesian Networks Multimedia Computing, Universität Augsburg [email protected] www.multimedia-computing.{de,org} References
Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall
Automatic Photo Quality Assessment Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Estimating i the photorealism of images: Distinguishing i i paintings from photographs h Florin
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
High Quality Image Magnification using Cross-Scale Self-Similarity
High Quality Image Magnification using Cross-Scale Self-Similarity André Gooßen 1, Arne Ehlers 1, Thomas Pralow 2, Rolf-Rainer Grigat 1 1 Vision Systems, Hamburg University of Technology, D-21079 Hamburg
Nonlinear Iterative Partial Least Squares Method
Numerical Methods for Determining Principal Component Analysis Abstract Factors Béchu, S., Richard-Plouet, M., Fernandez, V., Walton, J., and Fairley, N. (2016) Developments in numerical treatments for
Visibility optimization for data visualization: A Survey of Issues and Techniques
Visibility optimization for data visualization: A Survey of Issues and Techniques Ch Harika, Dr.Supreethi K.P Student, M.Tech, Assistant Professor College of Engineering, Jawaharlal Nehru Technological
Fast Matching of Binary Features
Fast Matching of Binary Features Marius Muja and David G. Lowe Laboratory for Computational Intelligence University of British Columbia, Vancouver, Canada {mariusm,lowe}@cs.ubc.ca Abstract There has been
Component Ordering in Independent Component Analysis Based on Data Power
Component Ordering in Independent Component Analysis Based on Data Power Anne Hendrikse Raymond Veldhuis University of Twente University of Twente Fac. EEMCS, Signals and Systems Group Fac. EEMCS, Signals
An Energy-Based Vehicle Tracking System using Principal Component Analysis and Unsupervised ART Network
Proceedings of the 8th WSEAS Int. Conf. on ARTIFICIAL INTELLIGENCE, KNOWLEDGE ENGINEERING & DATA BASES (AIKED '9) ISSN: 179-519 435 ISBN: 978-96-474-51-2 An Energy-Based Vehicle Tracking System using Principal
Augmented Reality Tic-Tac-Toe
Augmented Reality Tic-Tac-Toe Joe Maguire, David Saltzman Department of Electrical Engineering [email protected], [email protected] Abstract: This project implements an augmented reality version
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
Edge tracking for motion segmentation and depth ordering
Edge tracking for motion segmentation and depth ordering P. Smith, T. Drummond and R. Cipolla Department of Engineering University of Cambridge Cambridge CB2 1PZ,UK {pas1001 twd20 cipolla}@eng.cam.ac.uk
Manifold Learning Examples PCA, LLE and ISOMAP
Manifold Learning Examples PCA, LLE and ISOMAP Dan Ventura October 14, 28 Abstract We try to give a helpful concrete example that demonstrates how to use PCA, LLE and Isomap, attempts to provide some intuition
Face detection is a process of localizing and extracting the face region from the
Chapter 4 FACE NORMALIZATION 4.1 INTRODUCTION Face detection is a process of localizing and extracting the face region from the background. The detected face varies in rotation, brightness, size, etc.
Self Organizing Maps: Fundamentals
Self Organizing Maps: Fundamentals Introduction to Neural Networks : Lecture 16 John A. Bullinaria, 2004 1. What is a Self Organizing Map? 2. Topographic Maps 3. Setting up a Self Organizing Map 4. Kohonen
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Tracking in flussi video 3D. Ing. Samuele Salti
Seminari XXIII ciclo Tracking in flussi video 3D Ing. Tutors: Prof. Tullio Salmon Cinotti Prof. Luigi Di Stefano The Tracking problem Detection Object model, Track initiation, Track termination, Tracking
Adaptive Face Recognition System from Myanmar NRC Card
Adaptive Face Recognition System from Myanmar NRC Card Ei Phyo Wai University of Computer Studies, Yangon, Myanmar Myint Myint Sein University of Computer Studies, Yangon, Myanmar ABSTRACT Biometrics is
BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES
BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES 123 CHAPTER 7 BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES 7.1 Introduction Even though using SVM presents
Understanding and Applying Kalman Filtering
Understanding and Applying Kalman Filtering Lindsay Kleeman Department of Electrical and Computer Systems Engineering Monash University, Clayton 1 Introduction Objectives: 1. Provide a basic understanding
Automatic parameter regulation for a tracking system with an auto-critical function
Automatic parameter regulation for a tracking system with an auto-critical function Daniela Hall INRIA Rhône-Alpes, St. Ismier, France Email: [email protected] Abstract In this article we propose
Reconstructing 3D Pose and Motion from a Single Camera View
Reconstructing 3D Pose and Motion from a Single Camera View R Bowden, T A Mitchell and M Sarhadi Brunel University, Uxbridge Middlesex UB8 3PH [email protected] Abstract This paper presents a
COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS
COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS B.K. Mohan and S. N. Ladha Centre for Studies in Resources Engineering IIT
Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data
CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear
Face Recognition in Low-resolution Images by Using Local Zernike Moments
Proceedings of the International Conference on Machine Vision and Machine Learning Prague, Czech Republic, August14-15, 014 Paper No. 15 Face Recognition in Low-resolution Images by Using Local Zernie
Class-specific Sparse Coding for Learning of Object Representations
Class-specific Sparse Coding for Learning of Object Representations Stephan Hasler, Heiko Wersing, and Edgar Körner Honda Research Institute Europe GmbH Carl-Legien-Str. 30, 63073 Offenbach am Main, Germany
A Comparison of Photometric Normalisation Algorithms for Face Verification
A Comparison of Photometric Normalisation Algorithms for Face Verification James Short, Josef Kittler and Kieron Messer Centre for Vision, Speech and Signal Processing University of Surrey Guildford, Surrey,
CATEGORIZATION OF SIMILAR OBJECTS USING BAG OF VISUAL WORDS AND k NEAREST NEIGHBOUR CLASSIFIER
TECHNICAL SCIENCES Abbrev.: Techn. Sc., No 15(2), Y 2012 CATEGORIZATION OF SIMILAR OBJECTS USING BAG OF VISUAL WORDS AND k NEAREST NEIGHBOUR CLASSIFIER Piotr Artiemjew, Przemysław Górecki, Krzysztof Sopyła
Introduction to Pattern Recognition
Introduction to Pattern Recognition Selim Aksoy Department of Computer Engineering Bilkent University [email protected] CS 551, Spring 2009 CS 551, Spring 2009 c 2009, Selim Aksoy (Bilkent University)
Automatic Labeling of Lane Markings for Autonomous Vehicles
Automatic Labeling of Lane Markings for Autonomous Vehicles Jeffrey Kiske Stanford University 450 Serra Mall, Stanford, CA 94305 [email protected] 1. Introduction As autonomous vehicles become more popular,
Learning Enhanced 3D Models for Vehicle Tracking
Learning Enhanced 3D Models for Vehicle Tracking J.M. Ferryman, A. D. Worrall and S. J. Maybank Computational Vision Group, Department of Computer Science The University of Reading RG6 6AY, UK [email protected]
Efficient Attendance Management: A Face Recognition Approach
Efficient Attendance Management: A Face Recognition Approach Badal J. Deshmukh, Sudhir M. Kharad Abstract Taking student attendance in a classroom has always been a tedious task faultfinders. It is completely
Biometric Authentication using Online Signatures
Biometric Authentication using Online Signatures Alisher Kholmatov and Berrin Yanikoglu [email protected], [email protected] http://fens.sabanciuniv.edu Sabanci University, Tuzla, Istanbul,
Going Big in Data Dimensionality:
LUDWIG- MAXIMILIANS- UNIVERSITY MUNICH DEPARTMENT INSTITUTE FOR INFORMATICS DATABASE Going Big in Data Dimensionality: Challenges and Solutions for Mining High Dimensional Data Peer Kröger Lehrstuhl für
Feature Selection vs. Extraction
Feature Selection In many applications, we often encounter a very large number of potential features that can be used Which subset of features should be used for the best classification? Need for a small
Principal components analysis
CS229 Lecture notes Andrew Ng Part XI Principal components analysis In our discussion of factor analysis, we gave a way to model data x R n as approximately lying in some k-dimension subspace, where k
AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION
AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION Saurabh Asija 1, Rakesh Singh 2 1 Research Scholar (Computer Engineering Department), Punjabi University, Patiala. 2 Asst.
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 269 Class Project Report
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 69 Class Project Report Junhua Mao and Lunbo Xu University of California, Los Angeles [email protected] and lunbo
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
Review Jeopardy. Blue vs. Orange. Review Jeopardy
Review Jeopardy Blue vs. Orange Review Jeopardy Jeopardy Round Lectures 0-3 Jeopardy Round $200 How could I measure how far apart (i.e. how different) two observations, y 1 and y 2, are from each other?
Face Model Fitting on Low Resolution Images
Face Model Fitting on Low Resolution Images Xiaoming Liu Peter H. Tu Frederick W. Wheeler Visualization and Computer Vision Lab General Electric Global Research Center Niskayuna, NY, 1239, USA {liux,tu,wheeler}@research.ge.com
Environmental Remote Sensing GEOG 2021
Environmental Remote Sensing GEOG 2021 Lecture 4 Image classification 2 Purpose categorising data data abstraction / simplification data interpretation mapping for land cover mapping use land cover class
Introduction to Support Vector Machines. Colin Campbell, Bristol University
Introduction to Support Vector Machines Colin Campbell, Bristol University 1 Outline of talk. Part 1. An Introduction to SVMs 1.1. SVMs for binary classification. 1.2. Soft margins and multi-class classification.
Simultaneous Gamma Correction and Registration in the Frequency Domain
Simultaneous Gamma Correction and Registration in the Frequency Domain Alexander Wong [email protected] William Bishop [email protected] Department of Electrical and Computer Engineering University
CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C.
CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES From Exploratory Factor Analysis Ledyard R Tucker and Robert C MacCallum 1997 180 CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES In
Supporting Online Material for
www.sciencemag.org/cgi/content/full/313/5786/504/dc1 Supporting Online Material for Reducing the Dimensionality of Data with Neural Networks G. E. Hinton* and R. R. Salakhutdinov *To whom correspondence
A Survey on Outlier Detection Techniques for Credit Card Fraud Detection
IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661, p- ISSN: 2278-8727Volume 16, Issue 2, Ver. VI (Mar-Apr. 2014), PP 44-48 A Survey on Outlier Detection Techniques for Credit Card Fraud
Image Segmentation and Registration
Image Segmentation and Registration Dr. Christine Tanner ([email protected]) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation
Open-Set Face Recognition-based Visitor Interface System
Open-Set Face Recognition-based Visitor Interface System Hazım K. Ekenel, Lorant Szasz-Toth, and Rainer Stiefelhagen Computer Science Department, Universität Karlsruhe (TH) Am Fasanengarten 5, Karlsruhe
University of Lille I PC first year list of exercises n 7. Review
University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients
Limitations of Human Vision. What is computer vision? What is computer vision (cont d)?
What is computer vision? Limitations of Human Vision Slide 1 Computer vision (image understanding) is a discipline that studies how to reconstruct, interpret and understand a 3D scene from its 2D images
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS Norbert Buch 1, Mark Cracknell 2, James Orwell 1 and Sergio A. Velastin 1 1. Kingston University, Penrhyn Road, Kingston upon Thames, KT1 2EE,
Local features and matching. Image classification & object localization
Overview Instance level search Local features and matching Efficient visual recognition Image classification & object localization Category recognition Image classification: assigning a class label to
Visualization of large data sets using MDS combined with LVQ.
Visualization of large data sets using MDS combined with LVQ. Antoine Naud and Włodzisław Duch Department of Informatics, Nicholas Copernicus University, Grudziądzka 5, 87-100 Toruń, Poland. www.phys.uni.torun.pl/kmk
Force/position control of a robotic system for transcranial magnetic stimulation
Force/position control of a robotic system for transcranial magnetic stimulation W.N. Wan Zakaria School of Mechanical and System Engineering Newcastle University Abstract To develop a force control scheme
Least-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
ROBOTRACKER A SYSTEM FOR TRACKING MULTIPLE ROBOTS IN REAL TIME. by Alex Sirota, [email protected]
ROBOTRACKER A SYSTEM FOR TRACKING MULTIPLE ROBOTS IN REAL TIME by Alex Sirota, [email protected] Project in intelligent systems Computer Science Department Technion Israel Institute of Technology Under the
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA
A PHOTOGRAMMETRIC APPRAOCH FOR AUTOMATIC TRAFFIC ASSESSMENT USING CONVENTIONAL CCTV CAMERA N. Zarrinpanjeh a, F. Dadrassjavan b, H. Fattahi c * a Islamic Azad University of Qazvin - [email protected]
Poker Vision: Playing Cards and Chips Identification based on Image Processing
Poker Vision: Playing Cards and Chips Identification based on Image Processing Paulo Martins 1, Luís Paulo Reis 2, and Luís Teófilo 2 1 DEEC Electrical Engineering Department 2 LIACC Artificial Intelligence
Object class recognition using unsupervised scale-invariant learning
Object class recognition using unsupervised scale-invariant learning Rob Fergus Pietro Perona Andrew Zisserman Oxford University California Institute of Technology Goal Recognition of object categories
Medical Information Management & Mining. You Chen Jan,15, 2013 [email protected]
Medical Information Management & Mining You Chen Jan,15, 2013 [email protected] 1 Trees Building Materials Trees cannot be used to build a house directly. How can we transform trees to building materials?
Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability
Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller
Bayesian Image Super-Resolution
Bayesian Image Super-Resolution Michael E. Tipping and Christopher M. Bishop Microsoft Research, Cambridge, U.K..................................................................... Published as: Bayesian
T-61.3050 : Email Classification as Spam or Ham using Naive Bayes Classifier. Santosh Tirunagari : 245577
T-61.3050 : Email Classification as Spam or Ham using Naive Bayes Classifier Santosh Tirunagari : 245577 January 20, 2011 Abstract This term project gives a solution how to classify an email as spam or
Linear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
Are we ready for Autonomous Driving? The KITTI Vision Benchmark Suite
Are we ready for Autonomous Driving? The KITTI Vision Benchmark Suite Philip Lenz 1 Andreas Geiger 2 Christoph Stiller 1 Raquel Urtasun 3 1 KARLSRUHE INSTITUTE OF TECHNOLOGY 2 MAX-PLANCK-INSTITUTE IS 3
The Delicate Art of Flower Classification
The Delicate Art of Flower Classification Paul Vicol Simon Fraser University University Burnaby, BC [email protected] Note: The following is my contribution to a group project for a graduate machine learning
CS231M Project Report - Automated Real-Time Face Tracking and Blending
CS231M Project Report - Automated Real-Time Face Tracking and Blending Steven Lee, [email protected] June 6, 2015 1 Introduction Summary statement: The goal of this project is to create an Android
Colour Image Segmentation Technique for Screen Printing
60 R.U. Hewage and D.U.J. Sonnadara Department of Physics, University of Colombo, Sri Lanka ABSTRACT Screen-printing is an industry with a large number of applications ranging from printing mobile phone
Terrain Traversability Analysis using Organized Point Cloud, Superpixel Surface Normals-based segmentation and PCA-based Classification
Terrain Traversability Analysis using Organized Point Cloud, Superpixel Surface Normals-based segmentation and PCA-based Classification Aras Dargazany 1 and Karsten Berns 2 Abstract In this paper, an stereo-based
Graduate Co-op Students Information Manual. Department of Computer Science. Faculty of Science. University of Regina
Graduate Co-op Students Information Manual Department of Computer Science Faculty of Science University of Regina 2014 1 Table of Contents 1. Department Description..3 2. Program Requirements and Procedures
Visualization of Breast Cancer Data by SOM Component Planes
International Journal of Science and Technology Volume 3 No. 2, February, 2014 Visualization of Breast Cancer Data by SOM Component Planes P.Venkatesan. 1, M.Mullai 2 1 Department of Statistics,NIRT(Indian
Relating Vanishing Points to Catadioptric Camera Calibration
Relating Vanishing Points to Catadioptric Camera Calibration Wenting Duan* a, Hui Zhang b, Nigel M. Allinson a a Laboratory of Vision Engineering, University of Lincoln, Brayford Pool, Lincoln, U.K. LN6
Introduction. Chapter 1
Chapter 1 Introduction The area of fault detection and diagnosis is one of the most important aspects in process engineering. This area has received considerable attention from industry and academia because
Tracking Moving Objects In Video Sequences Yiwei Wang, Robert E. Van Dyck, and John F. Doherty Department of Electrical Engineering The Pennsylvania State University University Park, PA16802 Abstract{Object
Feature Tracking and Optical Flow
02/09/12 Feature Tracking and Optical Flow Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem Many slides adapted from Lana Lazebnik, Silvio Saverse, who in turn adapted slides from Steve
Towards Online Recognition of Handwritten Mathematics
Towards Online Recognition of Handwritten Mathematics Vadim Mazalov, joint work with Oleg Golubitsky and Stephen M. Watt Ontario Research Centre for Computer Algebra Department of Computer Science Western
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
Segmentation & Clustering
EECS 442 Computer vision Segmentation & Clustering Segmentation in human vision K-mean clustering Mean-shift Graph-cut Reading: Chapters 14 [FP] Some slides of this lectures are courtesy of prof F. Li,
An Overview of Knowledge Discovery Database and Data mining Techniques
An Overview of Knowledge Discovery Database and Data mining Techniques Priyadharsini.C 1, Dr. Antony Selvadoss Thanamani 2 M.Phil, Department of Computer Science, NGM College, Pollachi, Coimbatore, Tamilnadu,
Tracking Groups of Pedestrians in Video Sequences
Tracking Groups of Pedestrians in Video Sequences Jorge S. Marques Pedro M. Jorge Arnaldo J. Abrantes J. M. Lemos IST / ISR ISEL / IST ISEL INESC-ID / IST Lisbon, Portugal Lisbon, Portugal Lisbon, Portugal
Perception-based Design for Tele-presence
Perception-based Design for Tele-presence Santanu Chaudhury 1, Shantanu Ghosh 1,2, Amrita Basu 3, Brejesh Lall 1, Sumantra Dutta Roy 1, Lopamudra Choudhury 3, Prashanth R 1, Ashish Singh 1, and Amit Maniyar
