View-Invariant Dynamic Texture Recognition using a Bag of Dynamical Systems
|
|
|
- Margaret Greer
- 10 years ago
- Views:
Transcription
1 View-Invariant Dynamic Texture Recognition using a Bag of Dynamical Systems Avinash Ravichandran, Rizwan Chaudhry and René Vidal Center for Imaging Science, Johns Hopkins University, Baltimore, MD 21218, USA Abstract In this paper, we consider the problem of categorizing videos of dynamic textures under varying view-point. We propose to model each video with a collection of Linear Dynamics Systems (LDSs) describing the dynamics of spatiotemporal video patches. This bag of systems (BoS) representation is analogous to the bag of features (BoF) representation, except that we use LDSs as feature descriptors. This poses several technical challenges to the BoF framework. Most notably, LDSs do not live in a Euclidean space, hence novel methods for clustering LDSs and computing codewords of LDSs need to be developed. Our framework makes use of nonlinear dimensionality reduction and clustering techniques combined with the Martin distance for LDSs for tackling these issues. Our experiments show that our BoS approach can be used for recognizing dynamic textures in challenging scenarios, which could not be handled by existing dynamic texture recognition methods. 1. Introduction Dynamic textures are video sequences of complex scenes such as water on the surface of a lake, a flag fluttering in the wind, etc. The constant change in the geometry of the these objects poses a challenge for applying traditional vision algorithms to these video sequences. Over the past few years, several algorithms have been proposed for vision tasks such as registration, segmentation and categorization of such scenes. Among these tasks, the categorization of such video sequences is of specific interest to us, because it is critical to surveillance applications, such as detecting fires in tunnels, leaks in pipes on ships, etc. Doretto et al. in [6], showed that dynamic textures can be modeled using a LDS. Using the LDS representation, several categorization algorithms have been proposed based on the LDS parameters, which live in a non-euclidean space. One popular genre of methods first defines a distance or a kernel between the model parameters of two dynamical systems. Once such a distance or kernel has been defined, classifiers such as nearest neighbors or support vector machines (SVM) [7] can be used to categorize a query video sequence based on the training data. Among these methods, Saisan et al. [12] used distances based on the principal angles between the subspaces generated by the parameters of the LDSs. Vishwanathan et al. [15], used Binet-Cauchy kernels to compare the parameters of two LDSs. Further, Chan et al. used both the KL divergence [1] and the Martin distance [2] as a metric between dynamical system. Finally, Woolfe et al. [18] used the family of Chernoff distances and distances between cepstrum coefficients as a metric between LDSs. Other types of approaches for dynamic texture categorization, such as Fujita et al. [8], divide the video sequences into blocks and compare the trajectories of the states in order to perform the inference. Alternatively, Vidal et al. [14] extended boosting to LDSs by using dynamical systems as weak classifiers. Although there exists several methods for dynamic texture categorization, most of these methods have validated their performance on the database from [12], which consists of 50 classes with 4 video sequences per class. However, if one examines this database, several of the classes are from a single category, e.g., water, but from different viewpoints. Thus, in reality there are only 9 different semantic categories. In addition to the lack of intraclass variability, most existing methods work with a manually extracted subsequence at the center of the video, as opposed to using the whole video sequence. In this paper, we aim to address a more challenging problem, namely the categorization of dynamic textures with invariance to changes in view point. To the best of our knowledge, apart from [18], which addressed shift invariance, there is no other work addressing the dynamic texture categorization problem with invariance to viewpoint. The bag of features (BoF) approach for object recognition has been shown to handle variability in viewpoints, illumination and scales. Recently, the BoF method was extended to categorizing video sequences [5, 17, 16, 9], wherein traditional image features were replaced by spatiotemporal features. Specifically, once feature points are extracted, a descriptor is formed based on gradients in a spatiotemporal neighborhood around the feature point. After this step, the pipeline is exactly the same as in BoF for images. In principle, one could follow the same approach for categorizing dynamic textures. However, the fact that dynamic textures can be modeled with a LDS would not be exploited /09/$ IEEE 1651
2 Moreover, videos of dynamic textures tend to be fairly homogeneous. Even though existing detectors can be used to find feature points in such video sequences, using gradients as descriptors may not be discriminative enough. Paper Contributions. In this paper, we propose to extend the BoF approach to dynamic textures, by replacing traditional features by LDSs. By using LDSs as features, we obtain an alternate representation of these video sequences which preserves their dynamics, as opposed to image gradients. The fact that the parameters of a LDS live in a non-euclidean space poses several challenges to the BoF framework. We propose a method to form a codebook from these non-euclidean parameters by performing nonlinear dimensionality reduction and clustering in the space of LDSs. We evaluate various weighting schemes used to represent a video sequence using the codebook. We also show that by using the BoS approach, we are able to recognize dynamic textures under changes in viewpoint and outperform existing methods. 2. Preliminaries In this section, we first introduce the dynamic texture framework and show how video sequences can be modeled using this framework. We then introduce the Martin distance between two LDSs. This distance will serve as our similarity metric in the Bag of Systems framework Dynamic Texture Model Given a video sequence or a spatiotemporal patch, {I(t)} F t=1, we model it as the output of a LDS as z(t + 1) = Az(t) + Bv(t) (1) I(t) = Cz(t) + w(t), (2) where z(t) R n is the hidden state at time t, A R n n represents the dynamics of the system, C R p n maps the hidden state to the output of the system, and w(t) N (0, R) and v(t) N (0, Q) are the measurement and process noise respectively. The order of the system is given by n and p is the number of pixels in one frame of the sequence or patch. Given a video sequence, the model parameters can be identified using a Principal Component Analysis (PCA) based approach outlined in [6]. As a consequence of the identification algorithm, the parameter C belongs to the Stiefel Manfiold ST(p, n) = {C C C = I R n n }. The advantage of this model is that, it decouples the appearance of the spatiotemporal patch, which is modeled by C, from the dynamics, which are represented by A. Hence, given a spatiotemporal patch, we describe it using the tuple M = (A, C). Such a feature descriptor models both the dynamics and the appearance in the spatiotemporal patch as opposed to gradients that only model local texture Subspace Angles Based Distance Between LDSs In any categorization algorithm, features from a new query video sequence need to be compared with the features from the training set. Given two feature descriptors, M 1 = (A 1, C 1 ) and M 2 = (A 2, C 2 ), we first need to define a notion of similarity between these two descriptors. One such family of distances between two LDSs is based on the subspace angles between the two systems. The subspace angles are defined as the principal angles between the observability subspaces associated with the two model parameters. The observability subspace is the range-space of the extended observability matrix of the LDS defined by O (M) = [C, (CA), (CA 2 ),...] R n. The calculation of the subspace angles between the two models is performed by first solving for P from the Lyapunov equation A PA P = C C, where P = [ ] P11 P 12 R 2n 2n, A = P 21 P 22 C = [ C 1 C 2 ] R p 2n. [ ] A1 0 R 2n 2n, 0 A 2 The cosines of the subspace angles {θ i } n i=1 are calculated as cos 2 θ i = i-th eigenvalue(p 1 11 P 12P 1 22 P 21). (3) The Martin distance between M 1 and M 2 is now defined as 3. Bag of Systems d M (M 1, M 2 ) 2 = ln n cos 2 θ i. (4) i=1 The BoF approach in computer vision draws its inspiration from the document retrieval community. Here, a document is hypothesized to be identifiable by the distribution of certain keywords. In the BoF approach, a similar philosophy is adopted, wherein feature descriptors are used in lieu of words. Images are now categorized by observing the distribution of a small collection of features called codewords. The typical steps followed in the BoF framework are: (1) features and their corresponding descriptors are extracted from all the images in the training set, (2) a codebook is formed using clustering methods such as K-means, where the cluster centers represent codewords in the codebook, (3) each image in the training dataset is represented using the codebook, (4) a classifier is chosen to compare a new query image to the training set and thus infer its category. In this section we describe the corresponding BoS analogue of the BoF steps. We wish to point out that our approach is not a mere change in the feature descriptor. As a consequence of using the parameters of a LDS as a feature descriptor, we now have features in a non-euclidean space. Therefore, clustering these descriptors is highly challenging. 1652
3 I(t) GL(n) ST(p, n) {M i,d M (, )} R de {e i, 2 } NLDR x x (6σ + 1) (6σ + 1) (6τ + 1) {F i } {k i } VIDEO SEQUENCE CODEWORDS CLUSTER CENTERS Figure 1. Overview of the bag of systems framework for categorizing dynamic textures with invariance to changes in view point. In this paper we propose a framework for codebook formation by clustering in the space of LDSs. An overview of our framework is shown in Fig Feature Extraction and Descriptors The first step in our framework is to extract feature points from the video sequences. Over the past few years, several spatiotemporal feature point methods have been proposed. Laptev et al. [9] extend the Harris corner detector used for images to video sequences by considering a second moment matrix of the spatiotemporal gradients. Willems et al. [16] use the spatiotemporal Hessian. Dollar et al. [5] convolve the video sequence spatially with a Gaussian filter and temporally with a quadrature pair of 1D Gabor filters. Local maxima of this response function are then selected as feature points. Wong et al. [17] model the video sequence with a LDS and extract features from the parameters of the LDS, as opposed to directly from the video sequence. A detailed comparison of these methods can be found in [16]. These methods however, have concentrated on classifying video sequences of human activity, gestures and mouse activity. In this paper, we use existing feature point selection methods for detecting feature points. However, a key difference is that we use LDS parameters as a feature descriptor for a spatiotemporal volume. Once a feature point and its corresponding spatiotemporal volume is extracted, a LDS is identified from this volume as described in 2.1, and represented using M = (A, C) Codebook Formation In the traditional BoF framework, once feature points and their corresponding descriptors are extracted, the descriptors are clustered using an algorithm such as K-means to form the codebook. However, in our case the descriptors are parameters of a LDS, i.e. (A, C), which lie on a non-euclidean space. More specifically, (A, C) GL(n) ST(p, n), where GL(n) is the group of all invertible matrices of size n and ST(p, n) is the Stiefel manifold. Hence, one cannot directly apply clustering algorithms that are used in the Euclidean case. However, one could find a low-dimensional Euclidean embedding of these points and apply the clustering algorithm in this space. There exist several techniques for Non-Linear Dimensionality Reduction (NLDR). Some methods such as Locally Linear Embedding (LLE) [11], work directly with the points in the higher dimensional space, while other methods such as Multidimensional Scaling (MDS) [3], Isomap [13] work with the pairwise distances between the points in the higher dimensional space. In our case, we exploit the fact that the space of LDSs is endowed with several distances, e.g., the Martin distance ( 2.2), and use this distance to perform dimensionality reduction and clustering to form the codebook. Given the set of features {M i } T i=1, where T represents the total number of features extracted from the videos in the training set, we first form the matrix D R T T such that D ij = d M (M i, M j ). Once pairwise distances are available between the features, techniques such as Multidimensional Scaling (MDS), Isomap [13], etc. can be used to obtain a low-dimensional embedding of these points {e i R de } T i=1, where d e is the dimension of the embedding. This lowdimensional representation gives us a set of Euclidean points, which preserve the relationships in the high-dimensional nonlinear space. Clustering algorithms can be applied to {e i } T i=1, as the the low-dimensional space is Euclidean. After this clustering, we now have K cluster centers {k i } K i=1. However, these cluster centers do not correspond to any of the original LDSs. Moreover, there is no explicit map to go from the lower dimensional embedding to the original space. Hence, in order to select our codewords {F i } K i=1, we choose the corresponding systems in the highdimensional space whose distance to the cluster center in the lower dimensional space is the least, i.e. F i = M p, p = arg min j e j k i 2. (5) 1653
4 In this way, we obtain our codebook C = {F 1,..., F K }, where F i = (A i, C i ). During the query phase, each detected feature is associated with model parameters M = (A, C). The membership to the codeword is given by m = arg min i d M (M, F i ) Representing Videos using the Codebook Once the K codewords are available, each video sequence needs to be represented using this vocabulary. This is done using a weight vector W = {w 1, w 2,... w K } R K. There are several choices for such a representation. Let us assume that codeword k occurs N ki times in the i th video sequence and there are total of N i codewords in the i th video sequence. Let V be the total number of video sequences and V i be the number of video sequences in which codeword i occurs. The simplest representation is called the term frequency (TF) and is defined as w ik = N ki N i. (6) A more popular approach is the term frequency - inverse document frequency (TF-IDF) defined as w ik = N ki N i ( ) V ln. (7) V i Another recently proposed approach termed as the softweighting (SW) is defined as k 0 L i w ik = i=1 j=1 1 2 i 1 d(f k, M j ) (8) where k 0 represents the number of neighbors (typically k 0 = 4) and L i represents the number of codewords that have F k as their i th neighbor. This method as opposed to the TF, exploits the entire structure of the codebook. Once a weight vector W is computed, it is normalized by its L 1 norm to become a histogram. In the experimental section, we evaluate the various representation schemes outlined above and comment on their optimality for our approach Classification The final step in our framework is to classify a new query video sequence using the training data. Given the training dataset {(W i, l i )} V i=1, where l i {1,..., m} is the class label of the weight vector, our goal is to infer the class label of a new weight vector W t. To compare the training weight vectors and the query weight vector we can use standard distances between histograms such as the χ 2 distance or the square root distance on the sphere [7], which are defined as d χ 2(W 1, W 2 ) = 1 2 K i=1 w 1i w 2i w 1i + w 2i, (9) K d S (W 1, W 2 ) = arccos( w1i w 2i ). (10) i=1 A simple classification approach is the k-nearest neighbors (k-nn) classifier [7], where the query video sequence is assigned the majority class label of its k closest weight vectors from the training dataset. A generative classification approach is the naive Bayes (NB) classifier used in Csurka et al. [4]. The posterior i.e. P (l = j I i (t)) is proportional to the product of the prior P (l = j) and the likelihood P (I i (t) l = j). Given the fact that the video sequences I i (t) can be represented by the codewords and assuming independence of the codewords, the likelihood on the codewords are learnt using the equation P (F k l = j) = 1 + {i l i=j} N ki K + K s=1 {i l i=j} N si. (11) Given a new video sequence I n (t), the inference is then computed as the maximum aposterori estimate of P (l = j I n (t)) P (l = j)p (I n (t) l = j) (12) K = P (l = j) P (F s l = j) Nsn (13) s=1 A discriminative classification approach is SVM with kernels. The standard kernel used for the weight vector is the radial basis kernel, defined as K(W 1, W 2 ) = e γd(w1,w2), where γ is a free parameter usually learnt by cross validation and d(w 1, W 2 ) is any distance on the space of histograms Implementation Details In our framework, we use the approach outlined in [5] for extracting the features. We find this feature detector more appropriate for dynamic textures than those of [9, 16], which are based on image gradients. Once a feature point is detected at a spatial scale σ and temporal scale τ, a volume of size (6σ + 1) (6σ + 1) (6τ + 1) centered around the feature point is extracted. For each of these volumes, a dynamic texture model of order n is identified using PCAbased approach. Unlike [9, 16], we do not have a method to calculate scales automatically. Hence, we use multiple values of σ and τ, in order to detect feature points at different scales. Also, the Martin distance can be applied to systems of different orders, but requires the dimensions of the output spaces of the two systems, i.e. the number of pixels in a frame, to be the same. Hence, we normalize the spatial size 1654
5 Smoke Boiling Fire Flowers Sea Water Fountain Waterfall Figure 2. Sample snapshots from the reorganized dynamic texture database. This represents the 8 classes we consider in our framework, while in the original database these are considered as 22 classes. of the spatiotemporal volumes to be uniform. We however do not require the same temporal size for these volumes. In order to perform the clustering in the low-dimensional space, we use Hierarchical K-means (HKM)[10]. It is well known that K-means is sensitive to initialization. Hence, K-means is run multiple times and the cluster centers with the minimum intraclass variance are selected as the cluster centers. Since we want stable codewords, as opposed to minimum variance, we run HKM multiple times and select the most frequent cluster centers as our codewords. In our experiments, since the number of codewords used was small, we directly use the leaves of the HKM as the cluster centers. 4. Experiments In this section we present a detailed experimental evaluation of various aspects of our algorithm. We use the dataset from [12]. This dataset consists of 50 classes of 4 video sequences each. The state of the art recognition results on this dataset reported by Chan et al. in [2], is 97.5% by using the kernel version of the dynamic texture model and 96.5% by using just the Martin distance on the LDS parameters. We wish to point out that these numbers are based on the 50 class structure and the reported results are not on the entire video sequences, but on a manually extracted patch of size However, if one combines the sequences that are taken from different viewpoints, the dataset can be reduced to a 9 class dataset with the classes being boiling water (8), fire (8), flowers (12), fountains (20), plants (108), sea (12), smoke (4), water (12) and waterfall (16). Here the numbers represent the number of sequences in the dataset. Since the number of sequences of plants far outnumbered the number of sequences for the other classes, we used the remaining 8 classes in our experiments. Sample frames from the video sequences in the database are show in Fig. 2. We show results for four different scenarios in this paper. We first consider an easy two class problem namely the case of water vs. fountain. Then a more challenging two class problem namely the fountain vs. waterfall. Our third set of experiments is on a four class problem and the last set is on the reorganized database with 8 classes. All the results presented in this section have been averaged over 20 trials. Codebook formation. We experimented with the two NLDR methods, namely MDS and Isomap. However, we noticed that, with all other experimental conditions remaining the same, recognition rates by using MDS were higher than those of Isomap. We also found the optimal codebook size to be K = 8 and the number of restarts for the HKM to be 16. Also, the order of the LDSs was set to n = 8 and we did not optimize this parameter. Representation and Classification. Once the codebook is formed, we have 3 choices for representing a video sequence, namely TF, TF-IDF and SW, and two choices of distances between histograms, namely χ 2 distance and the square root distance on the sphere. In addition, the classifier can k-nn, naive Bayes or SVM. Overall this gives us 14 ways to perform the inference, given the codebook. In our experiments, we noticed that the TF-IDF representation exhibited inconsistent performance. For some scenarios, this was the best representation and for others this was the worst. Since we would like a stable representation, which works across different scenarios, we omit this representation in our analyis. We also noticed that using the square root distance on the sphere always marginally lowered the recognition rate. Hence, we present results based only on the χ 2 distance. As a baseline for comparison, we categorize the different scenarios using the existing single dynamic texture framework as outlined in [2, 12]. We model the entire video sequence using a single LDS. We calculate distances between LDSs using the Martin distance and categorize using both the nearest neighbor and SVM classifiers. We abbreviate these methods as DK and DS respectively. The methods based on the TF representation are abbreviated as TF-1NN, TF-3NN for the k-nn using k = 1 and k = 3 neighbors, and TF-SVM for the SVM classification. Similarly the methods using the SW representation are abbreviated as SW-1NN, SW-3NN and SW-SVM. Note that for the naive Bayes classifier there is no representation (TF or SW) that is required. Water vs. Fountain. These classes are visually very different. Hence, we expect high classification rates for this problem. We consider 12 video sequences from the two classes and train using 50% of the data and test using the rest. The classification results for the various metrics are 1655
6 Four Classes Eight Classes Recognition Results Recognition Results TF-1NN SW-1NN TF-3NN SW-3NN NB TF-SVM SW-SVM 0 DK DS Water Vs Fountain Fountain Vs Waterfall Four Classes Eight Classes (a) Results using our approach. Figure 3. Classification results for the different scenarios. (b) Results using [12] (a) Our approach (b) Approach from [12] Figure 4. Confusion matrix for the 8 class scenario. shown in Fig. 3. Although DK and DS are not designed to be invariant to view points, we notice that their performance on this scenario are 67% and % respectively. Our proposed method is also able to achieve % classification rate. Fountain vs. Waterfall. This is a more challenging example in which there are no clear appearance differences. We consider 16 video sequences for each class from the database. We train using 50% of the data and test using the rest. It must be noted that in this scenario, the testing involved video sequence from novel viewpoints, which were not used in the training. The classification results for the various metrics are shown in Fig. 3. The baseline results for this case are 50% using DK and 56.25% using DS. Our results for this scenario significantly outperform the baseline at 98%. Notice from the two class scenarios that the best method to use is the naive Bayes approach. The SW-SVM approach is the second best method and outperforms the baseline. However, using the naive Bayes for more than two classes results in poor performance, while the SW-SVM still performs well, as we will results shown below. Four Class. This scenario combines the categories from the above two scenarios along with the sea category. We use a total of 12 sequences per class and train using 50% of the data and test using the rest. The classification results for this scenario are shown in Fig. 3. We see that the proposed method is still able to outperform the baseline of 75% and 83% using DK and DS respectively. The best classification result we obtain using the BoS approach is 89%. In this scenario, one can notice that the increase in the classification rate is only 6%. This is primarily because there is only one novel viewpoint in the testing phase. Although the baseline methods are not invariant to changes in view point, in this database they are extremely effective in recognizing video sequences in the same viewpoint. Eight Class. Our last and most challenging scenario is using all the classes from the database. We have a total of 88 video sequences with varying number of sequences per class. Similar to the other scenarios, we trained on 50% of the dataset and tested on the rest. The classification rates for this scenario are shown in Fig. 3. Our approach has a classification rate of 80%, while DK and DS are at 52.27% and 47.73% respectively. The confusion matrix for BoS and DK is shown in Fig. 4. In our methods, the smoke, fountain and water categories exhibit lower classification rates than the other categories. Except for the case of the water category, the confusion is between closely related categories. The DK method on the other hand, shows poor performance for all categories except sea, smoke and waterfall. In fact, most of the categories are wrongly recognized as waterfall. 1656
7 5. Conclusions In this paper, we have presented a method for the categorization of dynamic textures which is invariant to view point changes. Traditional methods cannot handle this variability and this fact is clearly reflected in the results we have presented. For the easy two class problem, our approach performs as well as the traditional methods. While for the hard two class problem, we show a gain of approximately 42%. In the case of the multiclass scenarios, the increase in performance is approximately 6% and 28% for four and eight classes, respectively. We achieve this by modeling a video sequence with a bag of of LDSs as opposed to using a single model for the video. To the best of our knowledge, no other existing work addresses this problem. We have also evaluated the various weighting schemes. Based on our experiments, we observed that for a two class problem, using the naive Bayes classifier is the best approach. However, for higher number of classes using the SW with SVM is the best approach. [12] P. Saisan, G. Doretto, Y. N. Wu, and S. Soatto. Dynamic texture recognition. In CVPR, volume II, pages 58 63, [13] J. B. Tenenbaum, V. de Silva, and J. C. Langford. A global geometric framework for nonlinear dimensionality reduction. Science, 290(5500): , [14] R. Vidal and P. Favaro. Dynamicboost: Boosting time series generated by dynamical systems. In ICCV, [15] S. Vishwanathan, A. Smola, and R. Vidal. Binet-Cauchy kernels on dynamical systems and its application to the analysis of dynamic scenes. IJCV, 73(1):95 119, [16] G. Willems, T. Tuytelaars, and L. J. V. Gool. An efficient dense and scale-invariant spatio-temporal interest point detector. In ECCV, volume 5303, pages , [17] S.-F. Wong and R. Cipolla. Extracting spatiotemporal interest points using global information. In ICCV, pages 1 8, [18] F. Woolfe and A. Fitzgibbon. Shift-invariant dynamic texture recognition. In ECCV, pages II: , Acknowledgements This work was partially supported by startup funds from JHU, by grants ONR N , ONR N , NSF CAREER and ARL Robotics-CTA 80014MC. References [1] A. Chan and N. Vasconcelos. Probabilistic kernels for the classification of auto-regressive visual processes. In CVPR, volume 1, pages , [2] A. Chan and N. Vasconcelos. Classifying video with kernel dynamic textures. In CVPR, pages 1 6, [3] T. F. Cox and M. A. A. Cox. Multidimensional Scaling. Chapman and Hall, [4] C. Dance, J. Willamowski, L. Fan, C. Bray, and G. Csurka. Visual categorization with bags of keypoints. In ECCV, [5] P. Dollár, V. Rabaud, G. Cottrell, and S. Belongie. Behavior recognition via sparse spatio-temporal features. In VS-PETS, October [6] G. Doretto, A. Chiuso, Y. Wu, and S. Soatto. Dynamic textures. IJCV, 51(2):91 109, [7] R. Duda, P. Hart, and D. Stork. Pattern Classification. Wiley- Interscience, October [8] K. Fujita and S. Nayar. Recognition of Dynamic Textures using Impulse Responses of State Variables. In Texture 2003, Oct [9] I. Laptev. On space-time interest points. IJCV, 64(2-3): , [10] D. Nister and H. Stewenius. Scalable recognition with a vocabulary tree. In CVPR, pages , [11] S. Roweis and L. Saul. Nonlinear dimensionality reduction by locally linear embedding. Science, 290(5500): ,
Probabilistic Latent Semantic Analysis (plsa)
Probabilistic Latent Semantic Analysis (plsa) SS 2008 Bayesian Networks Multimedia Computing, Universität Augsburg [email protected] www.multimedia-computing.{de,org} References
Local features and matching. Image classification & object localization
Overview Instance level search Local features and matching Efficient visual recognition Image classification & object localization Category recognition Image classification: assigning a class label to
Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data
CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear
Recognizing Cats and Dogs with Shape and Appearance based Models. Group Member: Chu Wang, Landu Jiang
Recognizing Cats and Dogs with Shape and Appearance based Models Group Member: Chu Wang, Landu Jiang Abstract Recognizing cats and dogs from images is a challenging competition raised by Kaggle platform
CATEGORIZATION OF SIMILAR OBJECTS USING BAG OF VISUAL WORDS AND k NEAREST NEIGHBOUR CLASSIFIER
TECHNICAL SCIENCES Abbrev.: Techn. Sc., No 15(2), Y 2012 CATEGORIZATION OF SIMILAR OBJECTS USING BAG OF VISUAL WORDS AND k NEAREST NEIGHBOUR CLASSIFIER Piotr Artiemjew, Przemysław Górecki, Krzysztof Sopyła
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
Fast Matching of Binary Features
Fast Matching of Binary Features Marius Muja and David G. Lowe Laboratory for Computational Intelligence University of British Columbia, Vancouver, Canada {mariusm,lowe}@cs.ubc.ca Abstract There has been
Learning Motion Categories using both Semantic and Structural Information
Learning Motion Categories using both Semantic and Structural Information Shu-Fai Wong, Tae-Kyun Kim and Roberto Cipolla Department of Engineering, University of Cambridge, Cambridge, CB2 1PZ, UK {sfw26,
Recognition. Sanja Fidler CSC420: Intro to Image Understanding 1 / 28
Recognition Topics that we will try to cover: Indexing for fast retrieval (we still owe this one) History of recognition techniques Object classification Bag-of-words Spatial pyramids Neural Networks Object
3D Model based Object Class Detection in An Arbitrary View
3D Model based Object Class Detection in An Arbitrary View Pingkun Yan, Saad M. Khan, Mubarak Shah School of Electrical Engineering and Computer Science University of Central Florida http://www.eecs.ucf.edu/
Group Sparse Coding. Fernando Pereira Google Mountain View, CA [email protected]. Dennis Strelow Google Mountain View, CA strelow@google.
Group Sparse Coding Samy Bengio Google Mountain View, CA [email protected] Fernando Pereira Google Mountain View, CA [email protected] Yoram Singer Google Mountain View, CA [email protected] Dennis Strelow
Object Recognition. Selim Aksoy. Bilkent University [email protected]
Image Classification and Object Recognition Selim Aksoy Department of Computer Engineering Bilkent University [email protected] Image classification Image (scene) classification is a fundamental
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
So which is the best?
Manifold Learning Techniques: So which is the best? Todd Wittman Math 8600: Geometric Data Analysis Instructor: Gilad Lerman Spring 2005 Note: This presentation does not contain information on LTSA, which
Cluster Analysis: Advanced Concepts
Cluster Analysis: Advanced Concepts and dalgorithms Dr. Hui Xiong Rutgers University Introduction to Data Mining 08/06/2006 1 Introduction to Data Mining 08/06/2006 1 Outline Prototype-based Fuzzy c-means
Visualization by Linear Projections as Information Retrieval
Visualization by Linear Projections as Information Retrieval Jaakko Peltonen Helsinki University of Technology, Department of Information and Computer Science, P. O. Box 5400, FI-0015 TKK, Finland [email protected]
siftservice.com - Turning a Computer Vision algorithm into a World Wide Web Service
siftservice.com - Turning a Computer Vision algorithm into a World Wide Web Service Ahmad Pahlavan Tafti 1, Hamid Hassannia 2, and Zeyun Yu 1 1 Department of Computer Science, University of Wisconsin -Milwaukee,
Journal of Industrial Engineering Research. Adaptive sequence of Key Pose Detection for Human Action Recognition
IWNEST PUBLISHER Journal of Industrial Engineering Research (ISSN: 2077-4559) Journal home page: http://www.iwnest.com/aace/ Adaptive sequence of Key Pose Detection for Human Action Recognition 1 T. Sindhu
Cees Snoek. Machine. Humans. Multimedia Archives. Euvision Technologies The Netherlands. University of Amsterdam The Netherlands. Tree.
Visual search: what's next? Cees Snoek University of Amsterdam The Netherlands Euvision Technologies The Netherlands Problem statement US flag Tree Aircraft Humans Dog Smoking Building Basketball Table
Tracking and Recognition in Sports Videos
Tracking and Recognition in Sports Videos Mustafa Teke a, Masoud Sattari b a Graduate School of Informatics, Middle East Technical University, Ankara, Turkey [email protected] b Department of Computer
HDDVis: An Interactive Tool for High Dimensional Data Visualization
HDDVis: An Interactive Tool for High Dimensional Data Visualization Mingyue Tan Department of Computer Science University of British Columbia [email protected] ABSTRACT Current high dimensional data visualization
The Delicate Art of Flower Classification
The Delicate Art of Flower Classification Paul Vicol Simon Fraser University University Burnaby, BC [email protected] Note: The following is my contribution to a group project for a graduate machine learning
Statistical Models in Data Mining
Statistical Models in Data Mining Sargur N. Srihari University at Buffalo The State University of New York Department of Computer Science and Engineering Department of Biostatistics 1 Srihari Flood of
Face Recognition in Low-resolution Images by Using Local Zernike Moments
Proceedings of the International Conference on Machine Vision and Machine Learning Prague, Czech Republic, August14-15, 014 Paper No. 15 Face Recognition in Low-resolution Images by Using Local Zernie
A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow
, pp.233-237 http://dx.doi.org/10.14257/astl.2014.51.53 A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow Giwoo Kim 1, Hye-Youn Lim 1 and Dae-Seong Kang 1, 1 Department of electronices
Automatic parameter regulation for a tracking system with an auto-critical function
Automatic parameter regulation for a tracking system with an auto-critical function Daniela Hall INRIA Rhône-Alpes, St. Ismier, France Email: [email protected] Abstract In this article we propose
Analysis of kiva.com Microlending Service! Hoda Eydgahi Julia Ma Andy Bardagjy December 9, 2010 MAS.622j
Analysis of kiva.com Microlending Service! Hoda Eydgahi Julia Ma Andy Bardagjy December 9, 2010 MAS.622j What is Kiva? An organization that allows people to lend small amounts of money via the Internet
International Journal of Computer Science Trends and Technology (IJCST) Volume 3 Issue 3, May-June 2015
RESEARCH ARTICLE OPEN ACCESS Data Mining Technology for Efficient Network Security Management Ankit Naik [1], S.W. Ahmad [2] Student [1], Assistant Professor [2] Department of Computer Science and Engineering
Tensor Methods for Machine Learning, Computer Vision, and Computer Graphics
Tensor Methods for Machine Learning, Computer Vision, and Computer Graphics Part I: Factorizations and Statistical Modeling/Inference Amnon Shashua School of Computer Science & Eng. The Hebrew University
Machine Learning using MapReduce
Machine Learning using MapReduce What is Machine Learning Machine learning is a subfield of artificial intelligence concerned with techniques that allow computers to improve their outputs based on previous
Machine Learning in Computer Vision A Tutorial. Ajay Joshi, Anoop Cherian and Ravishankar Shivalingam Dept. of Computer Science, UMN
Machine Learning in Computer Vision A Tutorial Ajay Joshi, Anoop Cherian and Ravishankar Shivalingam Dept. of Computer Science, UMN Outline Introduction Supervised Learning Unsupervised Learning Semi-Supervised
Tracking in flussi video 3D. Ing. Samuele Salti
Seminari XXIII ciclo Tracking in flussi video 3D Ing. Tutors: Prof. Tullio Salmon Cinotti Prof. Luigi Di Stefano The Tracking problem Detection Object model, Track initiation, Track termination, Tracking
A Partially Supervised Metric Multidimensional Scaling Algorithm for Textual Data Visualization
A Partially Supervised Metric Multidimensional Scaling Algorithm for Textual Data Visualization Ángela Blanco Universidad Pontificia de Salamanca [email protected] Spain Manuel Martín-Merino Universidad
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
A Computational Framework for Exploratory Data Analysis
A Computational Framework for Exploratory Data Analysis Axel Wismüller Depts. of Radiology and Biomedical Engineering, University of Rochester, New York 601 Elmwood Avenue, Rochester, NY 14642-8648, U.S.A.
The use of computer vision technologies to augment human monitoring of secure computing facilities
The use of computer vision technologies to augment human monitoring of secure computing facilities Marius Potgieter School of Information and Communication Technology Nelson Mandela Metropolitan University
Search Taxonomy. Web Search. Search Engine Optimization. Information Retrieval
Information Retrieval INFO 4300 / CS 4300! Retrieval models Older models» Boolean retrieval» Vector Space model Probabilistic Models» BM25» Language models Web search» Learning to Rank Search Taxonomy!
Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall
Automatic Photo Quality Assessment Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Estimating i the photorealism of images: Distinguishing i i paintings from photographs h Florin
DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS
DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS 1 AND ALGORITHMS Chiara Renso KDD-LAB ISTI- CNR, Pisa, Italy WHAT IS CLUSTER ANALYSIS? Finding groups of objects such that the objects in a group will be similar
Predict Influencers in the Social Network
Predict Influencers in the Social Network Ruishan Liu, Yang Zhao and Liuyu Zhou Email: rliu2, yzhao2, [email protected] Department of Electrical Engineering, Stanford University Abstract Given two persons
MACHINE LEARNING IN HIGH ENERGY PHYSICS
MACHINE LEARNING IN HIGH ENERGY PHYSICS LECTURE #1 Alex Rogozhnikov, 2015 INTRO NOTES 4 days two lectures, two practice seminars every day this is introductory track to machine learning kaggle competition!
Time Series Prediction by Chaotic Modeling of Nonlinear Dynamical Systems
Time Series Prediction by Chaotic Modeling of Nonlinear Dynamical Systems Arslan Basharat + Kitware Inc. Clifton Park, NY, USA [email protected] Mubarak Shah + + University of Central Florida
203.4770: Introduction to Machine Learning Dr. Rita Osadchy
203.4770: Introduction to Machine Learning Dr. Rita Osadchy 1 Outline 1. About the Course 2. What is Machine Learning? 3. Types of problems and Situations 4. ML Example 2 About the course Course Homepage:
Distance Metric Learning in Data Mining (Part I) Fei Wang and Jimeng Sun IBM TJ Watson Research Center
Distance Metric Learning in Data Mining (Part I) Fei Wang and Jimeng Sun IBM TJ Watson Research Center 1 Outline Part I - Applications Motivation and Introduction Patient similarity application Part II
Visual Categorization with Bags of Keypoints
Visual Categorization with Bags of Keypoints Gabriella Csurka, Christopher R. Dance, Lixin Fan, Jutta Willamowski, Cédric Bray Xerox Research Centre Europe 6, chemin de Maupertuis 38240 Meylan, France
An Introduction to Data Mining. Big Data World. Related Fields and Disciplines. What is Data Mining? 2/12/2015
An Introduction to Data Mining for Wind Power Management Spring 2015 Big Data World Every minute: Google receives over 4 million search queries Facebook users share almost 2.5 million pieces of content
1816 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 15, NO. 7, JULY 2006. Principal Components Null Space Analysis for Image and Video Classification
1816 IEEE TRANSACTIONS ON IMAGE PROCESSING, VOL. 15, NO. 7, JULY 2006 Principal Components Null Space Analysis for Image and Video Classification Namrata Vaswani, Member, IEEE, and Rama Chellappa, Fellow,
Data, Measurements, Features
Data, Measurements, Features Middle East Technical University Dep. of Computer Engineering 2009 compiled by V. Atalay What do you think of when someone says Data? We might abstract the idea that data are
TouchPaper - An Augmented Reality Application with Cloud-Based Image Recognition Service
TouchPaper - An Augmented Reality Application with Cloud-Based Image Recognition Service Feng Tang, Daniel R. Tretter, Qian Lin HP Laboratories HPL-2012-131R1 Keyword(s): image recognition; cloud service;
Classifying Manipulation Primitives from Visual Data
Classifying Manipulation Primitives from Visual Data Sandy Huang and Dylan Hadfield-Menell Abstract One approach to learning from demonstrations in robotics is to make use of a classifier to predict if
Content-Based Recommendation
Content-Based Recommendation Content-based? Item descriptions to identify items that are of particular interest to the user Example Example Comparing with Noncontent based Items User-based CF Searches
Social Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
Supporting Online Material for
www.sciencemag.org/cgi/content/full/313/5786/504/dc1 Supporting Online Material for Reducing the Dimensionality of Data with Neural Networks G. E. Hinton* and R. R. Salakhutdinov *To whom correspondence
A Survey on Pre-processing and Post-processing Techniques in Data Mining
, pp. 99-128 http://dx.doi.org/10.14257/ijdta.2014.7.4.09 A Survey on Pre-processing and Post-processing Techniques in Data Mining Divya Tomar and Sonali Agarwal Indian Institute of Information Technology,
Visualization of Large Font Databases
Visualization of Large Font Databases Martin Solli and Reiner Lenz Linköping University, Sweden ITN, Campus Norrköping, Linköping University, 60174 Norrköping, Sweden [email protected], [email protected]
Clustering. Data Mining. Abraham Otero. Data Mining. Agenda
Clustering 1/46 Agenda Introduction Distance K-nearest neighbors Hierarchical clustering Quick reference 2/46 1 Introduction It seems logical that in a new situation we should act in a similar way as in
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Open issues and research trends in Content-based Image Retrieval
Open issues and research trends in Content-based Image Retrieval Raimondo Schettini DISCo Universita di Milano Bicocca [email protected] www.disco.unimib.it/schettini/ IEEE Signal Processing Society
RANDOM PROJECTIONS FOR SEARCH AND MACHINE LEARNING
= + RANDOM PROJECTIONS FOR SEARCH AND MACHINE LEARNING Stefan Savev Berlin Buzzwords June 2015 KEYWORD-BASED SEARCH Document Data 300 unique words per document 300 000 words in vocabulary Data sparsity:
Categorical Data Visualization and Clustering Using Subjective Factors
Categorical Data Visualization and Clustering Using Subjective Factors Chia-Hui Chang and Zhi-Kai Ding Department of Computer Science and Information Engineering, National Central University, Chung-Li,
Incremental PCA: An Alternative Approach for Novelty Detection
Incremental PCA: An Alternative Approach for Detection Hugo Vieira Neto and Ulrich Nehmzow Department of Computer Science University of Essex Wivenhoe Park Colchester CO4 3SQ {hvieir, udfn}@essex.ac.uk
Practical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006
Practical Tour of Visual tracking David Fleet and Allan Jepson January, 2006 Designing a Visual Tracker: What is the state? pose and motion (position, velocity, acceleration, ) shape (size, deformation,
FACE RECOGNITION BASED ATTENDANCE MARKING SYSTEM
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 2, February 2014,
Image Segmentation and Registration
Image Segmentation and Registration Dr. Christine Tanner ([email protected]) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation
Behavior Analysis in Crowded Environments. XiaogangWang Department of Electronic Engineering The Chinese University of Hong Kong June 25, 2011
Behavior Analysis in Crowded Environments XiaogangWang Department of Electronic Engineering The Chinese University of Hong Kong June 25, 2011 Behavior Analysis in Sparse Scenes Zelnik-Manor & Irani CVPR
Manifold Learning Examples PCA, LLE and ISOMAP
Manifold Learning Examples PCA, LLE and ISOMAP Dan Ventura October 14, 28 Abstract We try to give a helpful concrete example that demonstrates how to use PCA, LLE and Isomap, attempts to provide some intuition
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Dimension Reduction. Wei-Ta Chu 2014/10/22. Multimedia Content Analysis, CSIE, CCU
1 Dimension Reduction Wei-Ta Chu 2014/10/22 2 1.1 Principal Component Analysis (PCA) Widely used in dimensionality reduction, lossy data compression, feature extraction, and data visualization Also known
Classification of Bad Accounts in Credit Card Industry
Classification of Bad Accounts in Credit Card Industry Chengwei Yuan December 12, 2014 Introduction Risk management is critical for a credit card company to survive in such competing industry. In addition
Surgical Tools Recognition and Pupil Segmentation for Cataract Surgical Process Modeling
Surgical Tools Recognition and Pupil Segmentation for Cataract Surgical Process Modeling David Bouget, Florent Lalys, Pierre Jannin To cite this version: David Bouget, Florent Lalys, Pierre Jannin. Surgical
Waffles: A Machine Learning Toolkit
Journal of Machine Learning Research 12 (2011) 2383-2387 Submitted 6/10; Revised 3/11; Published 7/11 Waffles: A Machine Learning Toolkit Mike Gashler Department of Computer Science Brigham Young University
Support Vector Machines with Clustering for Training with Very Large Datasets
Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France [email protected] Massimiliano
Towards better accuracy for Spam predictions
Towards better accuracy for Spam predictions Chengyan Zhao Department of Computer Science University of Toronto Toronto, Ontario, Canada M5S 2E4 [email protected] Abstract Spam identification is crucial
MVA ENS Cachan. Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos [email protected]
Machine Learning for Computer Vision 1 MVA ENS Cachan Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos [email protected] Department of Applied Mathematics Ecole Centrale Paris Galen
FAST APPROXIMATE NEAREST NEIGHBORS WITH AUTOMATIC ALGORITHM CONFIGURATION
FAST APPROXIMATE NEAREST NEIGHBORS WITH AUTOMATIC ALGORITHM CONFIGURATION Marius Muja, David G. Lowe Computer Science Department, University of British Columbia, Vancouver, B.C., Canada [email protected],
Open-Set Face Recognition-based Visitor Interface System
Open-Set Face Recognition-based Visitor Interface System Hazım K. Ekenel, Lorant Szasz-Toth, and Rainer Stiefelhagen Computer Science Department, Universität Karlsruhe (TH) Am Fasanengarten 5, Karlsruhe
SZTAKI @ ImageCLEF 2011
SZTAKI @ ImageCLEF 2011 Bálint Daróczy Róbert Pethes András A. Benczúr Data Mining and Web search Research Group, Informatics Laboratory Computer and Automation Research Institute of the Hungarian Academy
Geometric-Guided Label Propagation for Moving Object Detection
MITSUBISHI ELECTRIC RESEARCH LABORATORIES http://www.merl.com Geometric-Guided Label Propagation for Moving Object Detection Kao, J.-Y.; Tian, D.; Mansour, H.; Ortega, A.; Vetro, A. TR2016-005 March 2016
See All by Looking at A Few: Sparse Modeling for Finding Representative Objects
See All by Looking at A Few: Sparse Modeling for Finding Representative Objects Ehsan Elhamifar Johns Hopkins University Guillermo Sapiro University of Minnesota René Vidal Johns Hopkins University Abstract
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Macario O. Cordel II and Arnulfo P. Azcarraga College of Computer Studies *Corresponding Author: [email protected]
Pixels Description of scene contents. Rob Fergus (NYU) Antonio Torralba (MIT) Yair Weiss (Hebrew U.) William T. Freeman (MIT) Banksy, 2006
Object Recognition Large Image Databases and Small Codes for Object Recognition Pixels Description of scene contents Rob Fergus (NYU) Antonio Torralba (MIT) Yair Weiss (Hebrew U.) William T. Freeman (MIT)
COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS
COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS B.K. Mohan and S. N. Ladha Centre for Studies in Resources Engineering IIT
Image Classification for Dogs and Cats
Image Classification for Dogs and Cats Bang Liu, Yan Liu Department of Electrical and Computer Engineering {bang3,yan10}@ualberta.ca Kai Zhou Department of Computing Science [email protected] Abstract
Visualization of General Defined Space Data
International Journal of Computer Graphics & Animation (IJCGA) Vol.3, No.4, October 013 Visualization of General Defined Space Data John R Rankin La Trobe University, Australia Abstract A new algorithm
Supervised Feature Selection & Unsupervised Dimensionality Reduction
Supervised Feature Selection & Unsupervised Dimensionality Reduction Feature Subset Selection Supervised: class labels are given Select a subset of the problem features Why? Redundant features much or
Classification algorithm in Data mining: An Overview
Classification algorithm in Data mining: An Overview S.Neelamegam #1, Dr.E.Ramaraj *2 #1 M.phil Scholar, Department of Computer Science and Engineering, Alagappa University, Karaikudi. *2 Professor, Department
Environmental Remote Sensing GEOG 2021
Environmental Remote Sensing GEOG 2021 Lecture 4 Image classification 2 Purpose categorising data data abstraction / simplification data interpretation mapping for land cover mapping use land cover class
Analyzing The Role Of Dimension Arrangement For Data Visualization in Radviz
Analyzing The Role Of Dimension Arrangement For Data Visualization in Radviz Luigi Di Caro 1, Vanessa Frias-Martinez 2, and Enrique Frias-Martinez 2 1 Department of Computer Science, Universita di Torino,
Taking Inverse Graphics Seriously
CSC2535: 2013 Advanced Machine Learning Taking Inverse Graphics Seriously Geoffrey Hinton Department of Computer Science University of Toronto The representation used by the neural nets that work best
Big Data Analytics CSCI 4030
High dim. data Graph data Infinite data Machine learning Apps Locality sensitive hashing PageRank, SimRank Filtering data streams SVM Recommen der systems Clustering Community Detection Web advertising
Combining SVM classifiers for email anti-spam filtering
Combining SVM classifiers for email anti-spam filtering Ángela Blanco Manuel Martín-Merino Abstract Spam, also known as Unsolicited Commercial Email (UCE) is becoming a nightmare for Internet users and
The Visual Internet of Things System Based on Depth Camera
The Visual Internet of Things System Based on Depth Camera Xucong Zhang 1, Xiaoyun Wang and Yingmin Jia Abstract The Visual Internet of Things is an important part of information technology. It is proposed
