Generating Hyperdimensional Distributed Representations from Continuous-
|
|
|
- Sheila Alison Ryan
- 10 years ago
- Views:
Transcription
1 Generating Hyperdimensional Distributed Representations from Continuous- Valued Multivariate Sensory Input Okko Räsänen Department of Signal Processing and Acoustics. Aalto University PO Box 76 AALTO, Finland Abstract Hyperdimensional computing (HDC refers to the representation and manipulation of data in a very high dimensional space using random vectors. Due to the high dimensionality, vectors of the space can code large amounts of information in a distributed manner, are robust to variation, and are easily distinguished from random noise. More importantly, HDC can be used to represent compositional and hierarchical relationships and recursive operations between entities using fixed-size representations, making it intriguing from a cognitive modeling point of view. However, the majority of the existing work in this area has focused on modeling discrete categorical data. This paper presents a new method for mapping continuous-valued multivariate data into hypervectors, enabling construction of compositional representations from non-categorical data. The mapping is studied in a word classification task, showing how rich distributed representations of spoken words can be encoded using HDC-based representations. Keywords: hyperdimensional computing; distributed representations; speech recognition; memory Introduction Hyperdimensional computing (HDC was first introduced by Kanerva (988 in the context of his neurally inspired memory model called sparse distributed memory (SDM. HDC is based on the idea that the distances between concepts in our minds correspond to distances between points in a very high-dimensional space (Kanerva, 29. Since its introduction, HDC has been used successfully in modeling of analogical processing (Plate, 995; see also Eliasmith & Thagard, 2, latent semantic analysis (Kanerva et al., 2, multimodal data fusion and prediction (Räsänen & Kakouros, 24, robotics (Jockel, 2, and, e.g., cognitive architectures (Rachkovskij et al., 23; see also Levy & Gayler, 28; Kelly & West, 22 as it successfully bridges the gap between symbolic processing and connectionist systems. In typical systems using HDC, discrete entities w i (e.g., symbols, states or words are represented with randomly generated binary, ternary, or continuous-valued vectors y i of huge dimensionality h, typically counted in thousands (e.g., Kanerva, 988; Kanerva et al., 2. These vectors can have only a small number of non-zero elements (as in SDM, or they can be fully dense. In all cases, the large dimensionality of such vectors leads to a number of interesting properties (see Gallant & Okaywe, 23, for a recent overview. Firstly, the representations are highly robust against distortions, noise, and degradation due to the distribution of information across numerous dimensions. Secondly, the distribution of the mutual distances between all possible random vectors is tightly packed around the mean of the distances. In the case of random hypervectors with zero mean, the pair-wise linear correlation ρ(y a,y b [-, ] between almost any two randomly drawn vectors y a and y b is very close to zero (Kanerva, 29. This quasi-orthogonality of random vectors leads to the practical property that a set of unrelated items can be represented as the sum of the hypervectors corresponding to the items in the set. For example, a set {w, w 2, w 3 } can be coded as y set = y +y 2 +y 3, and this process is usually referred to as chunking. The obtained representation y set is much more similar to its components than any unrelated vectors in the hyperspace, and therefore the individual items can still be recovered from the holistic representation if the codes of all possible items are known (see Gallant & Okaywe, 23, for a capacity analysis. In addition, HDC can overcome the superposition catastrophe of distributed representations by using invertible vector operations such as circular convolution to bind vectors together (Plate, 995. For instance, correct attribute encoding of a sentence black cats and red balls could be represented with y = y black y cats + y red y balls + y and if each unique word is assigned with a random vector and where denotes the binding operation. Importantly, the dimension of the representations always stays fixed during the chunking and binding operations, ensuring that distance metrics between representations of different granularity and combinatorial complexity are always defined. However, a major challenge in applying HDC to many real world problems has been that the world, as sensed by a number of senses (or sensors, does not give rise to inherently categorical (discrete representations before some learning takes place. The idea of using random vectors for different inputs is only applicable after the data has been clustered or quantized into a finite number of representative states or receptive fields. Given the theoretically interesting properties of HDC, it would be useful to be able to represent non-categorical multivariate inputs such as speech in a HDC-compatible form without imposing hard quantization on the input features before further processing. In order to address this issue, the present paper describes a method for transforming continuous multivariate data into quasi-orthogonal random vectors. The transformation maintains local distance metrics of the original feature 943
2 space, allowing generalization across similar tokens, while simultaneously mapping more distant inputs into nearly orthogonal random vectors that is a requirement for the efficient use of chunking and binding operations. In comparison to the previously suggested scatter code (Stanford & Smith, 994, the present method is not limited to binary vectors, enabling higher representational capacity in vector spaces of the same dimension. The proposed method is evaluated in a spoken word classification task using a simple prototype memory for acoustic modeling. S-WARP mapping for multivariate data The core of the mapping problem is that many types of data such as spectral features of speech do not come in discrete and mutually exclusive categories w i w j (i j that can be assigned with unique random vectors but as multivariate observations x t with varying degrees of similarity (correlation between the vectors. The correlation is a problem because it significantly affects the coding capacity of the hyperspace as the entire idea of HDC is to operate on quasi-orthogonal representations. However, in order to generalize between different tokens of the same category, the correlation between the original features should be also reflected in the derived hyperdimensional representations, and therefore arbitrarily small differences in the input cannot lead to orthogonal codes in the hyperspace. Given this, the minimal set of desired properties in the mapping y = f(x from a low-dimensional space F to a highdimensional space H can be listed as follows: Local similarities between input vectors must be approximately maintained, enabling generalization towards new input tokens. 2 Distant inputs should be coded with quasi-orthogonal vectors, maximizing coding capacity of the hyperspace. 3 A continuous distance metric between original vectors should be also continuous and smooth in the hyperspace. 4 The local/distant trade-off in the requirements in and 2 should be controllable. The desired properties are illustrated in Fig.. In order to approach a solution to the mapping problem, let x denote a feature vector of dimension d = x with feature values x i, i = [, 2,, d] from the feature space F. In addition, let M denote a mapping matrix of size h x d where h is the dimension of the hyperspace H (h >> d. In the case of a trivial random mapping from F to H, one can initialize M as a randomly generated binary matrix (all values randomly set to + or - and then linearly expand the original feature vector x as: y = Mx ( This type of random mapping approximately preserves the relative distances in F (the Johnson-Lindenstrauss Lemma. However, this only makes use of a subspace S H of the entire hyperspace due to the fixed mapping from each x i to a set of y j, j [, h]. In other words, M acts as a single basis in H, and the distance metrics are linearly preserved. In Fig.. An example of desired hyperspace mapping properties in terms of distance metrics. The x-axis shows the correlation ρ(x a [-, ] between two data points in the low-dimensional input space and the y-axis shows the crosscorrelation ρ(y a,y b between the corresponding hypervectors. Ideally, local similarity (high ρ is carried over to the hyperspace while the hypervectors of distant inputs are independent of each other. Preservation of anticorrelations (ρ - can also be beneficial for some tasks. order to achieve orthogonalization between distant inputs, different mapping matrices M a and M b should be used for feature vectors x a and x b that are far apart in the original feature space. Simultaneously, the same mapping matrix M cd should be used for two vectors x c and x d that are similar to each other in F in order to maintain similarity in H also. The problem then is the selection of the best possible mapping matrix M i for each input vector x i. In addition, the transition between matrices M i and M j should be smooth so that the mapping does not introduce points of discontinuity in the distance metrics between inputs x i and x j. We propose that a deterministic mapping with efficient use of the entire hyperspace can be achieved as a linear combination of individual mappings. More specifically, let us define M i (i = [, 2,, v] as a set of v random mapping matrices. Then the mapping x à y can be written as v y = λ i M i x (2 i= Since each individual mapping with a random matrix M i approximately preserves the distance metrics in a linear manner, the weights λ i can be used to control the rate of change from one basis to another (Fig. 2. From now on, we will refer to the formulation in Eq. (2 as Weighted Accumulation of Random Projections (WARP. In the absence of any external knowledge of the input, the weights λ i of each single mapping are determined by the input vector itself: λ i = f (x (3 In other words, the hypervector is a result of v random mappings i = [, 2,, v] into the hyperspace H with each individual mapping i weighted by a value that is derived from the vector itself that is being mapped. We will refer to this self-regulated mapping as S-WARP. 944
3 = = 3 = 5 = 7 (y a,y b d= (y a,y b d=3 (y a,y b d=5 Fig. 2. A schematic view of the mapping in Eq. (2 utilizing a linear combination of multiple individual mappings. Each individual mapping matrix M i acts as a pointer to a subspace of H, and pairs of data points ending up in different sub-spaces become quasi-orthogonal with a high probability. Smooth transitions between sub-spaces are ensured by a weighting function that is a continuous function of the input itself. Possibly the simplest way to implement S-WARP would be to use the elements of the input vector x directly as the weighting coefficients but then the mapping would be indifferent to the sign of the input vector, i.e., y =f(x = f(- x. This problem can be avoided by using the absolute value of the coefficients instead: α λ i = ( x i / x α j (4 j The additional parameter α in Eq. (4 controls the amount of orthogonalization by controlling the rate at which the hyperspace basis matrices M i change when values of x change. When α has a high value, two vectors have to be very close in the original space F in order to end up close in the hyperspace whereas more distant vectors tend towards quasi-orthogonal representations (cf. Fig. When the weights of Eq. (4 are used in the mapping described in Eq. (2, all previously listed requirements are satisfied. The mapping is also scale invariant with respect to the resulting hypervector direction, i.e., ρ(f(x a,f(x b = ρ(f(αx a,f(βx b, where f(x denotes the mapping operation and α and β are constants, while the magnitude of the resulting hypervector will be affected. This is not the case for the previously introduced scatter code (Stanford & Smith, 994 where the direction of the vector changes if the input vector is multiplied by a constant. However, the weighting scheme in Eq. (4 still has a shortcoming. Consider two vectors x a and x b with possibly different signs and scale but similar relative order of magnitudes within the set of largest elements. After applying Eq. (4, the weights λ i become similar for the two vectors, and they are mapped using a similar set of mapping matrices M. Since the non-linearity of the distances in the hyperspace is caused by the use of different weights for different vectors, the distance between the two different vectors x a and x b using the similar weights λ becomes linearized. With large values of α, the mapping becomes Fig. 3. Correlation ρ(x a of random vectors in the original space F (x axis and the corresponding correlation ρ(y a,y b in the hyperspace H (y axis as a function of α in Eq. (6 (columns and dimension d of the input vectors (rows. dependent only on the largest magnitude elements in the input and thus the probability of linearization increases. In practice, this phenomenon limits the maximum value of α that still leads to a consistent mapping with a sufficiently high probability. The risk of linearization is also dependent on the dimension d of the original vectors. The effects of α and d on the mapping of Eq. (2 and (4 are illustrated in Fig. 3, where correlations between pairs of randomly generated original low-dimensional vectors and the corresponding hypervectors are plotted. As can be observed, α successfully controls the non-linearity of the distances in the hyperspace, but the non-linearity breaks down for a large ratio of α/d. For increasing α, increasingly many vector pairs maintain linear or near-linear mutual distance across the mapping. As the largest useful non-linearity factor α of a single hyperspace mapping is determined by the dimension d of the input vectors, the problem can be easily solved by simply first expanding the original d-dimensional input data into a higher dimension h > d using linear random mapping in Eq. ( before applying S-WARP in Eq. (2. Another option is to recursively apply S-WARP mapping with a smaller value of α, in which case the non-linearity will gradually increase towards a desired level. The linear expansion approach is demonstrated in Fig. 4, where random x of original dimension d = are first mapped into a 3-dimensional space with a randomly generated fully dense expansion matrix E (all elements + or - according to y = Ex, and then the resulting y are mapped into hypervectors y according to Eq. (2 with weights according to Eq. (4. As can be observed, the linearization problem is now absent, confirming that the linear expansion is sufficient for avoiding the linearization artifacts occurring with small d and/or large α. In general, the method is successful at generating quasi-orthogonal representations for weakly-correlated inputs. Spoken word classification with HDC S-WARP was studied in word recognition from speech. Since the current goal was to study hypervectors capability 945
4 to code structural information of time-varying signals, the experiment was limited to the classification of a small vocabulary of words that had been segmented from continuous speech using the available word-level annotation of the data. (y a, y b.5.5 =.5.5 = = 9 Data The speech data consisted of 2397 utterances from the four main talkers of the CAREGIVER Y2 UK corpus (Altosaar et al., 2. The material consists of child directed speech with an overall vocabulary size of 79 unique words (silences excluded. Each signal corresponding to an utterance was segmented into individual words using the associated wordlevel annotation. Separate training and testing sets were created for each of the four talkers by choosing 8% of the first words as the training samples (N = 423 ±4.6 for each talker and the remaining 2% as the test samples (N = 266 ±. in the order of appearance in the corpus. A total of 79 unique words occurred in the training data of which 7 also occurred in the test set. All models were always trained on the full set of 79 words. Experimental setup The entire word classification architecture is based on learning a hypervector prototype m w for each word w in the training data, where the prototype is constructed incrementally from the short-term spectral features extracted from the acoustic realizations of the word (Fig. 5. The audio signal corresponding to a spoken word is first fed to a pre-processing block where standard Mel-frequency cepstral coefficient (MFCC features, including delta and delta-delta, are extracted using a 32-ms Hamming window with a step size of ms (a total of 39 coefficients including energy. Each MFCC vector x t is then used as an input to the hyperdimensional mapping processs (S-WARP or scatter code, yielding a hypervector y t for the corresponding time frame. The temporal structure of the words is encoded with the binding operation by computing pair-wise circular convolutions z t, k = y t y P t-k between all vectors within 25 ms from each other (k [, 2,, 25] (cf., Plate, 995. In the convolution, the preceding vector is always permuted with a fixed permutation (denoted with y P in order to encode temporal order information, since otherwise y t y t-k = y t-k y t, i.e., making the representation invariant with respect to the direction of time. Finally, all the obtained hypervectors y t and z t,k are additively combined to form a single hypervector y input for the current input, and the result is summed to the existing hypervector model m w for the word w in order to have an updated model m w. y input = z t,k + y t t,k t (7 m" w m w + y input As a result of processing the training data, a word model m w is the sum of all word w realizations, where each realization is the sum of all frame-based hypervectors and their pair-.5.5, x b.5.5, x b.5.5, x b Fig. 4. Examples of cross-correlation plots for a two-stage process where the low-dimensional input vectors are first expanded to a larger dimension with a linear random mapping and then used as an input to the non-linear mapping in Eq. (2. Results for three different values of non-linearity, namely α =, 3, and 9, are shown from left to right, respectively. Fig. 5. A schematic view of the word recognition system used in the current experiment (training stage. wise circular convolutions (note that the model is of same dimension as each individual feature frame or each individual realization of the word. During the training, the word identity w is always known due to labeling, and the word models m w for all W unique words are accumulated as row vectors of a memory matrix H of size W x h. During the recognition, the input segment is again coded into y input and the most likely word label w is obtained by computing the activation distribution p with p = H y input (8 where H denotes H with each row normalized into a unit vector. The hypothesis w i for the current input is determined by finding the p i (i [, W] with the largest value. The experiment was conducted using both binary scatter code and the continuous-valued S-WARP proposed in the current paper. The test was repeated for different values of the non-linearity parameter s of the scatter code, for different values of α in the present S-WARP formulation, and with and without the linear expansion layer before the non-linear mapping. Based on preliminary tests, the size of the linear expansion layer in S-WARP was always fixed to d E = 3 in order to ensure that no linearization occurs for the studied values of α. 946
5 recognition rate (% 95 9 GMM HMM (fullcov GMM HMM (diag scatter code recognition rate (% 95 9 S WARP with LinExp S WARP direct mapping recognition rate (% 95 9 S WARP with LinExp s x d E Fig. 6. Word classification accuracies (UAR % with talker specific models (mean and one standard deviation of the results across the four talkers. Left: Scatter code (red dashed line as a function of the non-linearity parameter s and the reference HMM systems with diagonal and full covariance matrices ( diag and fullcov, respectively. Center: Performance of the S- WARP with and without the linear expansion layer as a function of the α parameter and with d E = 3. Right: S-WARP performance as a function of the linear expansion layer size d E with fixed α =.5. Hyperspace dimensionality is always set to h = In the scatter code, the integer values of each dimension of the input space F are first sorted into numerical order and one of the integers is mapped into a randomly generated binary vector of dimension h. Then a code for a neighboring integer is generated by randomly choosing b locations of the first hypervector and flipping the corresponding bits. The new vector is then used as a starting point for the next integer, and the random flipping process is repeated until the entire range of the input space is covered. In this manner, the expected Hamming distance of two codes in the hyperspace is equal to h/2*(-(-2/h*(b*t/h, where h is the dimension of the hyperspace and t is the distance in the original space, i.e., the rate of orthogonalization is controlled by the proportion s = b/h of flipped bits per iteration (Smith & Stanford, 99. After the process is repeated for each input dimension separately, the resulting hypervectors are combined with the XOR operation (Stanford & Smith, 994 in order to obtain the final hypervector describing the entire multivariate input vector. Two other reference systems were also used. The basic reference was exactly the same setup as the system in Fig. 5 except that the hypervectors y were replaced with the original low-dimensional MFCC feature vectors x before the convolution and accumulation. This provides a test for the benefits of hyperdimensionality in comparison to operating in low-dimensional spaces. The second reference system was a standard Gaussian mixture -based continuous-density hidden-markov model (GMM-HMM, one HMM trained for each word. For benchmarking purposes, the number of states and Gaussians in the HMMs were optimized directly on the test data, leading to Q = 3 states for all words and M = 3 Gaussians per mixture. The Gaussians were initialized using the k-means algorithm, and parameters were estimated using the Baum-Welch algorithm with four iterations, as this was found to perform best on the test set. Classification performance was evaluated in terms of unweighted average recall (UAR computed across the words occurring at least once in the test data (the mean of word-specific classification accuracies. Results recognition rate (% S WARP LinExp hyperspace dimensionality h Fig. 7. The effect of hyperspace dimensionality h on the classification accuracy for linear (Eq. and S-WARP (Eq. 2 mappings. Fig. 6 shows the average results for the speaker-dependent models across all four speakers. The S-WARP template system performs at a level comparable with the HMM system using full covariance matrices, with S-WARP achieving an UAR of 97.2% (α =, d E = 3, h = 4 while the HMM reaches on average an UAR of 97.%. Without the linear expansion layer, S-WARP achieves an UAR of 96.6%. The scatter code achieves best performance of 92.3% correct recognitions at s =.75 (h = 4 The word recognition accuracy using the original MFCCs is 67.9% with convolutional encoding of temporal dependencies. If only the sum of the individual MFCC vectors x t is used as a model for each word (i.e., no temporal convolution, the performance drops to 3.4%. This means that the S-WARP and scatter code -based HDC representations are able to maintain information about not only the average spectrum of a word, but also the evolution of the spectrum across the word duration and a typical variability of this trajectory across different realizations of the word. The latter aspects are lost in a low-dimensional average MFCC template. The results also reveal that the degree of distance metric non-linearity in the mapping has an effect on the overall 947
6 results. This is revealed by the scatter code results (Fig. 6, left and in the effects of α and d E that control the orthogonalization in S-WARP (Fig. 6 middle and right. Note that the use of Eq. (2 with α = and the original data dimensionality of d = 39 (the MFCC coefficients already leads to a relatively large degree of non-linearity, and therefore the performance of S-WARP without the linear expansion layer is already optimal at α =. However, if α is fixed and d E is gradually decreased from the default value of, the performance drops significantly (Fig. 6, right. Finally, Fig. 7 shows the effect of hyperspace dimensionality on the results when using standard nonorthogonalizing random mapping and the proposed S- WARP mapping. As can be observed, the S-WARP outperforms the linear mapping with a clear margin but both benefit from an increasing hypervector dimensionality. This is expected as the chunking and binding operations assume that the data lies in a sufficiently high dimensional space. Conclusions The current work describes a new method for converting multivariate inputs into continuous-valued hyperdimensional random vectors. The method attempts to solve the conflicting requirements of similarity preservation and decorrelation by performing recursive application of selfmodulated random mappings. This leads to orthogonalization of the input data in a manner that still allows detection of the degree of similarity in the inputs. This is a highly relevant property for any cognitive system utilizing distributed representations, since no learning can occur without a consistent mapping from sensory input to internal representations in the memory and without the ability to measure similarity between the representations. In contrast to standard (deep neural networks, the proposed approach does not involve learning and therefore the quality of the distributed representations is not dependent on the amount of training data available for training of the mapping network. On the other hand, S-WARP does not learn abstracted features from the data, but simply makes noncategorical data available for use in HDC based systems. The present work also shows that both S-WARP and scatter code based HDC representations can be used to encode the complex time-frequency structure of spoken words by utilizing the chunking and binding operations of HDC systems. However, a more systematical approach to encoding temporally evolving multivariate inputs should be investigated in future work. In addition, it should be noted that despite its mathematical simplicity, the large number of vector operations in S-WARP makes its computational complexity non-trivial for large-scale experiments. Acknowledgement This research was funded by the Academy of Finland and by Tekes D2I program. The author would like to thank Unto K. Laine, Sofoklis Kakouros and Jukka P. Saarinen for their useful comments on the manuscript. References Altosaar, T., ten Bosch, L., Aimetti, G., Koniaris, C., Demuynck, K., & van den Heuvel, H. (2. A Speech Corpus for Modeling Language Acquisition: CAREGIVER. Proc. LREC 2, Malta. Eliasmith, C. & Thagard, P. (2. Integrating Structure and Meaning: a distributed model of analogical mapping. Cognitive Science, 25, Gallant, S. I., & Okaywe, T. W. (23. Representing Objects, Relations, and Sequences. Neural Computation, 25, Jockel, S (2. Crossmodal Learning and Prediction of Autobiographical Episodic Experiences using a Sparse Distributed Memory. Ph.D. dissertation, Department of Informatics, University of Hamburg. Kanerva, P. (988. Sparse distributed memory. Cambridge, Mass.: Bradford/MIT Press. Kanerva, P., (29. Hyperdimensional Computing: An Introduction to Computing in Distributed Representation with High-Dimensional Random Vectors. Cognitive Computation,, Kanerva, P., Kristoferson, J., & Holst, A. (2. Random Indexing of Text Samples for Latent Semantic Analysis. Proc. 22nd Annual Conference of the Cognitive Science Society, Mahwah, New Jersey, pp. 36. Kelly, M. A., & West, R. L. (22. From Vectors to Symbols to Cognition: The Symbolic and Sub-Symbolic Aspects of Vector-Symbolic Cognitive Models. Proc. 34th Annual Conference of the Cognitive Science Society, Austin, TX, pp Levy, S. D., & Gayler, R. (28. Vector Symbolic Architectures: A New Building Material for Artificial General Intelligence. Proc. Conf. on Artificial General Intelligence 28, IOS Press, Amsterdam, pp Plate, T. (995. Holographic reduced representations. IEEE Trans. Neural Networks, 6, Plate, T. (2. Analogy retrieval and processing with distributed vector representations. Expert Systems: Int. J. Knowledge Eng. and Neural Networks, 7, Rachkovskij, D. A., Kussul, E. M., & Baidyk, T. N. (23. Building a world model with structure-sensitive sparse binary distributed representations. Biologically Inspired Cognitive Architectures, 3, Räsänen, O., & Kakouros, S. (24. Modeling Dependencies in Multiple Parallel Data Streams with Hyperdimensional Computing. IEEE Signal Processing Letters, Smith, D. J., & Stanford, P. H. (99. A Random Walk in Hamming Space. Proc. IEEE Intl. Conf. on Neural Networks, San Diego, California, pp Stanford, P. H., & Smith, D. J. (994. The Multidimensional Scatter Code: A Data Fusion Technique with Exponential Capacity. Proc. Int. Conf. on Artificial Neural Networks (ICANN 94, Sorrento, Italy, pp
Data, Measurements, Features
Data, Measurements, Features Middle East Technical University Dep. of Computer Engineering 2009 compiled by V. Atalay What do you think of when someone says Data? We might abstract the idea that data are
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Macario O. Cordel II and Arnulfo P. Azcarraga College of Computer Studies *Corresponding Author: [email protected]
AUTOMATIC PHONEME SEGMENTATION WITH RELAXED TEXTUAL CONSTRAINTS
AUTOMATIC PHONEME SEGMENTATION WITH RELAXED TEXTUAL CONSTRAINTS PIERRE LANCHANTIN, ANDREW C. MORRIS, XAVIER RODET, CHRISTOPHE VEAUX Very high quality text-to-speech synthesis can be achieved by unit selection
How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm
IJSTE - International Journal of Science Technology & Engineering Volume 1 Issue 10 April 2015 ISSN (online): 2349-784X Image Estimation Algorithm for Out of Focus and Blur Images to Retrieve the Barcode
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Hardware Implementation of Probabilistic State Machine for Word Recognition
IJECT Vo l. 4, Is s u e Sp l - 5, Ju l y - Se p t 2013 ISSN : 2230-7109 (Online) ISSN : 2230-9543 (Print) Hardware Implementation of Probabilistic State Machine for Word Recognition 1 Soorya Asokan, 2
Investigations on Error Minimizing Training Criteria for Discriminative Training in Automatic Speech Recognition
, Lisbon Investigations on Error Minimizing Training Criteria for Discriminative Training in Automatic Speech Recognition Wolfgang Macherey Lars Haferkamp Ralf Schlüter Hermann Ney Human Language Technology
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin
Establishing the Uniqueness of the Human Voice for Security Applications
Proceedings of Student/Faculty Research Day, CSIS, Pace University, May 7th, 2004 Establishing the Uniqueness of the Human Voice for Security Applications Naresh P. Trilok, Sung-Hyuk Cha, and Charles C.
Component Ordering in Independent Component Analysis Based on Data Power
Component Ordering in Independent Component Analysis Based on Data Power Anne Hendrikse Raymond Veldhuis University of Twente University of Twente Fac. EEMCS, Signals and Systems Group Fac. EEMCS, Signals
MUSICAL INSTRUMENT FAMILY CLASSIFICATION
MUSICAL INSTRUMENT FAMILY CLASSIFICATION Ricardo A. Garcia Media Lab, Massachusetts Institute of Technology 0 Ames Street Room E5-40, Cambridge, MA 039 USA PH: 67-53-0 FAX: 67-58-664 e-mail: rago @ media.
Advanced Signal Processing and Digital Noise Reduction
Advanced Signal Processing and Digital Noise Reduction Saeed V. Vaseghi Queen's University of Belfast UK WILEY HTEUBNER A Partnership between John Wiley & Sons and B. G. Teubner Publishers Chichester New
Developing an Isolated Word Recognition System in MATLAB
MATLAB Digest Developing an Isolated Word Recognition System in MATLAB By Daryl Ning Speech-recognition technology is embedded in voice-activated routing systems at customer call centres, voice dialling
Lecture 9: Introduction to Pattern Analysis
Lecture 9: Introduction to Pattern Analysis g Features, patterns and classifiers g Components of a PR system g An example g Probability definitions g Bayes Theorem g Gaussian densities Features, patterns
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
Nonlinear Iterative Partial Least Squares Method
Numerical Methods for Determining Principal Component Analysis Abstract Factors Béchu, S., Richard-Plouet, M., Fernandez, V., Walton, J., and Fairley, N. (2016) Developments in numerical treatments for
THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS
THREE DIMENSIONAL REPRESENTATION OF AMINO ACID CHARAC- TERISTICS O.U. Sezerman 1, R. Islamaj 2, E. Alpaydin 2 1 Laborotory of Computational Biology, Sabancı University, Istanbul, Turkey. 2 Computer Engineering
Emotion Detection from Speech
Emotion Detection from Speech 1. Introduction Although emotion detection from speech is a relatively new field of research, it has many potential applications. In human-computer or human-human interaction
Visualization of large data sets using MDS combined with LVQ.
Visualization of large data sets using MDS combined with LVQ. Antoine Naud and Włodzisław Duch Department of Informatics, Nicholas Copernicus University, Grudziądzka 5, 87-100 Toruń, Poland. www.phys.uni.torun.pl/kmk
How To Use Neural Networks In Data Mining
International Journal of Electronics and Computer Science Engineering 1449 Available Online at www.ijecse.org ISSN- 2277-1956 Neural Networks in Data Mining Priyanka Gaur Department of Information and
ViSOM A Novel Method for Multivariate Data Projection and Structure Visualization
IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 13, NO. 1, JANUARY 2002 237 ViSOM A Novel Method for Multivariate Data Projection and Structure Visualization Hujun Yin Abstract When used for visualization of
A Robust Method for Solving Transcendental Equations
www.ijcsi.org 413 A Robust Method for Solving Transcendental Equations Md. Golam Moazzam, Amita Chakraborty and Md. Al-Amin Bhuiyan Department of Computer Science and Engineering, Jahangirnagar University,
Artificial Neural Network for Speech Recognition
Artificial Neural Network for Speech Recognition Austin Marshall March 3, 2005 2nd Annual Student Research Showcase Overview Presenting an Artificial Neural Network to recognize and classify speech Spoken
USING SPECTRAL RADIUS RATIO FOR NODE DEGREE TO ANALYZE THE EVOLUTION OF SCALE- FREE NETWORKS AND SMALL-WORLD NETWORKS
USING SPECTRAL RADIUS RATIO FOR NODE DEGREE TO ANALYZE THE EVOLUTION OF SCALE- FREE NETWORKS AND SMALL-WORLD NETWORKS Natarajan Meghanathan Jackson State University, 1400 Lynch St, Jackson, MS, USA [email protected]
General Framework for an Iterative Solution of Ax b. Jacobi s Method
2.6 Iterative Solutions of Linear Systems 143 2.6 Iterative Solutions of Linear Systems Consistent linear systems in real life are solved in one of two ways: by direct calculation (using a matrix factorization,
SPEAKER IDENTIFICATION FROM YOUTUBE OBTAINED DATA
SPEAKER IDENTIFICATION FROM YOUTUBE OBTAINED DATA Nitesh Kumar Chaudhary 1 and Shraddha Srivastav 2 1 Department of Electronics & Communication Engineering, LNMIIT, Jaipur, India 2 Bharti School Of Telecommunication,
Statistical Models in Data Mining
Statistical Models in Data Mining Sargur N. Srihari University at Buffalo The State University of New York Department of Computer Science and Engineering Department of Biostatistics 1 Srihari Flood of
Data Mining Project Report. Document Clustering. Meryem Uzun-Per
Data Mining Project Report Document Clustering Meryem Uzun-Per 504112506 Table of Content Table of Content... 2 1. Project Definition... 3 2. Literature Survey... 3 3. Methods... 4 3.1. K-means algorithm...
Factor Analysis. Principal components factor analysis. Use of extracted factors in multivariate dependency models
Factor Analysis Principal components factor analysis Use of extracted factors in multivariate dependency models 2 KEY CONCEPTS ***** Factor Analysis Interdependency technique Assumptions of factor analysis
6.2.8 Neural networks for data mining
6.2.8 Neural networks for data mining Walter Kosters 1 In many application areas neural networks are known to be valuable tools. This also holds for data mining. In this chapter we discuss the use of neural
Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability
Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller
An Introduction to Random Indexing
MAGNUS SAHLGREN SICS, Swedish Institute of Computer Science Box 1263, SE-164 29 Kista, Sweden [email protected] Introduction Word space models enjoy considerable attention in current research on semantic indexing.
DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS
DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS 1 AND ALGORITHMS Chiara Renso KDD-LAB ISTI- CNR, Pisa, Italy WHAT IS CLUSTER ANALYSIS? Finding groups of objects such that the objects in a group will be similar
D-optimal plans in observational studies
D-optimal plans in observational studies Constanze Pumplün Stefan Rüping Katharina Morik Claus Weihs October 11, 2005 Abstract This paper investigates the use of Design of Experiments in observational
Linear Codes. Chapter 3. 3.1 Basics
Chapter 3 Linear Codes In order to define codes that we can encode and decode efficiently, we add more structure to the codespace. We shall be mainly interested in linear codes. A linear code of length
Ericsson T18s Voice Dialing Simulator
Ericsson T18s Voice Dialing Simulator Mauricio Aracena Kovacevic, Anna Dehlbom, Jakob Ekeberg, Guillaume Gariazzo, Eric Lästh and Vanessa Troncoso Dept. of Signals Sensors and Systems Royal Institute of
An Overview of Knowledge Discovery Database and Data mining Techniques
An Overview of Knowledge Discovery Database and Data mining Techniques Priyadharsini.C 1, Dr. Antony Selvadoss Thanamani 2 M.Phil, Department of Computer Science, NGM College, Pollachi, Coimbatore, Tamilnadu,
Support Vector Machines with Clustering for Training with Very Large Datasets
Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France [email protected] Massimiliano
Solutions to Exam in Speech Signal Processing EN2300
Solutions to Exam in Speech Signal Processing EN23 Date: Thursday, Dec 2, 8: 3: Place: Allowed: Grades: Language: Solutions: Q34, Q36 Beta Math Handbook (or corresponding), calculator with empty memory.
Stabilization by Conceptual Duplication in Adaptive Resonance Theory
Stabilization by Conceptual Duplication in Adaptive Resonance Theory Louis Massey Royal Military College of Canada Department of Mathematics and Computer Science PO Box 17000 Station Forces Kingston, Ontario,
Spatial Statistics Chapter 3 Basics of areal data and areal data modeling
Spatial Statistics Chapter 3 Basics of areal data and areal data modeling Recall areal data also known as lattice data are data Y (s), s D where D is a discrete index set. This usually corresponds to data
Secure-Access System via Fixed and Mobile Telephone Networks using Voice Biometrics
Secure-Access System via Fixed and Mobile Telephone Networks using Voice Biometrics Anastasis Kounoudes 1, Anixi Antonakoudi 1, Vasilis Kekatos 2 1 The Philips College, Computing and Information Systems
Neural Network Add-in
Neural Network Add-in Version 1.5 Software User s Guide Contents Overview... 2 Getting Started... 2 Working with Datasets... 2 Open a Dataset... 3 Save a Dataset... 3 Data Pre-processing... 3 Lagging...
System Identification for Acoustic Comms.:
System Identification for Acoustic Comms.: New Insights and Approaches for Tracking Sparse and Rapidly Fluctuating Channels Weichang Li and James Preisig Woods Hole Oceanographic Institution The demodulation
Predict Influencers in the Social Network
Predict Influencers in the Social Network Ruishan Liu, Yang Zhao and Liuyu Zhou Email: rliu2, yzhao2, [email protected] Department of Electrical Engineering, Stanford University Abstract Given two persons
2. Simple Linear Regression
Research methods - II 3 2. Simple Linear Regression Simple linear regression is a technique in parametric statistics that is commonly used for analyzing mean response of a variable Y which changes according
Subspace Analysis and Optimization for AAM Based Face Alignment
Subspace Analysis and Optimization for AAM Based Face Alignment Ming Zhao Chun Chen College of Computer Science Zhejiang University Hangzhou, 310027, P.R.China [email protected] Stan Z. Li Microsoft
Adaptive Demand-Forecasting Approach based on Principal Components Time-series an application of data-mining technique to detection of market movement
Adaptive Demand-Forecasting Approach based on Principal Components Time-series an application of data-mining technique to detection of market movement Toshio Sugihara Abstract In this study, an adaptive
Least Squares Estimation
Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David
Monotonicity Hints. Abstract
Monotonicity Hints Joseph Sill Computation and Neural Systems program California Institute of Technology email: [email protected] Yaser S. Abu-Mostafa EE and CS Deptartments California Institute of Technology
Multivariate Analysis of Ecological Data
Multivariate Analysis of Ecological Data MICHAEL GREENACRE Professor of Statistics at the Pompeu Fabra University in Barcelona, Spain RAUL PRIMICERIO Associate Professor of Ecology, Evolutionary Biology
Force/position control of a robotic system for transcranial magnetic stimulation
Force/position control of a robotic system for transcranial magnetic stimulation W.N. Wan Zakaria School of Mechanical and System Engineering Newcastle University Abstract To develop a force control scheme
Classifying Large Data Sets Using SVMs with Hierarchical Clusters. Presented by :Limou Wang
Classifying Large Data Sets Using SVMs with Hierarchical Clusters Presented by :Limou Wang Overview SVM Overview Motivation Hierarchical micro-clustering algorithm Clustering-Based SVM (CB-SVM) Experimental
0.1 Phase Estimation Technique
Phase Estimation In this lecture we will describe Kitaev s phase estimation algorithm, and use it to obtain an alternate derivation of a quantum factoring algorithm We will also use this technique to design
Part 2: Analysis of Relationship Between Two Variables
Part 2: Analysis of Relationship Between Two Variables Linear Regression Linear correlation Significance Tests Multiple regression Linear Regression Y = a X + b Dependent Variable Independent Variable
Coding and decoding with convolutional codes. The Viterbi Algor
Coding and decoding with convolutional codes. The Viterbi Algorithm. 8 Block codes: main ideas Principles st point of view: infinite length block code nd point of view: convolutions Some examples Repetition
How To Filter Spam Image From A Picture By Color Or Color
Image Content-Based Email Spam Image Filtering Jianyi Wang and Kazuki Katagishi Abstract With the population of Internet around the world, email has become one of the main methods of communication among
Comparison of K-means and Backpropagation Data Mining Algorithms
Comparison of K-means and Backpropagation Data Mining Algorithms Nitu Mathuriya, Dr. Ashish Bansal Abstract Data mining has got more and more mature as a field of basic research in computer science and
Dimensionality Reduction: Principal Components Analysis
Dimensionality Reduction: Principal Components Analysis In data mining one often encounters situations where there are a large number of variables in the database. In such situations it is very likely
Understanding and Applying Kalman Filtering
Understanding and Applying Kalman Filtering Lindsay Kleeman Department of Electrical and Computer Systems Engineering Monash University, Clayton 1 Introduction Objectives: 1. Provide a basic understanding
Character Image Patterns as Big Data
22 International Conference on Frontiers in Handwriting Recognition Character Image Patterns as Big Data Seiichi Uchida, Ryosuke Ishida, Akira Yoshida, Wenjie Cai, Yaokai Feng Kyushu University, Fukuoka,
Manifold Learning Examples PCA, LLE and ISOMAP
Manifold Learning Examples PCA, LLE and ISOMAP Dan Ventura October 14, 28 Abstract We try to give a helpful concrete example that demonstrates how to use PCA, LLE and Isomap, attempts to provide some intuition
2.2 Creaseness operator
2.2. Creaseness operator 31 2.2 Creaseness operator Antonio López, a member of our group, has studied for his PhD dissertation the differential operators described in this section [72]. He has compared
MANAGING QUEUE STABILITY USING ART2 IN ACTIVE QUEUE MANAGEMENT FOR CONGESTION CONTROL
MANAGING QUEUE STABILITY USING ART2 IN ACTIVE QUEUE MANAGEMENT FOR CONGESTION CONTROL G. Maria Priscilla 1 and C. P. Sumathi 2 1 S.N.R. Sons College (Autonomous), Coimbatore, India 2 SDNB Vaishnav College
Artificial Neural Networks and Support Vector Machines. CS 486/686: Introduction to Artificial Intelligence
Artificial Neural Networks and Support Vector Machines CS 486/686: Introduction to Artificial Intelligence 1 Outline What is a Neural Network? - Perceptron learners - Multi-layer networks What is a Support
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Keywords: Image complexity, PSNR, Levenberg-Marquardt, Multi-layer neural network.
Global Journal of Computer Science and Technology Volume 11 Issue 3 Version 1.0 Type: Double Blind Peer Reviewed International Research Journal Publisher: Global Journals Inc. (USA) Online ISSN: 0975-4172
RN-Codings: New Insights and Some Applications
RN-Codings: New Insights and Some Applications Abstract During any composite computation there is a constant need for rounding intermediate results before they can participate in further processing. Recently
Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus
Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus Tihomir Asparouhov and Bengt Muthén Mplus Web Notes: No. 15 Version 8, August 5, 2014 1 Abstract This paper discusses alternatives
Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data
CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear
Text Mining in JMP with R Andrew T. Karl, Senior Management Consultant, Adsurgo LLC Heath Rushing, Principal Consultant and Co-Founder, Adsurgo LLC
Text Mining in JMP with R Andrew T. Karl, Senior Management Consultant, Adsurgo LLC Heath Rushing, Principal Consultant and Co-Founder, Adsurgo LLC 1. Introduction A popular rule of thumb suggests that
Social Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
ECE 533 Project Report Ashish Dhawan Aditi R. Ganesan
Handwritten Signature Verification ECE 533 Project Report by Ashish Dhawan Aditi R. Ganesan Contents 1. Abstract 3. 2. Introduction 4. 3. Approach 6. 4. Pre-processing 8. 5. Feature Extraction 9. 6. Verification
Synchronization of sampling in distributed signal processing systems
Synchronization of sampling in distributed signal processing systems Károly Molnár, László Sujbert, Gábor Péceli Department of Measurement and Information Systems, Budapest University of Technology and
By choosing to view this document, you agree to all provisions of the copyright laws protecting it.
This material is posted here with permission of the IEEE Such permission of the IEEE does not in any way imply IEEE endorsement of any of Helsinki University of Technology's products or services Internal
Statistics Graduate Courses
Statistics Graduate Courses STAT 7002--Topics in Statistics-Biological/Physical/Mathematics (cr.arr.).organized study of selected topics. Subjects and earnable credit may vary from semester to semester.
A Reliability Point and Kalman Filter-based Vehicle Tracking Technique
A Reliability Point and Kalman Filter-based Vehicle Tracing Technique Soo Siang Teoh and Thomas Bräunl Abstract This paper introduces a technique for tracing the movement of vehicles in consecutive video
Class-specific Sparse Coding for Learning of Object Representations
Class-specific Sparse Coding for Learning of Object Representations Stephan Hasler, Heiko Wersing, and Edgar Körner Honda Research Institute Europe GmbH Carl-Legien-Str. 30, 63073 Offenbach am Main, Germany
CORRELATED TO THE SOUTH CAROLINA COLLEGE AND CAREER-READY FOUNDATIONS IN ALGEBRA
We Can Early Learning Curriculum PreK Grades 8 12 INSIDE ALGEBRA, GRADES 8 12 CORRELATED TO THE SOUTH CAROLINA COLLEGE AND CAREER-READY FOUNDATIONS IN ALGEBRA April 2016 www.voyagersopris.com Mathematical
E190Q Lecture 5 Autonomous Robot Navigation
E190Q Lecture 5 Autonomous Robot Navigation Instructor: Chris Clark Semester: Spring 2014 1 Figures courtesy of Siegwart & Nourbakhsh Control Structures Planning Based Control Prior Knowledge Operator
http://www.jstor.org This content downloaded on Tue, 19 Feb 2013 17:28:43 PM All use subject to JSTOR Terms and Conditions
A Significance Test for Time Series Analysis Author(s): W. Allen Wallis and Geoffrey H. Moore Reviewed work(s): Source: Journal of the American Statistical Association, Vol. 36, No. 215 (Sep., 1941), pp.
Mathematical Modelling of Computer Networks: Part II. Module 1: Network Coding
Mathematical Modelling of Computer Networks: Part II Module 1: Network Coding Lecture 3: Network coding and TCP 12th November 2013 Laila Daniel and Krishnan Narayanan Dept. of Computer Science, University
Operation Count; Numerical Linear Algebra
10 Operation Count; Numerical Linear Algebra 10.1 Introduction Many computations are limited simply by the sheer number of required additions, multiplications, or function evaluations. If floating-point
Visualization by Linear Projections as Information Retrieval
Visualization by Linear Projections as Information Retrieval Jaakko Peltonen Helsinki University of Technology, Department of Information and Computer Science, P. O. Box 5400, FI-0015 TKK, Finland [email protected]
Linear Threshold Units
Linear Threshold Units w x hx (... w n x n w We assume that each feature x j and each weight w j is a real number (we will relax this later) We will study three different algorithms for learning linear
Factoring Patterns in the Gaussian Plane
Factoring Patterns in the Gaussian Plane Steve Phelps Introduction This paper describes discoveries made at the Park City Mathematics Institute, 00, as well as some proofs. Before the summer I understood
Intrusion Detection via Machine Learning for SCADA System Protection
Intrusion Detection via Machine Learning for SCADA System Protection S.L.P. Yasakethu Department of Computing, University of Surrey, Guildford, GU2 7XH, UK. [email protected] J. Jiang Department
Probabilistic Latent Semantic Analysis (plsa)
Probabilistic Latent Semantic Analysis (plsa) SS 2008 Bayesian Networks Multimedia Computing, Universität Augsburg [email protected] www.multimedia-computing.{de,org} References
APPM4720/5720: Fast algorithms for big data. Gunnar Martinsson The University of Colorado at Boulder
APPM4720/5720: Fast algorithms for big data Gunnar Martinsson The University of Colorado at Boulder Course objectives: The purpose of this course is to teach efficient algorithms for processing very large
A secure face tracking system
International Journal of Information & Computation Technology. ISSN 0974-2239 Volume 4, Number 10 (2014), pp. 959-964 International Research Publications House http://www. irphouse.com A secure face tracking
SECRET sharing schemes were introduced by Blakley [5]
206 IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 52, NO. 1, JANUARY 2006 Secret Sharing Schemes From Three Classes of Linear Codes Jin Yuan Cunsheng Ding, Senior Member, IEEE Abstract Secret sharing has
METHODOLOGICAL CONSIDERATIONS OF DRIVE SYSTEM SIMULATION, WHEN COUPLING FINITE ELEMENT MACHINE MODELS WITH THE CIRCUIT SIMULATOR MODELS OF CONVERTERS.
SEDM 24 June 16th - 18th, CPRI (Italy) METHODOLOGICL CONSIDERTIONS OF DRIVE SYSTEM SIMULTION, WHEN COUPLING FINITE ELEMENT MCHINE MODELS WITH THE CIRCUIT SIMULTOR MODELS OF CONVERTERS. Áron Szûcs BB Electrical
Appendix 4 Simulation software for neuronal network models
Appendix 4 Simulation software for neuronal network models D.1 Introduction This Appendix describes the Matlab software that has been made available with Cerebral Cortex: Principles of Operation (Rolls
Tracking and Recognition in Sports Videos
Tracking and Recognition in Sports Videos Mustafa Teke a, Masoud Sattari b a Graduate School of Informatics, Middle East Technical University, Ankara, Turkey [email protected] b Department of Computer
IN current film media, the increase in areal density has
IEEE TRANSACTIONS ON MAGNETICS, VOL. 44, NO. 1, JANUARY 2008 193 A New Read Channel Model for Patterned Media Storage Seyhan Karakulak, Paul H. Siegel, Fellow, IEEE, Jack K. Wolf, Life Fellow, IEEE, and
Clustering & Visualization
Chapter 5 Clustering & Visualization Clustering in high-dimensional databases is an important problem and there are a number of different clustering paradigms which are applicable to high-dimensional data.
