Dimensionality Reduction of DNA Sequences through Wavelet Transforms
|
|
|
- Tabitha Sparks
- 10 years ago
- Views:
Transcription
1 Dimensionality Reduction of DNA Sequences through Wavelet Transforms MAMTA C. PADOLE Department of Computer Science and Engineering, The Maharaja Sayajirao University of Baroda Kalabhavan, Baroda, Gujarat, India 39 1 [email protected] Abstract: - In recent years, development in molecular biology and sequencing technology has led to large sequencing of genomes, metagenomics and transcriptome. As a result sequencing data is growing exponentially. This mounting data needs secured storage and frequent analysis to resolve several mysteries of nature. Although, cost of storage capacities in computers is reducing and processing power enhancing with, there still arises the need of storing Big data efficiently and applying various analysis techniques optimally. This paper discusses, the use of Haar Wavelet Transforms, a signal processing approach, for lossless data compaction for Genomic data. The use of Wavelet Transforms also guarantees perfect reconstruction. The advantage of Wavelet Transform is its computational complexity being O(log n) and is memory efficient as it can be computed in place, without a temporary buffer. Key-Words: - Wavelet Transforms, Signal Processing, Haar Wavelets, Dimensionality Reduction, Data Compaction 1 Introduction High throughput sequencing techniques has led to large sequencing of genomes, metagenomics and transcriptome. As a result sequencing data is growing exponentially. This mounting data causes lot of challenges to Computer Scientists to store this data efficiently and analyse it optimally. Applying various techniques for storage or analysis of raw data in terms of DNA sequences or protein sequences, involves string processing or textual processing, somes along with different data structures for string processing. String processing algorithms are consuming and usually involve linear or exponential computational complexity. The existing compression tools are primarily based on the classic Lempel-Ziv algorithm [1], the bwt (Burrows-Wheeler transform) [2], Huffman Coding [3], FM-Index [4] and Wavelet Trees [5] These algorithms or tools for DNA compression take statistical or dictionary based approaches. Statistical based algorithms need prior knowledge of the data in concern and hence are data dependent. formatdb [6] and GRS [7] uses Huffman coding, but deals efficiently only with minor symbols [8]. The Huffman code is an optimal prefix code in the case where exact symbols probabilities are known in advance and are integral powers of 1/2 [9][1]. The algorithms based on Arithmetic coders are slow in decompression process [8]. Arithmetic Coders and Huffman Coding use probability based compression which is difficult to predict, as DNA sequences comprises of nucleotide bases belonging to an alphabet size of 4, and hence each nucleotide base has almost similar probability [11]. The other algorithms applied in tools like GSR [12] or BioCompress [13] take into account characteristics of DNA like reverse complement or point mutation or characteristic structures of sequences like SNP s or repeat regions [14]. But SNP map may not be available, SNP may not be the common bi-allele [15], Index of repetitiveness may vary in different genomes [16], or organisms have very little variation and hence may become difficult to compress. If, a priori knowledge of the characteristics of the given DNA sequence is not available, then statistical approach becomes irrelevant and hence the problem of data compression becomes difficult [1]. Reference sequence is used for data compression of DNA [17]. Another algorithm used in software tool coil [18] uses Levenstein distances and encoding trees for compressing entire database of DNA data, but takes ISBN:
2 into account an initial DNA sequence to compare length-k substrings amongst all the other DNA sequences in database. This again is data dependent and will work efficiently only if all the sequences are nearly similar. For widely varied sequences in the database, the algorithm may prove inefficient [18] Techniques implemented on hardware based acceleration device, have also been used for data compression to accelerate DNA sequence alignment [19]. In this paper, we present an ab initio method of data transforms which can be applied for optimal analysis. This paper emphasizes on the use of Wavelet Transforms particularly the Haar Wavelets for data compaction of genomic data, principally the DNA sequences. After single level of Wavelet Transformation, there are always N elements, same as the original input sequence. But the fact is, that we do not need all N elements when we want to compare two sequences or need to perform sequence analysis. For local alignment, fine comparision is needed and this can be achieved through detailed co-efficient C d. For global alignment of two DNA sequences, the coarse components i.e. approximate co-efficient C a are enough. So, in all N/2 elements are sufficient for actual use. Thus, the paper proposes to reduce the number of elements used for further analysis, with an apriori approach. The content and positional information is preserved even after transform on which the analysis of data can be performed [2][21]. The complexity of Wavelet transforms is O(log n), n being the length of the input sequence. Also, use of Haar Wavelets is proved to be memory efficient. 2 WAVELET TRANSFORMS A wavelet transform is a linear transformation of a signal or data into coefficients on a basis of wavelet functions [22]. The wavelet transformations represent the data in one domain into another, from where hidden information can be explored. Wavelet transform provides the - information[23]. A signal is decomposed through Wavelet transforms into primarily two co-efficient vectors, which capture trends at different resolution levels [24]. These coefficient vectors of different resolution levels represent the characteristics of the data, at each different. Its performance efficiency is logarithmic, in and space. When a signal x is passed through low pass filters (scaling functions) and high pass filters (wavelet functions) simultaneously, it is defined as performing the discrete wavelet transform (DWT). The number of coefficients generated will be half the length of the original input to each filter, when a signal is passed simultaneously through low pass and high pass filters and on performing subsequent down-sampling. Therefore, the wavelet transform of a signal when passed through low pass filters and high pass filters generates the Approximate coefficient C a (1) and Detail co-efficient C d (2), as represented in equations (1) and (2) [25]. (1) (2) Here g is a scaling function and h is the wavelet function The property of the wavelet transform is that the lower order coefficients of the decomposition represent the broad trends or coarse s in the data, whereas the more localized trends or fine s are captured by the higher order coefficients. Filtering of the signal at each pass results in, a convolution of the signal x and the impulse response g of the filter. y(n) is the scalar product of functions x and g. Mathematically, this result can be represented as in (3): (3) Thus Wavelet Transform W T can also be represented as in (4), (4) Wavelets function f(t) consists of scalar product of the basis functions φ(x) and ψ(x) as in (5) (5) φ(x) is called scaling function to find the approximation and ψ(x) is called wavelet function that computes the details. Wavelet decomposition can be applied to any sequential data, including strings, where the position of a character in string represents the series ISBN:
3 data. Wavelets are tools used to study regularity and to conduct local studies [25]. The zero moments of the function are related to the regularity of scaling function and wavelets [26]. 2.1 Haar Wavelets The Haar transform is the simplest and oldest compact, dyadic, orthonormal wavelet transform [27]. The Haar function being an odd rectangular pulse pair, with compact support provides good possibility for local analysis of signal [28]. Haar wavelets are conceptually simple, fast and memory efficient [28], [29], can be computed in place, without a temporary buffer, are exactly reversible and can be perfectly reconstructed. The Haar transform decomposes a discrete signal x into two sub-signals of half its length. One sub-signal is a running average or trend (C a ) as in Table 1; the other sub-signal is a running difference or fluctuation (C d ) Computation of Haar Wavelet Transform of Time Series Data Consider a one-dimensional data vector X containing the N = 8 data values X = [ 18, 16, 6, 6, 12, 2, 4, 12] Where, X is having number of elements n = 8, which is usually the power of 2. Haar Wavelet Transform of the signal X can be computed by iteratively performing pair-wise averaging and semi differencing [3], more precisely, convolving with 1 1 the basis vector <, >. As shown in Table 1, 2 2 we can observe that, the Haar Transform W T of the original signal X is given as: W T = [ 47/, -1/, 11, 8, 1/,, -8/, -8/ ] TABLE 1. Representation of Computations of Haar Wavelet Transform TransformatScal Resoluti Leng Averag Differe ion Level or e on th of es / nces / Decomposit (Sca (ResolutiSign Approx Detail ion Level (j) le a on is = 2 j inverse al/ (L) imate Coefficien Coefficien ) of Scale t (C d ) i.e. 1/a ) t (C a ) C d = (Original Signal) C a = (X i + X i+1 ) / (X i - X i+1 ) / (a) ( b) (c) ( d) ( e) (f) 2 = = 2 1/2 = 1 8 [18, 16, 6, 6, 1 2, 2, 4, 12] 1/2 1 = 1/ = 1/2 2 = 4 1/ = 1/2 3 = 8 1/8 4 [34/, 12/, 32/, 16/ [1/,, -8 /, -8/ ] 2 [23, 24] [11, 8] 1 [47/ ] [ 1 ] The Haar Wavelet Transform consists of basis functions and. The scaling function determines the or dilation/expansion and mother wavelet function determines the translation or shift. and for Haar Wavelet Transform can be written as, (6) (7) Where, h(), h(1) and g(), g(1) are known as the low-pass and high-pass filters respectively. These filters are used to convolve with the scaling function and mother wavelet functions respectively to generate the wavelet transform of a function. In Haar Wavelet Transform, the values of these lowpass and high pass filters are expressed as given in (8) to (11) (8) (9) (1) (11) Where, h() and g() are analysis filters and h(1) and g(1) are synthesis filters. ISBN:
4 In other words, these filter values can also be represented in terms of basis vectors as and <. Since, the basis vectors used in Haar Discrete Wavelet Transform are the smallest possible basis vectors; it is not possible to do Haar Transform in one-pass. Thus, it becomes essential to recursively transform the input signal using these basis vectors [31]. This concept is clearly shown intable 1. The scaling function in Haar Wavelet Transform is defined as in (12) (12) The mother wavelet function or translation function in Haar Wavelet Transform is defined as in (13), [23] 3 METHODOLOGY Algorithm: (13) 1. Read Fasta File containing Pyrosequenced Reads. 2. Assign binary values to nucleotide bases, thus converting nucleotide sequence into single indicator sequence. 3. Combine four binary values and put it as a single byte. Thus compressing the sequence one-fourth of the original sequence. 4. Perform four levels Haar Wavelet Transform on Numerical Sequence, which in turn reduces the sequence by Perform Steps 2 to 4 for all sequences in the Fasta File. 6. If at each level of transform, repeating sequences are removed, it would give further reduction of data. The suggested algorithm for creating a compressed form of a given sequence is applied as follows: Read the Fasta file which contains the sequence S i = {s 1, s 2, s n } where s i ϵ = {A,C,G,T}, i = 1, n and n is the length of S i. Convert the nucleotide sequence S i into its numerical representation X i. The single indicator sequence using binary representation of {A, C, G, T} as {, 1, 1, 11} is used. This binary representation of the nucleotide bases is than combined into a group of four and each of these four binary representations of the four nucleotide bases is than stored in a single byte. Thus, only one byte is used to represent four consecutive nucleotide bases. The use of binary representation in a single byte reduces the computational overhead to 25% compared, to the other representation of nucleotide sequences like dipole moments [22] or EIIP values [32]. Only numerical representations can be applied for wavelet transformations. Thus, if a given nucleotide sequence is: S i = GTGCAGGATGCTGCAA i = [1, n], n being length of nucleotide sequence Then, its numerical representation i.e. a digital signal can be given as: X j = 185, 4, 231, 144, j = [1, m], m being length of numerical representation of binary values, as a single byte for four nucleotides. m = 1 n/4 Perform first-level decomposition of this signal, using Haar Wavelet Transform. This would reduce the signal into one-half of the original numerical representation of the nucleotide sequences. Thus 5% compaction is achieved by single level wavelet transform. Further decomposition of a signal, to 2nd level, 3rd level and 4th level would further reduce the size of the signal by two. Thus, applying repeated wavelet transforms reduces the signal size and hence requires, less storage space compared to storing the original nucleotide sequence. This reduced size would also reduce the involved in. Also, Searching on the transformed signal is equivalent to searching on a reduced data set of the original sequence and hence is more efficient. If the correct mapping/transformation is applied, the Parseval Theorem implies that frequency distance FD is less than or equal to edit distance ED (Distance preserving transformation) [33]. With every transform, we are discarding half the values as per Nyquist s rule [22] and hence optimizing the search, with complexity of O(log n). The Haar wavelet transforms Ca i.e. running average represents the trend; the other sub-signal, ISBN:
5 Cd a running difference, represents fluctuations [34]. Hence, Haar wavelet, can be used to transform the signals for data analysis too, besides applying it for compaction. Thus, use of this algorithm does compression at two levels. Initially converting four nucleotide bases into a single byte through binary representation compresses the data to one-fourth. Further, applying wavelet transforms four s reduces the data to 1/2 4 s i.e. one-sixteenth of the numerical representation of the original nucleotide sequence. Thus, the proposed algorithm applies compression twice i.e. The size of the compressed signal C k becomes, C k = 1/ /2 4 = 1/2 6 = 1/64. Thus, C k = S i * 1/64. So, 64 s compaction of data can be achieved using the proposed algorithm. Storing this compact data would result in storage efficiency as well, will be much optimal for any type of analysis. Overall, the complexity of the proposed algorithm applied for compression would be linear. Complexity of proposed algorithm for compression = Complexity of converting nucleotide sequence to binary representation + Complexity of Wavelet transform = O(n) + O(log n) = O(n). Numerical Representation through Binary Representation of {A, C, G,T} as {, 1, 1, 11} respectively and storing four consecutive characters as binary form in a single byte. The following are the decimal values of the converted binary form: (Total 3 elements i.e. 4 s compression) i.e. X j: 254, 169, 185, 4, 231, 144, 6, 243, 246, 144, 131, 152, 148, 3, 142, 82, 224, 174, 115, 242, 163, 188, 18, 21, 134, 27, 48, 197, 165, 184 After applying Fourth Level Wavelet Transform, Data appears (Total 2 elements i.e. 2 4 s thus 3/16 ~ 3/15 = 2 elements) as C k : , The graphical representation of the numerical form of original sequence, 4-levels of decompositions and the reconstructed sequence, is given in Fig. 1. The X axis represents the Time or Space Localization and the Y axis represents the Scale or Frequency Localization. Fig. 1 represents the output of the proposed algorithm for a sample sequence as explained in Example 1. The four level decompositions clearly display reducing length trend, each being the half of the previous signal, using multi-level Haar Wavelet Transform. The graph labeled reconstructed seqn clearly represents that the signal can be perfectly reconstructed, as it is exactly same as the graph labeled as original seqn using multi-level Haar Wavelet Transform. The graph labeled reconstructed seqn clearly represents that the signal can be perfectly reconstructed, as it is exactly same as the graph labeled as original seqn 4 RESULT AND DISCUSSION Example-1: Example of Synthetic Nucleotide sequence, before and after applying the proposed algorithm. The sequence can be perfectly reconstructed as demonstrated in Fig. 1. If, the Original Read (Total 121 characters) S i is: TTTGGGGCGTGCAGGATGCTGCAAAACGTT ATTTCGGCAAGAATGCGAGCCAACTGGATG CCAGTGAAGGTGCTATTTAGGGATGTTACGT AACCCGACGTATTATAATACCGGCCGTGAT original seqn level 1 decomposition level 3 decomposition reconstructed seqn level 2 decomposition level 4 decomposition Fig. 1: Visual representation of four level decompositions displaying reducing length trend, of a sequence as explained in Example 1, ISBN:
6 Example-2: SRA Read of Metagenomics sequence having Accession Number >gnl SRA SRR EXHS9OF1EH7NX.2 is used to apply the proposed algorithm. (Refer Fig. 2) original seqn level 1 decomposition reconstructed seqn level 2 decomposition 4 Example-3: Contig 46 of E. Coli K-12 strain having Accession Number gi gb AEFE146.1 is used to apply the proposed algorithm. Any type of Nucleotide sequence may be used for compression through application of the proposed algorithm. The sequence can be perfectly reconstructed, and can be compressed to almost 64 s, as demonstrated in Fig. 2. Table 2, further supports the compression efficiency of the proposed algorithm because the cases considered for compression, show nearly 64 s of compression ratio which is worth for a big data of genomics. The computational was calculated for an experiment conducted on regular laptop with Intel Core2 Duo CPU, 2.1 GHz x 2.1 GHz processor and 32-bit Windows 7 Ultimate Operating System with 4.GB RAM using MATLAB R29a version The computing environment did not consider any special arrangement; instead the experiments were performed, while all routine services including Oracle Database service and Apache Tomcat Application Server was running on an average personal machine. This type of environment used for computing, simply proves that special purpose computer or clustered computing environment is not essential for executing this algorithm. It can be smoothly executed on a simple desktop machine too. The computational can be optimized to a great extent, if special purpose computers are used instead of shared resources. Thus, the proposed algorithm can be applied on any form of genomic data. 1-1 Fig. 2: Visual representation of compressed form metagenomics sequence having Accession Number >gnl SRA SRR EXHS9OF1EH7NX.2, using proposed algorithm applying multi-level Haar Wavelet Transform. TABLE 2. Represents the Compression Ratio of Original Sequence and Compressed Sequence after binary representation and after four levels of Wavelet Transforms, as proposed in the algorithm Accessi on Numbe r level 3 decomposition Descrip tion of Sequen ce Origin al Nucleot ide Sequen ce Size (In base pairs) Size in Decima l values of Binary Form of Nucleot ide (In Numbe r of elemen ts) level 4 decomposition 5 Size 4 th level Wavele t Transf ormed Sequen ce (In Numbe r of elemen ts) Comp ressio n Ratio (a) (b) (c) = S i (d) = X j (e) = C k (f) = (c)/(e ) - Sample Read gnl SR A SRR Metage nomic Read 1.2 EXHS9 OF1E H7NX ISBN:
7 gi 356 Contig 946 g 46 of b AEFE E.Coli 1 K strain gi 2245 Chromo 1493 r some ef NT_ 7genom gi r ef NC_ gi r ef NC_ NC_ 7327 ic Contig of Homo Sapiens Chromo some 3 of Caenor habditis Elegans Chromo some 26 of Bos Taurus 5. CONCLUSION The research concludes that discrete wavelet transform, particularly Haar wavelets, can be used to compress the genomic data. The proposed algorithm using signal processing, compresses the DNA sequences in terms of the number of elements, preserving the details of the original signal. The information of original DNA sequence is preserved in a signal compressed after wavelet transform, be it content of the sequence or the positional information. Thus, using transformed data in place of original sequence for data analysis would also be possible, because there is no loss of data due to transformation. The wavelet transforms have a complexity of O(log n). Hence using Haar wavelet transforms is much efficient than other classic string or regex based algorithms, with exponential complexity. In all, the proposed algorithm using Haar Wavelet Transforms, is an efficient way to compress nucleotide sequences, in linear, without requirement of complex data structures and can be used for data analysis. The enhancement to this work is, of applying wavelet transforms on large sequences using distributed tool box of MATLAB. The research also aims to apply this transformed data using the Wavelet Transforms, for different types of sequence analysis. REFERENCES [1] Ziv J, Lempel A. A universal algorithm for sequential data compression, IEEE Transaction of Information Theory 1977; 23(3): [2] Burrows M, Wheeler D: A block sorting lossless data compression algorithm. Tech. Rep. 124, Digital Equipment Corporation [3] Huffman D.A. A Method for the Construction of Minimum-Redundancy Codes. Proceedings of the I.R.E., September 1952, pp Huffman's original article. [4] Paolo Ferragina, Giovanni Manzini, Veli M akinen, and Gonzalo Navarro, An Alphabet- Friendly FM-index? In International Symposium on String Processing and Information Retrieval (SPIRE), (24) pages [5] Oˇguzhan M, Vitter J S, Xu B, Efficient Maximal Repeat Finding Using the Burrows- Wheeler Transform and Wavelet Tree, IEEE/ACM Transactions On Computational Biology And Bioinformatics. [6] [7] Congmao Wang1 and Dabing Zhang, A novel compression tool for efficient storage of genome resequencing data [8] Hisahiko Sato1 Takashi Yoshioka, Akihiko Konagaya3 Tetsuro Toyoda,DNA Data Compression in the Post Genome Era, Genome Informatics 12: (21) [9] Ana Balevic, Lars Rockstroh, Marek Wroblewski, and Sven Simon, Using Arithmetic Coding for Reduction of Resulting Simulation Data Size on Massively Parallel GPGPUs [1] Sayood, K., Lossless Compression Handbook. Academic Press (23) [11] Toshiko Matsumoto Kunihiko Sadakane, Hiroshi Imai, Biological Sequence Compression Algorithms, Genome Informatics 11: (2) 43 [12] Chen, X., Kwong, S., and Li, M., A compression algorithm for DNA sequences and its applications in genome comparison, Genome Informatics, 1:52 61, [13] Grumbach, S. and Tahi, F., A new challenge for compression algorithms: genetic sequences, Information Processing & Management, 3: , [14] Kenny Daily, Paul Rigor, Scott Christley, Xiaohui Xie, Pierre Baldi, Data structures and compression algorithms for high-throughput ISBN:
8 sequencing technologies, BMC Bioinformatics 21, 11:514 [15] Samantha Woodward, A Critical Analysis of DNA Data Compression Methods BIOC 218, Winter [16] Bernhard Haubold andthomas Wiehe, How repetitive are genomes?, BMC Bioinformatics 26, 7:541 doi : / [17] Markus Hsi-Yang Fritz, Rasko Leinonen, Guy Cochrane, and Ewan Birney, Efficient storage of high throughput DNA sequencing data using reference-based compression, Genome Research, 21:734 74_211 by Cold Spring Harbor Laboratory Press; ISSN /11; [18] W Timothy J White* and Michael D Hendy,Compressing DNA sequence databases with coil, BMC Bioinformatics 28, 9:242 doi:1.1186/ [19] Al Junid, S. A. M. Tahir, N.M. Haron, M.A. Abd Majid, Z., Idros, M.F. Osman, F.N., Development and Implementation of Novel Data Compression Technique for Accelerate DNA Sequence Alignment Based on Smith Waterman Algorithm, IJSSST, Vol. 11, No. 3 ISSN: x online, [2] Mamta C. Padole, B. S. Parekh, D. P. Patel, Signal Processing Approach for Recognizing Identical Reads From DNA Sequencing of Bacillus Strains, IOSR Journal of Computer Engineering, Mar-Apr 213, pp [21] Mamta C. Padole, Recognizing Short Tandem Repeat Regions in Genomic Sequences Using Wavelet Transforms, unpublished [22] J. K. Meher, M. R. Panigrahi, G. N. Dash, P. K. Meher, Wavelet Based Lossless DNA Sequence Compression for Faster Detection of Eukaryotic Protein Coding Regions, I.J. Image, Graphics and Signal Processing 212, 7, [23] I. Daubechies, Ten Lectures On Wavelets, 1992 [24] Charu C. Aggarwal, On the Use of Wavelet Decomposition for String Classification, Springer - Data Mining and Knowledge Discovery, 1, , 25 [25] MATLAB manual: [26] C.S.Burrus, R.A. Gopinath, Haitau Guo, Introduction to Wavelets and Wavelet Transforms, Prentice Hall, 1998 [27] Moharir PS, Pattern recognition transforms, New York: Wiley, 1992 [28] Radomir S. Stankovic, Bogdan J. Falkowski, The Haar wavelet transform: its status and achievements, Elsevier Science Ltd. Computers and Electrical Engineering, (23) [29] Ruiz G, Michell JA, Buron A, Switch-level fault detection and diagnosis environment for MOS digital circuits using spectral techniques, IEE Proc Part E, 1992; 139(4): [3] Dimitris Sacharidis, Constructing Optimal Wavelet Synopses, Proceedings of the 26 International Conference on Current Trends in Database Technology EBDT 6, Pg 97-14, / _1 [31] Musawir Ali, An Introduction to Wavelets and the Haar Transform, [32] A. S. Nair and S. P. Sreenathan, A Coding Measure Scheme Employing Electron-Ion Interaction Pseudopotential (EIIP), Bioinformation, Vol. 1, No. 6, 26, pp [33] Aghili, Alireza A, Agrawal, D El Abbadi, A, Sequence Similarity Search Using Discrete Fourier and Wavelet Transformation Techniques, Postprints, UC Santa Barbara, 25 [34] I. Daubechies, Orthonormal Bases of Compactly Supported Wavelets, Comm. Pure Appl. Math, Vol 41, 1988, pp ISBN:
DNA Sequencing Data Compression. Michael Chung
DNA Sequencing Data Compression Michael Chung Problem DNA sequencing per dollar is increasing faster than storage capacity per dollar. Stein (2010) Data 3 billion base pairs in human genome Genomes are
DNA Mapping/Alignment. Team: I Thought You GNU? Lars Olsen, Venkata Aditya Kovuri, Nick Merowsky
DNA Mapping/Alignment Team: I Thought You GNU? Lars Olsen, Venkata Aditya Kovuri, Nick Merowsky Overview Summary Research Paper 1 Research Paper 2 Research Paper 3 Current Progress Software Designs to
Image Compression through DCT and Huffman Coding Technique
International Journal of Current Engineering and Technology E-ISSN 2277 4106, P-ISSN 2347 5161 2015 INPRESSCO, All Rights Reserved Available at http://inpressco.com/category/ijcet Research Article Rahul
HIGH DENSITY DATA STORAGE IN DNA USING AN EFFICIENT MESSAGE ENCODING SCHEME Rahul Vishwakarma 1 and Newsha Amiri 2
HIGH DENSITY DATA STORAGE IN DNA USING AN EFFICIENT MESSAGE ENCODING SCHEME Rahul Vishwakarma 1 and Newsha Amiri 2 1 Tata Consultancy Services, India [email protected] 2 Bangalore University, India ABSTRACT
Wavelet analysis. Wavelet requirements. Example signals. Stationary signal 2 Hz + 10 Hz + 20Hz. Zero mean, oscillatory (wave) Fast decay (let)
Wavelet analysis In the case of Fourier series, the orthonormal basis is generated by integral dilation of a single function e jx Every 2π-periodic square-integrable function is generated by a superposition
Analysis of Compression Algorithms for Program Data
Analysis of Compression Algorithms for Program Data Matthew Simpson, Clemson University with Dr. Rajeev Barua and Surupa Biswas, University of Maryland 12 August 3 Abstract Insufficient available memory
CHAPTER 2 LITERATURE REVIEW
11 CHAPTER 2 LITERATURE REVIEW 2.1 INTRODUCTION Image compression is mainly used to reduce storage space, transmission time and bandwidth requirements. In the subsequent sections of this chapter, general
Lossless Grey-scale Image Compression using Source Symbols Reduction and Huffman Coding
Lossless Grey-scale Image Compression using Source Symbols Reduction and Huffman Coding C. SARAVANAN [email protected] Assistant Professor, Computer Centre, National Institute of Technology, Durgapur,WestBengal,
Storage Optimization in Cloud Environment using Compression Algorithm
Storage Optimization in Cloud Environment using Compression Algorithm K.Govinda 1, Yuvaraj Kumar 2 1 School of Computing Science and Engineering, VIT University, Vellore, India [email protected] 2 School
SPEECH SIGNAL CODING FOR VOIP APPLICATIONS USING WAVELET PACKET TRANSFORM A
International Journal of Science, Engineering and Technology Research (IJSETR), Volume, Issue, January SPEECH SIGNAL CODING FOR VOIP APPLICATIONS USING WAVELET PACKET TRANSFORM A N.Rama Tej Nehru, B P.Sunitha
Co-integration of Stock Markets using Wavelet Theory and Data Mining
Co-integration of Stock Markets using Wavelet Theory and Data Mining R.Sahu P.B.Sanjeev [email protected] [email protected] ABV-Indian Institute of Information Technology and Management, India Abstract
How To Use Neural Networks In Data Mining
International Journal of Electronics and Computer Science Engineering 1449 Available Online at www.ijecse.org ISSN- 2277-1956 Neural Networks in Data Mining Priyanka Gaur Department of Information and
Introduction to Medical Image Compression Using Wavelet Transform
National Taiwan University Graduate Institute of Communication Engineering Time Frequency Analysis and Wavelet Transform Term Paper Introduction to Medical Image Compression Using Wavelet Transform 李 自
ANALYSIS OF THE EFFECTIVENESS IN IMAGE COMPRESSION FOR CLOUD STORAGE FOR VARIOUS IMAGE FORMATS
ANALYSIS OF THE EFFECTIVENESS IN IMAGE COMPRESSION FOR CLOUD STORAGE FOR VARIOUS IMAGE FORMATS Dasaradha Ramaiah K. 1 and T. Venugopal 2 1 IT Department, BVRIT, Hyderabad, India 2 CSE Department, JNTUH,
A Wavelet Based Prediction Method for Time Series
A Wavelet Based Prediction Method for Time Series Cristina Stolojescu 1,2 Ion Railean 1,3 Sorin Moga 1 Philippe Lenca 1 and Alexandru Isar 2 1 Institut TELECOM; TELECOM Bretagne, UMR CNRS 3192 Lab-STICC;
DYNAMIC DOMAIN CLASSIFICATION FOR FRACTAL IMAGE COMPRESSION
DYNAMIC DOMAIN CLASSIFICATION FOR FRACTAL IMAGE COMPRESSION K. Revathy 1 & M. Jayamohan 2 Department of Computer Science, University of Kerala, Thiruvananthapuram, Kerala, India 1 [email protected]
Fast Arithmetic Coding (FastAC) Implementations
Fast Arithmetic Coding (FastAC) Implementations Amir Said 1 Introduction This document describes our fast implementations of arithmetic coding, which achieve optimal compression and higher throughput by
Sachin Patel HOD I.T Department PCST, Indore, India. Parth Bhatt I.T Department, PCST, Indore, India. Ankit Shah CSE Department, KITE, Jaipur, India
Image Enhancement Using Various Interpolation Methods Parth Bhatt I.T Department, PCST, Indore, India Ankit Shah CSE Department, KITE, Jaipur, India Sachin Patel HOD I.T Department PCST, Indore, India
EFFICIENT DATA PRE-PROCESSING FOR DATA MINING
EFFICIENT DATA PRE-PROCESSING FOR DATA MINING USING NEURAL NETWORKS JothiKumar.R 1, Sivabalan.R.V 2 1 Research scholar, Noorul Islam University, Nagercoil, India Assistant Professor, Adhiparasakthi College
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Final Year Project Progress Report. Frequency-Domain Adaptive Filtering. Myles Friel. Supervisor: Dr.Edward Jones
Final Year Project Progress Report Frequency-Domain Adaptive Filtering Myles Friel 01510401 Supervisor: Dr.Edward Jones Abstract The Final Year Project is an important part of the final year of the Electronic
A Tutorial in Genetic Sequence Classification Tools and Techniques
A Tutorial in Genetic Sequence Classification Tools and Techniques Jake Drew Data Mining CSE 8331 Southern Methodist University [email protected] www.jakemdrew.com Sequence Characters IUPAC nucleotide
New Hash Function Construction for Textual and Geometric Data Retrieval
Latest Trends on Computers, Vol., pp.483-489, ISBN 978-96-474-3-4, ISSN 79-45, CSCC conference, Corfu, Greece, New Hash Function Construction for Textual and Geometric Data Retrieval Václav Skala, Jan
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 17 Shannon-Fano-Elias Coding and Introduction to Arithmetic Coding
Memory Allocation Technique for Segregated Free List Based on Genetic Algorithm
Journal of Al-Nahrain University Vol.15 (2), June, 2012, pp.161-168 Science Memory Allocation Technique for Segregated Free List Based on Genetic Algorithm Manal F. Younis Computer Department, College
A Dynamic Programming Approach for Generating N-ary Reflected Gray Code List
A Dynamic Programming Approach for Generating N-ary Reflected Gray Code List Mehmet Kurt 1, Can Atilgan 2, Murat Ersen Berberler 3 1 Izmir University, Department of Mathematics and Computer Science, Izmir
Computer Networks and Internets, 5e Chapter 6 Information Sources and Signals. Introduction
Computer Networks and Internets, 5e Chapter 6 Information Sources and Signals Modified from the lecture slides of Lami Kaya ([email protected]) for use CECS 474, Fall 2008. 2009 Pearson Education Inc., Upper
Non-negative Matrix Factorization (NMF) in Semi-supervised Learning Reducing Dimension and Maintaining Meaning
Non-negative Matrix Factorization (NMF) in Semi-supervised Learning Reducing Dimension and Maintaining Meaning SAMSI 10 May 2013 Outline Introduction to NMF Applications Motivations NMF as a middle step
International Language Character Code
, pp.161-166 http://dx.doi.org/10.14257/astl.2015.81.33 International Language Character Code with DNA Molecules Wei Wang, Zhengxu Zhao, Qian Xu School of Information Science and Technology, Shijiazhuang
FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM
International Journal of Innovative Computing, Information and Control ICIC International c 0 ISSN 34-48 Volume 8, Number 8, August 0 pp. 4 FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT
Efficient Parallel Execution of Sequence Similarity Analysis Via Dynamic Load Balancing
Efficient Parallel Execution of Sequence Similarity Analysis Via Dynamic Load Balancing James D. Jackson Philip J. Hatcher Department of Computer Science Kingsbury Hall University of New Hampshire Durham,
Reconfigurable FPGA Inter-Connect For Optimized High Speed DNA Sequencing
Reconfigurable FPGA Inter-Connect For Optimized High Speed DNA Sequencing 1 A.Nandhini, 2 C.Ramalingam, 3 N.Maheswari, 4 N.Krishnakumar, 5 A.Surendar 1,2,3,4 UG Students, 5 Assistant Professor K.S.R College
A Performance Study of Load Balancing Strategies for Approximate String Matching on an MPI Heterogeneous System Environment
A Performance Study of Load Balancing Strategies for Approximate String Matching on an MPI Heterogeneous System Environment Panagiotis D. Michailidis and Konstantinos G. Margaritis Parallel and Distributed
Compression techniques
Compression techniques David Bařina February 22, 2013 David Bařina Compression techniques February 22, 2013 1 / 37 Contents 1 Terminology 2 Simple techniques 3 Entropy coding 4 Dictionary methods 5 Conclusion
An Overview of Knowledge Discovery Database and Data mining Techniques
An Overview of Knowledge Discovery Database and Data mining Techniques Priyadharsini.C 1, Dr. Antony Selvadoss Thanamani 2 M.Phil, Department of Computer Science, NGM College, Pollachi, Coimbatore, Tamilnadu,
Admin stuff. 4 Image Pyramids. Spatial Domain. Projects. Fourier domain 2/26/2008. Fourier as a change of basis
Admin stuff 4 Image Pyramids Change of office hours on Wed 4 th April Mon 3 st March 9.3.3pm (right after class) Change of time/date t of last class Currently Mon 5 th May What about Thursday 8 th May?
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin
Wan Accelerators: Optimizing Network Traffic with Compression. Bartosz Agas, Marvin Germar & Christopher Tran
Wan Accelerators: Optimizing Network Traffic with Compression Bartosz Agas, Marvin Germar & Christopher Tran Introduction A WAN accelerator is an appliance that can maximize the services of a point-to-point(ptp)
Fast Multipole Method for particle interactions: an open source parallel library component
Fast Multipole Method for particle interactions: an open source parallel library component F. A. Cruz 1,M.G.Knepley 2,andL.A.Barba 1 1 Department of Mathematics, University of Bristol, University Walk,
Non-Data Aided Carrier Offset Compensation for SDR Implementation
Non-Data Aided Carrier Offset Compensation for SDR Implementation Anders Riis Jensen 1, Niels Terp Kjeldgaard Jørgensen 1 Kim Laugesen 1, Yannick Le Moullec 1,2 1 Department of Electronic Systems, 2 Center
RN-Codings: New Insights and Some Applications
RN-Codings: New Insights and Some Applications Abstract During any composite computation there is a constant need for rounding intermediate results before they can participate in further processing. Recently
Network (Tree) Topology Inference Based on Prüfer Sequence
Network (Tree) Topology Inference Based on Prüfer Sequence C. Vanniarajan and Kamala Krithivasan Department of Computer Science and Engineering Indian Institute of Technology Madras Chennai 600036 [email protected],
A Survey on Outlier Detection Techniques for Credit Card Fraud Detection
IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661, p- ISSN: 2278-8727Volume 16, Issue 2, Ver. VI (Mar-Apr. 2014), PP 44-48 A Survey on Outlier Detection Techniques for Credit Card Fraud
Neural Networks in Data Mining
IOSR Journal of Engineering (IOSRJEN) ISSN (e): 2250-3021, ISSN (p): 2278-8719 Vol. 04, Issue 03 (March. 2014), V6 PP 01-06 www.iosrjen.org Neural Networks in Data Mining Ripundeep Singh Gill, Ashima Department
Statistical Modeling of Huffman Tables Coding
Statistical Modeling of Huffman Tables Coding S. Battiato 1, C. Bosco 1, A. Bruna 2, G. Di Blasi 1, G.Gallo 1 1 D.M.I. University of Catania - Viale A. Doria 6, 95125, Catania, Italy {battiato, bosco,
A FAST STRING MATCHING ALGORITHM
Ravendra Singh et al, Int. J. Comp. Tech. Appl., Vol 2 (6),877-883 A FAST STRING MATCHING ALGORITHM H N Verma, 2 Ravendra Singh Department of CSE, Sachdeva Institute of Technology, Mathura, India, [email protected]
Management Challenge. Managing Hardware Assets. Central Processing Unit. What is a Computer System?
Management Challenge Managing Hardware Assets What computer processing and storage capability does our organization need to handle its information and business transactions? What arrangement of computers
A Review of Data Mining Techniques
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 4, April 2014,
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
Hybrid Lossless Compression Method For Binary Images
M.F. TALU AND İ. TÜRKOĞLU/ IU-JEEE Vol. 11(2), (2011), 1399-1405 Hybrid Lossless Compression Method For Binary Images M. Fatih TALU, İbrahim TÜRKOĞLU Inonu University, Dept. of Computer Engineering, Engineering
Parallel Compression and Decompression of DNA Sequence Reads in FASTQ Format
, pp.91-100 http://dx.doi.org/10.14257/ijhit.2014.7.4.09 Parallel Compression and Decompression of DNA Sequence Reads in FASTQ Format Jingjing Zheng 1,* and Ting Wang 1, 2 1,* Parallel Software and Computational
RN-coding of Numbers: New Insights and Some Applications
RN-coding of Numbers: New Insights and Some Applications Peter Kornerup Dept. of Mathematics and Computer Science SDU, Odense, Denmark & Jean-Michel Muller LIP/Arénaire (CRNS-ENS Lyon-INRIA-UCBL) Lyon,
Protein Protein Interaction Networks
Functional Pattern Mining from Genome Scale Protein Protein Interaction Networks Young-Rae Cho, Ph.D. Assistant Professor Department of Computer Science Baylor University it My Definition of Bioinformatics
ANALYSIS OF RSA ALGORITHM USING GPU PROGRAMMING
ANALYSIS OF RSA ALGORITHM USING GPU PROGRAMMING Sonam Mahajan 1 and Maninder Singh 2 1 Department of Computer Science Engineering, Thapar University, Patiala, India 2 Department of Computer Science Engineering,
Redundant Wavelet Transform Based Image Super Resolution
Redundant Wavelet Transform Based Image Super Resolution Arti Sharma, Prof. Preety D Swami Department of Electronics &Telecommunication Samrat Ashok Technological Institute Vidisha Department of Electronics
CS 591.03 Introduction to Data Mining Instructor: Abdullah Mueen
CS 591.03 Introduction to Data Mining Instructor: Abdullah Mueen LECTURE 3: DATA TRANSFORMATION AND DIMENSIONALITY REDUCTION Chapter 3: Data Preprocessing Data Preprocessing: An Overview Data Quality Major
A Web Based Software for Synonymous Codon Usage Indices
International Journal of Information and Computation Technology. ISSN 0974-2239 Volume 3, Number 3 (2013), pp. 147-152 International Research Publications House http://www. irphouse.com /ijict.htm A Web
SGN-1158 Introduction to Signal Processing Test. Solutions
SGN-1158 Introduction to Signal Processing Test. Solutions 1. Convolve the function ( ) with itself and show that the Fourier transform of the result is the square of the Fourier transform of ( ). (Hints:
Domain Classification of Technical Terms Using the Web
Systems and Computers in Japan, Vol. 38, No. 14, 2007 Translated from Denshi Joho Tsushin Gakkai Ronbunshi, Vol. J89-D, No. 11, November 2006, pp. 2470 2482 Domain Classification of Technical Terms Using
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets
Methodology for Emulating Self Organizing Maps for Visualization of Large Datasets Macario O. Cordel II and Arnulfo P. Azcarraga College of Computer Studies *Corresponding Author: [email protected]
PAPR and Bandwidth Analysis of SISO-OFDM/WOFDM and MIMO OFDM/WOFDM (Wimax) for Multi-Path Fading Channels
PAPR and Bandwidth Analysis of SISO-OFDM/WOFDM and MIMO OFDM/WOFDM (Wimax) for Multi-Path Fading Channels Ahsan Adeel Lecturer COMSATS Institute of Information Technology Islamabad Raed A. Abd-Alhameed
2695 P a g e. IV Semester M.Tech (DCN) SJCIT Chickballapur Karnataka India
Integrity Preservation and Privacy Protection for Digital Medical Images M.Krishna Rani Dr.S.Bhargavi IV Semester M.Tech (DCN) SJCIT Chickballapur Karnataka India Abstract- In medical treatments, the integrity
Functional Data Analysis of MALDI TOF Protein Spectra
Functional Data Analysis of MALDI TOF Protein Spectra Dean Billheimer [email protected]. Department of Biostatistics Vanderbilt University Vanderbilt Ingram Cancer Center FDA for MALDI TOF
Information, Entropy, and Coding
Chapter 8 Information, Entropy, and Coding 8. The Need for Data Compression To motivate the material in this chapter, we first consider various data sources and some estimates for the amount of data associated
Sachin Dhawan Deptt. of ECE, UIET, Kurukshetra University, Kurukshetra, Haryana, India
Abstract Image compression is now essential for applications such as transmission and storage in data bases. In this paper we review and discuss about the image compression, need of compression, its principles,
International Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 3, May-Jun 2014
RESEARCH ARTICLE OPEN ACCESS A Survey of Data Mining: Concepts with Applications and its Future Scope Dr. Zubair Khan 1, Ashish Kumar 2, Sunny Kumar 3 M.Tech Research Scholar 2. Department of Computer
Master's projects at ITMO University. Daniil Chivilikhin PhD Student @ ITMO University
Master's projects at ITMO University Daniil Chivilikhin PhD Student @ ITMO University General information Guidance from our lab's researchers Publishable results 2 Research areas Research at ITMO Evolutionary
Data, Measurements, Features
Data, Measurements, Features Middle East Technical University Dep. of Computer Engineering 2009 compiled by V. Atalay What do you think of when someone says Data? We might abstract the idea that data are
IMPROVING PERFORMANCE OF RANDOMIZED SIGNATURE SORT USING HASHING AND BITWISE OPERATORS
Volume 2, No. 3, March 2011 Journal of Global Research in Computer Science RESEARCH PAPER Available Online at www.jgrcs.info IMPROVING PERFORMANCE OF RANDOMIZED SIGNATURE SORT USING HASHING AND BITWISE
SCAN-CA Based Image Security System
SCAN-CA Based Image Security System Bhagyashree.S.Anantwar 1, S.P.Sonavane 2 Student, Department of Computer Science and Engg, Walchand College of Engg, Sanli, India 1 Asso. Professor, Department of Information
Time series analysis Matlab tutorial. Joachim Gross
Time series analysis Matlab tutorial Joachim Gross Outline Terminology Sampling theorem Plotting Baseline correction Detrending Smoothing Filtering Decimation Remarks Focus on practical aspects, exercises,
A Fast Pattern Matching Algorithm with Two Sliding Windows (TSW)
Journal of Computer Science 4 (5): 393-401, 2008 ISSN 1549-3636 2008 Science Publications A Fast Pattern Matching Algorithm with Two Sliding Windows (TSW) Amjad Hudaib, Rola Al-Khalid, Dima Suleiman, Mariam
Barcode Based Automated Parking Management System
IJSRD - International Journal for Scientific Research & Development Vol. 2, Issue 03, 2014 ISSN (online): 2321-0613 Barcode Based Automated Parking Management System Parth Rajeshbhai Zalawadia 1 Jasmin
Binary search tree with SIMD bandwidth optimization using SSE
Binary search tree with SIMD bandwidth optimization using SSE Bowen Zhang, Xinwei Li 1.ABSTRACT In-memory tree structured index search is a fundamental database operation. Modern processors provide tremendous
Data Mining Solutions for the Business Environment
Database Systems Journal vol. IV, no. 4/2013 21 Data Mining Solutions for the Business Environment Ruxandra PETRE University of Economic Studies, Bucharest, Romania [email protected] Over
A Novel Method to Improve Resolution of Satellite Images Using DWT and Interpolation
A Novel Method to Improve Resolution of Satellite Images Using DWT and Interpolation S.VENKATA RAMANA ¹, S. NARAYANA REDDY ² M.Tech student, Department of ECE, SVU college of Engineering, Tirupati, 517502,
A Simultaneous Solution for General Linear Equations on a Ring or Hierarchical Cluster
Acta Technica Jaurinensis Vol. 3. No. 1. 010 A Simultaneous Solution for General Linear Equations on a Ring or Hierarchical Cluster G. Molnárka, N. Varjasi Széchenyi István University Győr, Hungary, H-906
Computers. Hardware. The Central Processing Unit (CPU) CMPT 125: Lecture 1: Understanding the Computer
Computers CMPT 125: Lecture 1: Understanding the Computer Tamara Smyth, [email protected] School of Computing Science, Simon Fraser University January 3, 2009 A computer performs 2 basic functions: 1.
Research on the UHF RFID Channel Coding Technology based on Simulink
Vol. 6, No. 7, 015 Research on the UHF RFID Channel Coding Technology based on Simulink Changzhi Wang Shanghai 0160, China Zhicai Shi* Shanghai 0160, China Dai Jian Shanghai 0160, China Li Meng Shanghai
Selection of Optimal Discount of Retail Assortments with Data Mining Approach
Available online at www.interscience.in Selection of Optimal Discount of Retail Assortments with Data Mining Approach Padmalatha Eddla, Ravinder Reddy, Mamatha Computer Science Department,CBIT, Gandipet,Hyderabad,A.P,India.
Data Storage. Chapter 3. Objectives. 3-1 Data Types. Data Inside the Computer. After studying this chapter, students should be able to:
Chapter 3 Data Storage Objectives After studying this chapter, students should be able to: List five different data types used in a computer. Describe how integers are stored in a computer. Describe how
Compression algorithm for Bayesian network modeling of binary systems
Compression algorithm for Bayesian network modeling of binary systems I. Tien & A. Der Kiureghian University of California, Berkeley ABSTRACT: A Bayesian network (BN) is a useful tool for analyzing the
A Time Efficient Algorithm for Web Log Analysis
A Time Efficient Algorithm for Web Log Analysis Santosh Shakya Anju Singh Divakar Singh Student [M.Tech.6 th sem (CSE)] Asst.Proff, Dept. of CSE BU HOD (CSE), BUIT, BUIT,BU Bhopal Barkatullah University,
The enhancement of the operating speed of the algorithm of adaptive compression of binary bitmap images
The enhancement of the operating speed of the algorithm of adaptive compression of binary bitmap images Borusyak A.V. Research Institute of Applied Mathematics and Cybernetics Lobachevsky Nizhni Novgorod
Complex Network Visualization based on Voronoi Diagram and Smoothed-particle Hydrodynamics
Complex Network Visualization based on Voronoi Diagram and Smoothed-particle Hydrodynamics Zhao Wenbin 1, Zhao Zhengxu 2 1 School of Instrument Science and Engineering, Southeast University, Nanjing, Jiangsu
Evaluation of Bitmap Index Compression using Data Pump in Oracle Database
IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661, p- ISSN: 2278-8727Volume 16, Issue 3, Ver. III (May-Jun. 2014), PP 43-48 Evaluation of Bitmap Index Compression using Data Pump in Oracle
Big Data with Rough Set Using Map- Reduce
Big Data with Rough Set Using Map- Reduce Mr.G.Lenin 1, Mr. A. Raj Ganesh 2, Mr. S. Vanarasan 3 Assistant Professor, Department of CSE, Podhigai College of Engineering & Technology, Tirupattur, Tamilnadu,
Web Document Clustering
Web Document Clustering Lab Project based on the MDL clustering suite http://www.cs.ccsu.edu/~markov/mdlclustering/ Zdravko Markov Computer Science Department Central Connecticut State University New Britain,
Automated Model Based Testing for an Web Applications
Automated Model Based Testing for an Web Applications Agasarpa Mounica, Lokanadham Naidu Vadlamudi Abstract- As the development of web applications plays a major role in our day-to-day life. Modeling the
