Image Compression Example redundant Information theory lossless lossy data information data redundancy
|
|
- Mervyn Neal
- 7 years ago
- Views:
Transcription
1 Image Compression Images usually require an enormous amount of data to represent: Example: A standard 8.5 by sheet of paper scanned at samples per inch (dpi) and quantized to two gray levels (binary image) would require more than k bytes to represent. ( 8.5 )( ) / bytes. Image compression involves reducing the amount of data (bits) required to represent a digital image. This is done by removing redundant information. Image compression is crucial for efficient storage, retrieval, transmission, and manipulation of images. More generally, data compression involves the efficient representation of data in digital format. Information theory --- a field pioneered by Claude E. Shannon in the 94s --- is the theoretical basis for most data compression techniques. Compression can be either lossless (information preserving) or lossy. In lossy compression, one can achieve higher levels of data reduction with less than perfect reproduction of the original image. Data compression refers to the process of reducing the amount of data required to represent a given quantity of information. Various amounts of data may be used to represent/describe the same information. Some representations maybe less efficient in the sense that they have data redundancy.
2 If n and n 2 are the number of information carrying units (ex. bits) in two datasets that represent the same information, the relative redundancy R D of the first dataset is defined as RD, where CR C R n n 2 is called the compression ratio. For images, data redundancy can be of three types: Coding redundancy: This refers to the binary codewords used to represent grayvalues. Interpixel redundancy: This refers to the correlation between adjacent pixels in an image. Psychovisual redundancy: This refers to the unequal sensitivity of the human eye to different visual information. A fidelity criterion or error criterion is required to quantify the loss of information (if any) due to a compression scheme. Objective fidelity criteria are based on some quantitive function of the original input image and the compressed and subsequently decompressed image. Example: Root-mean-square (RMS) error, which is defined as the square-root of the MSE. e( m, fˆ( m, f ( m, f ( m, : Original image f ˆ( m, : Reconstructed image e ( m, : Error image
3 e M N 2 [ fˆ( m, f ( m, n ] rms ) MN m n A related measure is the mean-square signal-to-noise ratio (SNR ms ): SNR ms [ fˆ( m, ] M N [ fˆ( m, f ( m, ] m n M N m n The rms value of SNR, denoted SNR rms, is the square-root of SNR rms. When the ultimate image is to be viewed by a human, subjective fidelity criterion may be more appropriate. Here, image quality is measured by subjective evaluations by a human observer. Ratings by a number of human observers, based on typical decompressed images, are averaged to obtain this subjective fidelity criterion
4 Example of an absolute comparison scale: Value Rating Description Excellent An image of extremely high quality --- as good as desired. 2 Fine An image of high quality, providing enjoyable viewing. Interference is not objectionable. 3 Passable An image of acceptable quality. Interference is not objectionable. 4 Marginal An image of poor quality; you wish you could improve it. Interference is somewhat objectionable. 5 Inferior A very poor image, but you could watch it. Objectionable interference is definitely present. 6 Unusable An image so bad that you cold not watch it. Example of a relative comparison scale, based on a side-by-side comparison of the original image and the decompressed image: Value Rating Much Worse Worse Slightly Worse Same Slightly better Better Much Better
5 Image Compression model f ( m, Source Encoder Channel Encoder Channel Channel Decoder Source Decoder fˆ ( m, Source Encoder is used to remove redundancy in the input image. Channel Encoder is used to introduce redundancy in a controlled fashion to help combat noise. Example: Parity bit. This provides a certain level of immunity from noise that is inherent in any storage/transmission system. If the channel is not prone to noise, this block maybe eliminated. The Channel could be a communication link or a storage/retrieval system. Channel Decoder and Source Decoder invert the operations of the corresponding encoder blocks. We will mainly concentrate on the source encoder/decoder blocks and not on the channel encoder/decoder steps.
6 Source Encoder and Decoder Source encoder is responsible for reducing or eliminating any coding, interpixel, or psychovisual redundancy. f ( m, Mapper Quantizer Source Encoder Symbol Encoder To Channel (Compressed Image) The first block Mapper transforms the input data into a (usually nonvisual) format, designed to reduce interpixel redundancy. This block is reversible and may or may not reduce the amount of data. Example: run-length encoding, image transform. The Quantizer reduces accuracy of the mapper output in accordance with some fidelity criterion. This block reduces psychovisual redundancy and is usually not invertible. The Symbol Encoder creates a fixed or variable length codeword to represent the quantizer output and maps the output in accordance with this code. This block is reversible and reduces coding redundancy. From Channel (Compressed Image) Symbol Decoder Inverse Mapper fˆ ( m, Source Decoder The decoder blocks are inverse operations of the corresponding encoder blocks (except the quantizer block, which is not invertible).
7 Elements of Information Theory What is information --- how to quantify it? What is the minimum amount of data that is sufficient to represent an image without loss of information? What is theoretically the best compression possible? What is the theoretically best possible transmission rate for reliable communication over a noisy channel? Information theory provides answers to these and other related fundamental questions. The fundamental premise of information theory is that the generation of information can be modeled as a probabilistic process. A discrete source of information generates one of N possible symbols from a source alphabet set A { a, a,, an }, in unit time. Example: A { a, b, c,, z}, {, }, {,, 2,, 255}. The source output can be modeled as a discrete random variable E, which can take values in set A { a, a,, an }, with p p,. corresponding probabilities { },, p N We will denote the symbol probabilities by the vector T z a ), P( a ), P( a ) p, p p. [ ] [ ] T P( N, N N Naturally, p, and. i p i i The information source is characterized by the pair ( A,z).
8 Observing an occurrence (or realizatio of the random variable E results in some gain of information denoted by I(E). This gain of information was defined to be (Shanno: I ( E) log log P( E) P( E) The base for the logarithm depends on the units for measuring information. Usually, we use base 2, which gives the information in units of binary digits or bits. Using a base logarithm would give the entropy in the units of decimal digits. The amount of information attributed to an event E is inversely related to the probability of that event. Examples: Certain event: P ( E).. In this case I ( E) log(/). This agrees with intuition, since if the event E is certain to occur (has probability ), knowing that it has occurred has not led to any gain of information. Coin toss: P ( E Heads). 5. In this case I( E) log(/.5) log(2) bit. This again agrees with intuition. Rare event: P ( E).. In this case I( E) log(/.) log() 9.97 bits. This again agrees with intuition, since knowing that a rare event has occurred leads to a significant gain of information. The entropy H(z) of a source is defined as the average amount of information gained by observing a single source symbol: H ( z) N i p i log p i
9 By convention, in the above formula, we set log. The entropy of a source quantifies the randomness of a source. Higher the source entropy, more the uncertainty associated with a source output, and higher the information associated with a source. For a fixed number of source symbols, the entropy is maximized if all the symbols are equally likely (recall uniform histogram). Example: Symbol ai Probability p i ½ ¼ 2 2 /8 3 3 /6 4 4 / / /64 6 Information (in bits) I ( a i ) log p i Source entropy: 6 H ( z) p i log p i i [ 2log 2 + 4log log 64] [ ] bits 32
10 Given that a source produces the above symbols with indicated probabilities, how do we represent them using binary strings? Symbol ai Probability Binary String Length of (Codeword) codeword l i ½ 3 ¼ 3 2 /8 3 3 /6 3 4 / / /64 3 Average length of a codeword: L pili pi (3) 3 pi i i i 3 bits Is this is the best we can do (in terms of L )? For a fixed length codeword scheme, yes. How about if we employ a variable length scheme? Idea: Since the symbols are not all equally likely, assign shorter codewords to symbols with higher probability and longer codewords to symbols with lower probability, such that the average length is smaller.
11 Consider the following scheme Symbol ai Probability p i Binary String (Codeword) ½ ¼ 2 2 /8 3 3 /6 4 4 / / /64 6 L 6 p l i i 32 i Length of codeword l i [ ] bit Notice that this is the same as the source entropy! Shannon s noiseless coding theorem Let (A,z) be a discrete source with probability vector z and entropy H(z). The average codeword length of any distortionless (uniquely decodable) coding is bounded by In other words, no codes exist that can losslessly represent the source if L < H( z). L H ( z)
12 Note that Shannon s theorem is quite general in that it refers to any code, not a particular coding scheme. Also, it does not specify a scheme to construct codes whose average length satisfies L H( z), nor does it claim that a code satisfying L H ( z) exists. Indeed, the Huffman code (to be studied later) is a particular algorithm for assigning codewords, with average codeword length satisfying H ( z) L < H ( z) + Using a block coding scheme (code a block of n symbols at a time instead of a single symbol), one can obtain codewords with average codeword length satisfying L H ( z) L n < H ( z) + n The efficiency of an encoding scheme is defined as H ( z) η L
Reading.. IMAGE COMPRESSION- I IMAGE COMPRESSION. Image compression. Data Redundancy. Lossy vs Lossless Compression. Chapter 8.
Reading.. IMAGE COMPRESSION- I Week VIII Feb 25 Chapter 8 Sections 8.1, 8.2 8.3 (selected topics) 8.4 (Huffman, run-length, loss-less predictive) 8.5 (lossy predictive, transform coding basics) 8.6 Image
More informationIntroduction to image coding
Introduction to image coding Image coding aims at reducing amount of data required for image representation, storage or transmission. This is achieved by removing redundant data from an image, i.e. by
More informationInformation, Entropy, and Coding
Chapter 8 Information, Entropy, and Coding 8. The Need for Data Compression To motivate the material in this chapter, we first consider various data sources and some estimates for the amount of data associated
More informationImage Compression through DCT and Huffman Coding Technique
International Journal of Current Engineering and Technology E-ISSN 2277 4106, P-ISSN 2347 5161 2015 INPRESSCO, All Rights Reserved Available at http://inpressco.com/category/ijcet Research Article Rahul
More informationEntropy and Mutual Information
ENCYCLOPEDIA OF COGNITIVE SCIENCE 2000 Macmillan Reference Ltd Information Theory information, entropy, communication, coding, bit, learning Ghahramani, Zoubin Zoubin Ghahramani University College London
More informationA NEW LOSSLESS METHOD OF IMAGE COMPRESSION AND DECOMPRESSION USING HUFFMAN CODING TECHNIQUES
A NEW LOSSLESS METHOD OF IMAGE COMPRESSION AND DECOMPRESSION USING HUFFMAN CODING TECHNIQUES 1 JAGADISH H. PUJAR, 2 LOHIT M. KADLASKAR 1 Faculty, Department of EEE, B V B College of Engg. & Tech., Hubli,
More informationChapter 1 Introduction
Chapter 1 Introduction 1. Shannon s Information Theory 2. Source Coding theorem 3. Channel Coding Theory 4. Information Capacity Theorem 5. Introduction to Error Control Coding Appendix A : Historical
More information2695 P a g e. IV Semester M.Tech (DCN) SJCIT Chickballapur Karnataka India
Integrity Preservation and Privacy Protection for Digital Medical Images M.Krishna Rani Dr.S.Bhargavi IV Semester M.Tech (DCN) SJCIT Chickballapur Karnataka India Abstract- In medical treatments, the integrity
More informationKhalid Sayood and Martin C. Rost Department of Electrical Engineering University of Nebraska
PROBLEM STATEMENT A ROBUST COMPRESSION SYSTEM FOR LOW BIT RATE TELEMETRY - TEST RESULTS WITH LUNAR DATA Khalid Sayood and Martin C. Rost Department of Electrical Engineering University of Nebraska The
More informationJPEG Image Compression by Using DCT
International Journal of Computer Sciences and Engineering Open Access Research Paper Volume-4, Issue-4 E-ISSN: 2347-2693 JPEG Image Compression by Using DCT Sarika P. Bagal 1* and Vishal B. Raskar 2 1*
More informationIntroduction to Medical Image Compression Using Wavelet Transform
National Taiwan University Graduate Institute of Communication Engineering Time Frequency Analysis and Wavelet Transform Term Paper Introduction to Medical Image Compression Using Wavelet Transform 李 自
More informationAn Introduction to Information Theory
An Introduction to Information Theory Carlton Downey November 12, 2013 INTRODUCTION Today s recitation will be an introduction to Information Theory Information theory studies the quantification of Information
More informationComparison of different image compression formats. ECE 533 Project Report Paula Aguilera
Comparison of different image compression formats ECE 533 Project Report Paula Aguilera Introduction: Images are very important documents nowadays; to work with them in some applications they need to be
More informationInformation Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture - 17 Shannon-Fano-Elias Coding and Introduction to Arithmetic Coding
More informationLinear Codes. Chapter 3. 3.1 Basics
Chapter 3 Linear Codes In order to define codes that we can encode and decode efficiently, we add more structure to the codespace. We shall be mainly interested in linear codes. A linear code of length
More informationELEC3028 Digital Transmission Overview & Information Theory. Example 1
Example. A source emits symbols i, i 6, in the BCD format with probabilities P( i ) as given in Table, at a rate R s = 9.6 kbaud (baud=symbol/second). State (i) the information rate and (ii) the data rate
More informationA deterministic fractal is an image which has low information content and no inherent scale.
FRACTAL IMAGE COMPRESSION: A RESOLUTION INDEPENDENT REPRESENTATION FOR IMAGER?? Alan D. Sloan 5550 Peachtree Parkway Iterated Systems, Inc. Norcross, Georgia 30092 1. Background A deterministic fractal
More informationData Storage 3.1. Foundations of Computer Science Cengage Learning
3 Data Storage 3.1 Foundations of Computer Science Cengage Learning Objectives After studying this chapter, the student should be able to: List five different data types used in a computer. Describe how
More informationPrivacy and Security in the Internet of Things: Theory and Practice. Bob Baxley; bob@bastille.io HitB; 28 May 2015
Privacy and Security in the Internet of Things: Theory and Practice Bob Baxley; bob@bastille.io HitB; 28 May 2015 Internet of Things (IoT) THE PROBLEM By 2020 50 BILLION DEVICES NO SECURITY! OSI Stack
More information1. (Ungraded) A noiseless 2-kHz channel is sampled every 5 ms. What is the maximum data rate?
Homework 2 Solution Guidelines CSC 401, Fall, 2011 1. (Ungraded) A noiseless 2-kHz channel is sampled every 5 ms. What is the maximum data rate? 1. In this problem, the channel being sampled gives us the
More informationCHAPTER 2 LITERATURE REVIEW
11 CHAPTER 2 LITERATURE REVIEW 2.1 INTRODUCTION Image compression is mainly used to reduce storage space, transmission time and bandwidth requirements. In the subsequent sections of this chapter, general
More informationOn the Use of Compression Algorithms for Network Traffic Classification
On the Use of for Network Traffic Classification Christian CALLEGARI Department of Information Ingeneering University of Pisa 23 September 2008 COST-TMA Meeting Samos, Greece Outline Outline 1 Introduction
More informationData Storage. Chapter 3. Objectives. 3-1 Data Types. Data Inside the Computer. After studying this chapter, students should be able to:
Chapter 3 Data Storage Objectives After studying this chapter, students should be able to: List five different data types used in a computer. Describe how integers are stored in a computer. Describe how
More informationVoice---is analog in character and moves in the form of waves. 3-important wave-characteristics:
Voice Transmission --Basic Concepts-- Voice---is analog in character and moves in the form of waves. 3-important wave-characteristics: Amplitude Frequency Phase Voice Digitization in the POTS Traditional
More informationTCOM 370 NOTES 99-4 BANDWIDTH, FREQUENCY RESPONSE, AND CAPACITY OF COMMUNICATION LINKS
TCOM 370 NOTES 99-4 BANDWIDTH, FREQUENCY RESPONSE, AND CAPACITY OF COMMUNICATION LINKS 1. Bandwidth: The bandwidth of a communication link, or in general any system, was loosely defined as the width of
More informationPHASE ESTIMATION ALGORITHM FOR FREQUENCY HOPPED BINARY PSK AND DPSK WAVEFORMS WITH SMALL NUMBER OF REFERENCE SYMBOLS
PHASE ESTIMATION ALGORITHM FOR FREQUENCY HOPPED BINARY PSK AND DPSK WAVEFORMS WITH SMALL NUM OF REFERENCE SYMBOLS Benjamin R. Wiederholt The MITRE Corporation Bedford, MA and Mario A. Blanco The MITRE
More informationFUNDAMENTALS of INFORMATION THEORY and CODING DESIGN
DISCRETE "ICS AND ITS APPLICATIONS Series Editor KENNETH H. ROSEN FUNDAMENTALS of INFORMATION THEORY and CODING DESIGN Roberto Togneri Christopher J.S. desilva CHAPMAN & HALL/CRC A CRC Press Company Boca
More informationConceptual Framework Strategies for Image Compression: A Review
International Journal of Computer Sciences and Engineering Open Access Review Paper Volume-4, Special Issue-1 E-ISSN: 2347-2693 Conceptual Framework Strategies for Image Compression: A Review Sumanta Lal
More informationMIMO CHANNEL CAPACITY
MIMO CHANNEL CAPACITY Ochi Laboratory Nguyen Dang Khoa (D1) 1 Contents Introduction Review of information theory Fixed MIMO channel Fading MIMO channel Summary and Conclusions 2 1. Introduction The use
More informationVideo compression: Performance of available codec software
Video compression: Performance of available codec software Introduction. Digital Video A digital video is a collection of images presented sequentially to produce the effect of continuous motion. It takes
More informationJPEG compression of monochrome 2D-barcode images using DCT coefficient distributions
Edith Cowan University Research Online ECU Publications Pre. JPEG compression of monochrome D-barcode images using DCT coefficient distributions Keng Teong Tan Hong Kong Baptist University Douglas Chai
More informationDigital Modulation. David Tipper. Department of Information Science and Telecommunications University of Pittsburgh. Typical Communication System
Digital Modulation David Tipper Associate Professor Department of Information Science and Telecommunications University of Pittsburgh http://www.tele.pitt.edu/tipper.html Typical Communication System Source
More informationSampling Theorem Notes. Recall: That a time sampled signal is like taking a snap shot or picture of signal periodically.
Sampling Theorem We will show that a band limited signal can be reconstructed exactly from its discrete time samples. Recall: That a time sampled signal is like taking a snap shot or picture of signal
More informationExample/ an analog signal f ( t) ) is sample by f s = 5000 Hz draw the sampling signal spectrum. Calculate min. sampling frequency.
1 2 3 4 Example/ an analog signal f ( t) = 1+ cos(4000πt ) is sample by f s = 5000 Hz draw the sampling signal spectrum. Calculate min. sampling frequency. Sol/ H(f) -7KHz -5KHz -3KHz -2KHz 0 2KHz 3KHz
More informationCoding and decoding with convolutional codes. The Viterbi Algor
Coding and decoding with convolutional codes. The Viterbi Algorithm. 8 Block codes: main ideas Principles st point of view: infinite length block code nd point of view: convolutions Some examples Repetition
More informationencoding compression encryption
encoding compression encryption ASCII utf-8 utf-16 zip mpeg jpeg AES RSA diffie-hellman Expressing characters... ASCII and Unicode, conventions of how characters are expressed in bits. ASCII (7 bits) -
More informationFigure 1: Relation between codec, data containers and compression algorithms.
Video Compression Djordje Mitrovic University of Edinburgh This document deals with the issues of video compression. The algorithm, which is used by the MPEG standards, will be elucidated upon in order
More informationHD Radio FM Transmission System Specifications Rev. F August 24, 2011
HD Radio FM Transmission System Specifications Rev. F August 24, 2011 SY_SSS_1026s TRADEMARKS HD Radio and the HD, HD Radio, and Arc logos are proprietary trademarks of ibiquity Digital Corporation. ibiquity,
More informationImage Compression and Decompression using Adaptive Interpolation
Image Compression and Decompression using Adaptive Interpolation SUNILBHOOSHAN 1,SHIPRASHARMA 2 Jaypee University of Information Technology 1 Electronicsand Communication EngineeringDepartment 2 ComputerScience
More informationGambling and Data Compression
Gambling and Data Compression Gambling. Horse Race Definition The wealth relative S(X) = b(x)o(x) is the factor by which the gambler s wealth grows if horse X wins the race, where b(x) is the fraction
More informationDigital Transmission of Analog Data: PCM and Delta Modulation
Digital Transmission of Analog Data: PCM and Delta Modulation Required reading: Garcia 3.3.2 and 3.3.3 CSE 323, Fall 200 Instructor: N. Vlajic Digital Transmission of Analog Data 2 Digitization process
More informationLossless Grey-scale Image Compression using Source Symbols Reduction and Huffman Coding
Lossless Grey-scale Image Compression using Source Symbols Reduction and Huffman Coding C. SARAVANAN cs@cc.nitdgp.ac.in Assistant Professor, Computer Centre, National Institute of Technology, Durgapur,WestBengal,
More informationA comprehensive survey on various ETC techniques for secure Data transmission
A comprehensive survey on various ETC techniques for secure Data transmission Shaikh Nasreen 1, Prof. Suchita Wankhade 2 1, 2 Department of Computer Engineering 1, 2 Trinity College of Engineering and
More informationCoded Bidirectional Relaying in Wireless Networks
Coded Bidirectional Relaying in Wireless Networks Petar Popovski and Toshiaki Koike - Akino Abstract The communication strategies for coded bidirectional (two way) relaying emerge as a result of successful
More informationComputer Networks and Internets, 5e Chapter 6 Information Sources and Signals. Introduction
Computer Networks and Internets, 5e Chapter 6 Information Sources and Signals Modified from the lecture slides of Lami Kaya (LKaya@ieee.org) for use CECS 474, Fall 2008. 2009 Pearson Education Inc., Upper
More informationEx. 2.1 (Davide Basilio Bartolini)
ECE 54: Elements of Information Theory, Fall 00 Homework Solutions Ex.. (Davide Basilio Bartolini) Text Coin Flips. A fair coin is flipped until the first head occurs. Let X denote the number of flips
More informationGenetic Algorithms commonly used selection, replacement, and variation operators Fernando Lobo University of Algarve
Genetic Algorithms commonly used selection, replacement, and variation operators Fernando Lobo University of Algarve Outline Selection methods Replacement methods Variation operators Selection Methods
More informationCompression techniques
Compression techniques David Bařina February 22, 2013 David Bařina Compression techniques February 22, 2013 1 / 37 Contents 1 Terminology 2 Simple techniques 3 Entropy coding 4 Dictionary methods 5 Conclusion
More informationLZ77. Example 2.10: Let T = badadadabaab and assume d max and l max are large. phrase b a d adadab aa b
LZ77 The original LZ77 algorithm works as follows: A phrase T j starting at a position i is encoded as a triple of the form distance, length, symbol. A triple d, l, s means that: T j = T [i...i + l] =
More informationANALYSIS OF THE EFFECTIVENESS IN IMAGE COMPRESSION FOR CLOUD STORAGE FOR VARIOUS IMAGE FORMATS
ANALYSIS OF THE EFFECTIVENESS IN IMAGE COMPRESSION FOR CLOUD STORAGE FOR VARIOUS IMAGE FORMATS Dasaradha Ramaiah K. 1 and T. Venugopal 2 1 IT Department, BVRIT, Hyderabad, India 2 CSE Department, JNTUH,
More informationTransform-domain Wyner-Ziv Codec for Video
Transform-domain Wyner-Ziv Codec for Video Anne Aaron, Shantanu Rane, Eric Setton, and Bernd Girod Information Systems Laboratory, Department of Electrical Engineering Stanford University 350 Serra Mall,
More informationLecture 2 Outline. EE 179, Lecture 2, Handout #3. Information representation. Communication system block diagrams. Analog versus digital systems
Lecture 2 Outline EE 179, Lecture 2, Handout #3 Information representation Communication system block diagrams Analog versus digital systems Performance metrics Data rate limits Next lecture: signals and
More informationLog-Likelihood Ratio-based Relay Selection Algorithm in Wireless Network
Recent Advances in Electrical Engineering and Electronic Devices Log-Likelihood Ratio-based Relay Selection Algorithm in Wireless Network Ahmed El-Mahdy and Ahmed Walid Faculty of Information Engineering
More informationTechnical Specifications for KD5HIO Software
Technical Specifications for KD5HIO Software Version 0.2 12/12/2000 by Glen Hansen, KD5HIO HamScope Forward Error Correction Algorithms HamScope is a terminal program designed to support multi-mode digital
More informationCatch Me If You Can: A Practical Framework to Evade Censorship in Information-Centric Networks
Catch Me If You Can: A Practical Framework to Evade Censorship in Information-Centric Networks Reza Tourani, Satyajayant (Jay) Misra, Joerg Kliewer, Scott Ortegel, Travis Mick Computer Science Department
More informationDiffusion and Data compression for data security. A.J. Han Vinck University of Duisburg/Essen April 2013 Vinck@iem.uni-due.de
Diffusion and Data compression for data security A.J. Han Vinck University of Duisburg/Essen April 203 Vinck@iem.uni-due.de content Why diffusion is important? Why data compression is important? Unicity
More informationStudy and Implementation of Video Compression Standards (H.264/AVC and Dirac)
Project Proposal Study and Implementation of Video Compression Standards (H.264/AVC and Dirac) Sumedha Phatak-1000731131- sumedha.phatak@mavs.uta.edu Objective: A study, implementation and comparison of
More informationAnalog-to-Digital Voice Encoding
Analog-to-Digital Voice Encoding Basic Voice Encoding: Converting Analog to Digital This topic describes the process of converting analog signals to digital signals. Digitizing Analog Signals 1. Sample
More informationOn the Many-to-One Transport Capacity of a Dense Wireless Sensor Network and the Compressibility of Its Data
On the Many-to-One Transport Capacity of a Dense Wireless Sensor etwork and the Compressibility of Its Data Daniel Marco, Enrique J. Duarte-Melo Mingyan Liu, David L. euhoff University of Michigan, Ann
More informationA Comparison of Speech Coding Algorithms ADPCM vs CELP. Shannon Wichman
A Comparison of Speech Coding Algorithms ADPCM vs CELP Shannon Wichman Department of Electrical Engineering The University of Texas at Dallas Fall 1999 December 8, 1999 1 Abstract Factors serving as constraints
More informationParametric Comparison of H.264 with Existing Video Standards
Parametric Comparison of H.264 with Existing Video Standards Sumit Bhardwaj Department of Electronics and Communication Engineering Amity School of Engineering, Noida, Uttar Pradesh,INDIA Jyoti Bhardwaj
More informationArithmetic Coding: Introduction
Data Compression Arithmetic coding Arithmetic Coding: Introduction Allows using fractional parts of bits!! Used in PPM, JPEG/MPEG (as option), Bzip More time costly than Huffman, but integer implementation
More informationLarge-Scale IP Traceback in High-Speed Internet
2004 IEEE Symposium on Security and Privacy Large-Scale IP Traceback in High-Speed Internet Jun (Jim) Xu Networking & Telecommunications Group College of Computing Georgia Institute of Technology (Joint
More informationTeaching Convolutional Coding using MATLAB in Communication Systems Course. Abstract
Section T3C2 Teaching Convolutional Coding using MATLAB in Communication Systems Course Davoud Arasteh Department of Electronic Engineering Technology, LA 70813, USA Abstract Convolutional codes are channel
More informationHIGH DENSITY DATA STORAGE IN DNA USING AN EFFICIENT MESSAGE ENCODING SCHEME Rahul Vishwakarma 1 and Newsha Amiri 2
HIGH DENSITY DATA STORAGE IN DNA USING AN EFFICIENT MESSAGE ENCODING SCHEME Rahul Vishwakarma 1 and Newsha Amiri 2 1 Tata Consultancy Services, India derahul@ieee.org 2 Bangalore University, India ABSTRACT
More informationDATA SECURITY USING PRIVATE KEY ENCRYPTION SYSTEM BASED ON ARITHMETIC CODING
DATA SECURITY USING PRIVATE KEY ENCRYPTION SYSTEM BASED ON ARITHMETIC CODING Ajit Singh 1 and Rimple Gilhotra 2 Department of Computer Science & Engineering and Information Technology BPS Mahila Vishwavidyalaya,
More informationHSI BASED COLOUR IMAGE EQUALIZATION USING ITERATIVE n th ROOT AND n th POWER
HSI BASED COLOUR IMAGE EQUALIZATION USING ITERATIVE n th ROOT AND n th POWER Gholamreza Anbarjafari icv Group, IMS Lab, Institute of Technology, University of Tartu, Tartu 50411, Estonia sjafari@ut.ee
More informationStreaming Lossless Data Compression Algorithm (SLDC)
Standard ECMA-321 June 2001 Standardizing Information and Communication Systems Streaming Lossless Data Compression Algorithm (SLDC) Phone: +41 22 849.60.00 - Fax: +41 22 849.60.01 - URL: http://www.ecma.ch
More informationQUANTIZED PRINCIPAL COMPONENT ANALYSIS WITH APPLICATIONS TO LOW-BANDWIDTH IMAGE COMPRESSION AND COMMUNICATION. D. Wooden. M. Egerstedt. B.K.
International Journal of Innovative Computing, Information and Control ICIC International c 5 ISSN 1349-4198 Volume x, Number x, x 5 pp. QUANTIZED PRINCIPAL COMPONENT ANALYSIS WITH APPLICATIONS TO LOW-BANDWIDTH
More informationDepartment of Electrical and Computer Engineering Ben-Gurion University of the Negev. LAB 1 - Introduction to USRP
Department of Electrical and Computer Engineering Ben-Gurion University of the Negev LAB 1 - Introduction to USRP - 1-1 Introduction In this lab you will use software reconfigurable RF hardware from National
More informationAn introduction to OBJECTIVE ASSESSMENT OF IMAGE QUALITY. Harrison H. Barrett University of Arizona Tucson, AZ
An introduction to OBJECTIVE ASSESSMENT OF IMAGE QUALITY Harrison H. Barrett University of Arizona Tucson, AZ Outline! Approaches to image quality! Why not fidelity?! Basic premises of the task-based approach!
More informationThe Degrees of Freedom of Compute-and-Forward
The Degrees of Freedom of Compute-and-Forward Urs Niesen Jointly with Phil Whiting Bell Labs, Alcatel-Lucent Problem Setting m 1 Encoder m 2 Encoder K transmitters, messages m 1,...,m K, power constraint
More informationIntroduction to Learning & Decision Trees
Artificial Intelligence: Representation and Problem Solving 5-38 April 0, 2007 Introduction to Learning & Decision Trees Learning and Decision Trees to learning What is learning? - more than just memorizing
More informationVideo Coding Basics. Yao Wang Polytechnic University, Brooklyn, NY11201 yao@vision.poly.edu
Video Coding Basics Yao Wang Polytechnic University, Brooklyn, NY11201 yao@vision.poly.edu Outline Motivation for video coding Basic ideas in video coding Block diagram of a typical video codec Different
More informationBroadband Networks. Prof. Dr. Abhay Karandikar. Electrical Engineering Department. Indian Institute of Technology, Bombay. Lecture - 29.
Broadband Networks Prof. Dr. Abhay Karandikar Electrical Engineering Department Indian Institute of Technology, Bombay Lecture - 29 Voice over IP So, today we will discuss about voice over IP and internet
More informationplc numbers - 13.1 Encoded values; BCD and ASCII Error detection; parity, gray code and checksums
plc numbers - 3. Topics: Number bases; binary, octal, decimal, hexadecimal Binary calculations; s compliments, addition, subtraction and Boolean operations Encoded values; BCD and ASCII Error detection;
More informationBasics of information theory and information complexity
Basics of information theory and information complexity a tutorial Mark Braverman Princeton University June 1, 2013 1 Part I: Information theory Information theory, in its modern format was introduced
More informationNumerical Matrix Analysis
Numerical Matrix Analysis Lecture Notes #10 Conditioning and / Peter Blomgren, blomgren.peter@gmail.com Department of Mathematics and Statistics Dynamical Systems Group Computational Sciences Research
More informationLecture 13 - Basic Number Theory.
Lecture 13 - Basic Number Theory. Boaz Barak March 22, 2010 Divisibility and primes Unless mentioned otherwise throughout this lecture all numbers are non-negative integers. We say that A divides B, denoted
More informationTowards a Tight Finite Key Analysis for BB84
The Uncertainty Relation for Smooth Entropies joint work with Charles Ci Wen Lim, Nicolas Gisin and Renato Renner Institute for Theoretical Physics, ETH Zurich Group of Applied Physics, University of Geneva
More informationComplexity-bounded Power Control in Video Transmission over a CDMA Wireless Network
Complexity-bounded Power Control in Video Transmission over a CDMA Wireless Network Xiaoan Lu, David Goodman, Yao Wang, and Elza Erkip Electrical and Computer Engineering, Polytechnic University, Brooklyn,
More informationFast Arithmetic Coding (FastAC) Implementations
Fast Arithmetic Coding (FastAC) Implementations Amir Said 1 Introduction This document describes our fast implementations of arithmetic coding, which achieve optimal compression and higher throughput by
More informationT = 1 f. Phase. Measure of relative position in time within a single period of a signal For a periodic signal f(t), phase is fractional part t p
Data Transmission Concepts and terminology Transmission terminology Transmission from transmitter to receiver goes over some transmission medium using electromagnetic waves Guided media. Waves are guided
More informationTape Drive Data Compression Q & A
Tape Drive Data Compression Q & A Question What is data compression and how does compression work? Data compression permits increased storage capacities by using a mathematical algorithm that reduces redundant
More informationWan Accelerators: Optimizing Network Traffic with Compression. Bartosz Agas, Marvin Germar & Christopher Tran
Wan Accelerators: Optimizing Network Traffic with Compression Bartosz Agas, Marvin Germar & Christopher Tran Introduction A WAN accelerator is an appliance that can maximize the services of a point-to-point(ptp)
More informationSteganographyinaVideoConferencingSystem? AndreasWestfeld1andGrittaWolf2 2InstituteforOperatingSystems,DatabasesandComputerNetworks 1InstituteforTheoreticalComputerScience DresdenUniversityofTechnology
More informationExperiment #1, Analyze Data using Excel, Calculator and Graphs.
Physics 182 - Fall 2014 - Experiment #1 1 Experiment #1, Analyze Data using Excel, Calculator and Graphs. 1 Purpose (5 Points, Including Title. Points apply to your lab report.) Before we start measuring
More informationFor Articulation Purpose Only
E305 Digital Audio and Video (4 Modular Credits) This document addresses the content related abilities, with reference to the module. Abilities of thinking, learning, problem solving, team work, communication,
More informationSachin Dhawan Deptt. of ECE, UIET, Kurukshetra University, Kurukshetra, Haryana, India
Abstract Image compression is now essential for applications such as transmission and storage in data bases. In this paper we review and discuss about the image compression, need of compression, its principles,
More informationProbability Interval Partitioning Entropy Codes
SUBMITTED TO IEEE TRANSACTIONS ON INFORMATION THEORY 1 Probability Interval Partitioning Entropy Codes Detlev Marpe, Senior Member, IEEE, Heiko Schwarz, and Thomas Wiegand, Senior Member, IEEE Abstract
More informationSTUDY OF MUTUAL INFORMATION IN PERCEPTUAL CODING WITH APPLICATION FOR LOW BIT-RATE COMPRESSION
STUDY OF MUTUAL INFORMATION IN PERCEPTUAL CODING WITH APPLICATION FOR LOW BIT-RATE COMPRESSION Adiel Ben-Shalom, Michael Werman School of Computer Science Hebrew University Jerusalem, Israel. {chopin,werman}@cs.huji.ac.il
More informationSecured Lossless Medical Image Compression Based On Adaptive Binary Optimization
IOSR Journal of Computer Engineering (IOSR-JCE) e-issn: 2278-0661, p- ISSN: 2278-8727Volume 16, Issue 2, Ver. IV (Mar-Apr. 2014), PP 43-47 Secured Lossless Medical Image Compression Based On Adaptive Binary
More informationStatistical Modeling of Huffman Tables Coding
Statistical Modeling of Huffman Tables Coding S. Battiato 1, C. Bosco 1, A. Bruna 2, G. Di Blasi 1, G.Gallo 1 1 D.M.I. University of Catania - Viale A. Doria 6, 95125, Catania, Italy {battiato, bosco,
More informationhttp://www.springer.com/0-387-23402-0
http://www.springer.com/0-387-23402-0 Chapter 2 VISUAL DATA FORMATS 1. Image and Video Data Digital visual data is usually organised in rectangular arrays denoted as frames, the elements of these arrays
More informationPCM Encoding and Decoding:
PCM Encoding and Decoding: Aim: Introduction to PCM encoding and decoding. Introduction: PCM Encoding: The input to the PCM ENCODER module is an analog message. This must be constrained to a defined bandwidth
More informationCHAPTER 3: DIGITAL IMAGING IN DIAGNOSTIC RADIOLOGY. 3.1 Basic Concepts of Digital Imaging
Physics of Medical X-Ray Imaging (1) Chapter 3 CHAPTER 3: DIGITAL IMAGING IN DIAGNOSTIC RADIOLOGY 3.1 Basic Concepts of Digital Imaging Unlike conventional radiography that generates images on film through
More informationThe Effect of Network Cabling on Bit Error Rate Performance. By Paul Kish NORDX/CDT
The Effect of Network Cabling on Bit Error Rate Performance By Paul Kish NORDX/CDT Table of Contents Introduction... 2 Probability of Causing Errors... 3 Noise Sources Contributing to Errors... 4 Bit Error
More informationDiscrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 13. Random Variables: Distribution and Expectation
CS 70 Discrete Mathematics and Probability Theory Fall 2009 Satish Rao, David Tse Note 3 Random Variables: Distribution and Expectation Random Variables Question: The homeworks of 20 students are collected
More informationTechnical Information. Digital Signals. 1 bit. Part 1 Fundamentals
Technical Information Digital Signals 1 1 bit Part 1 Fundamentals t Technical Information Part 1: Fundamentals Part 2: Self-operated Regulators Part 3: Control Valves Part 4: Communication Part 5: Building
More informationEE4367 Telecom. Switching & Transmission. Prof. Murat Torlak
Path Loss Radio Wave Propagation The wireless radio channel puts fundamental limitations to the performance of wireless communications systems Radio channels are extremely random, and are not easily analyzed
More information