Low Quality Image Retrieval System For Generic Databases

Similar documents
Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall

Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches

How To Filter Spam Image From A Picture By Color Or Color

ISSN: A Review: Image Retrieval Using Web Multimedia Mining

The Scientific Data Mining Process

International Journal of Advanced Information in Arts, Science & Management Vol.2, No.2, December 2014

Circle Object Recognition Based on Monocular Vision for Home Security Robot

A Dynamic Approach to Extract Texts and Captions from Videos

Simultaneous Gamma Correction and Registration in the Frequency Domain

Knowledge Discovery from patents using KMX Text Analytics

EHR CURATION FOR MEDICAL MINING

Image Processing Based Automatic Visual Inspection System for PCBs

Colour Image Segmentation Technique for Screen Printing

ANALYSIS OF THE EFFECTIVENESS IN IMAGE COMPRESSION FOR CLOUD STORAGE FOR VARIOUS IMAGE FORMATS


ENHANCED WEB IMAGE RE-RANKING USING SEMANTIC SIGNATURES

A Simple Feature Extraction Technique of a Pattern By Hopfield Network

Image Compression through DCT and Huffman Coding Technique

Open issues and research trends in Content-based Image Retrieval

Face detection is a process of localizing and extracting the face region from the

Data Storage 3.1. Foundations of Computer Science Cengage Learning

Handwritten Character Recognition from Bank Cheque

Comparison of different image compression formats. ECE 533 Project Report Paula Aguilera

JPEG Image Compression by Using DCT

JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions

A Method of Caption Detection in News Video

Use of Data Mining Techniques to Improve the Effectiveness of Sales and Marketing

Natural Language Querying for Content Based Image Retrieval System

Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 269 Class Project Report

How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm

Clustering Data Streams

Robust Outlier Detection Technique in Data Mining: A Univariate Approach

Comparision of k-means and k-medoids Clustering Algorithms for Big Data Using MapReduce Techniques

An Automatic and Accurate Segmentation for High Resolution Satellite Image S.Saumya 1, D.V.Jiji Thanka Ligoshia 2

Document Image Retrieval using Signatures as Queries

Visibility optimization for data visualization: A Survey of Issues and Techniques

International Journal of Advanced Computer Technology (IJACT) ISSN: PRIVACY PRESERVING DATA MINING IN HEALTH CARE APPLICATIONS

BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES

CS Introduction to Data Mining Instructor: Abdullah Mueen

An Experimental Study of the Performance of Histogram Equalization for Image Enhancement

Environmental Remote Sensing GEOG 2021

Classifying Large Data Sets Using SVMs with Hierarchical Clusters. Presented by :Limou Wang

Big Data: Image & Video Analytics

The Delicate Art of Flower Classification

Classifying Manipulation Primitives from Visual Data

Content-Based Image Retrieval

Industrial Challenges for Content-Based Image Retrieval

An Ants Algorithm to Improve Energy Efficient Based on Secure Autonomous Routing in WSN

Saving Mobile Battery Over Cloud Using Image Processing

Face Recognition For Remote Database Backup System

Palmprint Recognition. By Sree Rama Murthy kora Praveen Verma Yashwant Kashyap

An Introduction to Data Mining. Big Data World. Related Fields and Disciplines. What is Data Mining? 2/12/2015

Printed Circuit Board Defect Detection using Wavelet Transform

SIGNATURE VERIFICATION

A Genetic Algorithm-Evolved 3D Point Cloud Descriptor

MusicGuide: Album Reviews on the Go Serdar Sali

Calculation of Minimum Distances. Minimum Distance to Means. Σi i = 1

Automatic License Plate Recognition using Python and OpenCV

Using Data Mining for Mobile Communication Clustering and Characterization

Automatic Recognition Algorithm of Quick Response Code Based on Embedded System

Active Learning SVM for Blogs recommendation

Data Mining in Web Search Engine Optimization and User Assisted Rank Results

Digital Image Fundamentals. Selim Aksoy Department of Computer Engineering Bilkent University

HSI BASED COLOUR IMAGE EQUALIZATION USING ITERATIVE n th ROOT AND n th POWER

Support Vector Machines with Clustering for Training with Very Large Datasets

Social Media Mining. Data Mining Essentials

Colorado School of Mines Computer Vision Professor William Hoff

SOURCE SCANNER IDENTIFICATION FOR SCANNED DOCUMENTS. Nitin Khanna and Edward J. Delp

DYNAMIC DOMAIN CLASSIFICATION FOR FRACTAL IMAGE COMPRESSION

Distance Metric Learning in Data Mining (Part I) Fei Wang and Jimeng Sun IBM TJ Watson Research Center

Object Recognition and Template Matching

SVM Based License Plate Recognition System

FACE RECOGNITION BASED ATTENDANCE MARKING SYSTEM

Volume 2, Issue 9, September 2014 International Journal of Advance Research in Computer Science and Management Studies

Extend Table Lens for High-Dimensional Data Visualization and Classification Mining

Comparison of K-means and Backpropagation Data Mining Algorithms

Introduction to Pattern Recognition

The Role of Size Normalization on the Recognition Rate of Handwritten Numerals

Domain Classification of Technical Terms Using the Web

Bisecting K-Means for Clustering Web Log data

A Study of Automatic License Plate Recognition Algorithms and Techniques

TouchPaper - An Augmented Reality Application with Cloud-Based Image Recognition Service

Filtering Noisy Contents in Online Social Network by using Rule Based Filtering System

Determining optimal window size for texture feature extraction methods

Implementation of OCR Based on Template Matching and Integrating it in Android Application

Applied Mathematical Sciences, Vol. 7, 2013, no. 112, HIKARI Ltd,

Automatic Extraction of Signatures from Bank Cheques and other Documents

International Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 3, May-Jun 2014

Combating Anti-forensics of Jpeg Compression

Analecta Vol. 8, No. 2 ISSN

ScienceDirect. Brain Image Classification using Learning Machine Approach and Brain Structure Analysis

Data Storage. Chapter 3. Objectives. 3-1 Data Types. Data Inside the Computer. After studying this chapter, students should be able to:

Clustering Big Data. Anil K. Jain. (with Radha Chitta and Rong Jin) Department of Computer Science Michigan State University November 29, 2012

Image Search by MapReduce

Clustering Technique in Data Mining for Text Documents

CHAPTER 2 LITERATURE REVIEW

Cross-Domain Collaborative Recommendation in a Cold-Start Context: The Impact of User Profile Size on the Quality of Recommendation

K-means Clustering Technique on Search Engine Dataset using Data Mining Tool

Module II: Multimedia Data Mining

Euler Vector: A Combinatorial Signature for Gray-Tone Images

Transcription:

Low Quality Image Retrieval System For Generic Databases W.A.D.N. Wijesekera, W.M.J.I. Wijayanayake Abstract: Content Based Image Retrieval (CBIR) systems have become the trend in image retrieval technologies, as the index or notation based image retrieval algorithms give less efficient results in high usage of images. These CBIR systems are mostly developed considering the availability of high or normal quality images. High availability of low quality images in databases, due to usage of different quality equipment to capture images, and different environmental conditions the photos are being captured, has opened up a new path in image retrieval research area. The algorithms which are developed for low quality image based image retrieval are only a few, and have been performed only for specific domains. Low quality image based image retrieval algorithm on a generic database with a considerable accuracy level for different industries is an area which remains unsolved. Through this study, an algorithm has been developed to achieve above mentioned gaps. By using images with inappropriate brightness and compressed images as low quality images, the proposed algorithm is tested on a generic database, which includes many categories of data, instead of using a specific domain. The new algorithm developed, gives better precision and recall values when they are clustered into the most appropriate number of clusters which changes according to the level of quality of the image. As the quality of the image decreases, the accuracy of the algorithm also tends to be reduced; a space for further improvement. Index Terms: Content Based Image Retrieval, Low Quality Images, K means Clustering Technique, Query Based Image Content, Image Binarization 1 INTRODUCTION WITH the current developments in the Information Technology the use of multimedia data has grown up drastically. Although in the past, storing of multimedia data which occupy higher memory capacities was an issue to use them, this problem has been solved with newer technologies. Among multimedia data, images are the most frequently added and updated media in the World Wide Web. With the advent of digital cameras the rate of uploading photos into web has reached the clouds. According to the web statistics, Facebook gets approximately 208,300 photos uploaded every minute. Instagram gets 27.800 photos a minute and Flickr receives about 8 million photos a month as in [13]. Apart from these, in almost every other industry, images play a vital role. For an example, in health sector medical imaging is a critical factor in treating the patients. In robotics, intelligent systems, remote sensing and Forensics, image processing is a pivotal factor. Transportation is also an area where image processing has become of importance with recent automobile developments such as automatically driven vehicles with path planning, obstacle avoidance and servo control. In military applications for object detection, tracking, and three dimensional reconstruction of territory etc. image processing is essential. Although the images created in modern world are normally of high quality, there are still many areas and industries that have to deal with low quality images. Some of the low quality image types are fax images, scanned images, low resolution images, noisy images and highly compressed images as in [1]. Low quality images are highly created through fax machines, image compressing, low resolution and noisy images or photos captured by different resolution profiles in digital cameras. Images which have inappropriate contrast and brightness and quantized images also are considered as of low quality. Although fax is technology that has been overridden by many newer findings such as email, cloud services, Google Docs etc. the usage of fax machines are still high. Among thousands of photos and images that are being uploaded to the web databases, there is a significant amount of compressed images as well. Compressed images are created to minimize the download time and make the web sites attractive for users, in the aspect of their efficiency as in [2]. The use of low resolution photos taken from digital cameras is also a considerable fact in this study. It is stated that many people use low resolution photo profiles, in order to make them consume low memory and efficient in uploading and downloading. The traditional image indexing methods have many problems in accuracy and efficiency, through which the need of image retrieval using image content, such as color, shape and texture aroused. This method is generally known as Content Based Image Retrieval (CBIR) as in [3]. CBIR systems are moving into commercialization through new findings such as Query Based Image Content (QBIC). However there are no efficient algorithms to retrieve low quality images from a generic database which includes low quality images itself. 2 RELATED WORK The researches that have been done in this area have addressed specific areas of the usage of low quality images. One research has been performed using a museum database, which is queried by fax images that are of low quality as in [1]. But this method doesn t provide for low resolution and noisy images well. As this is only performed for a museum database, the numbers of images that are to be queried against are considerably fewer than of a generic database and also it consists of only domain specific query images. The method also can be further improved on accuracy and efficiency aspects as well. Another research has been done on this area using scanned and hand painted images as low quality images. This is an efficient method done for databases even having nearly 20,000 images. This algorithm has used multiresolution image querying technique and wavelet decomposition [4], to develop the metric [5]. This method also has many areas to be improved, such as huge dependability on the size of the image database, not allowing general pattern matching of a small query such as an icon or a company logo. Usage of CBIR depends highly on the image domain that s been used, according to [1]. The two types of image domains are narrow domains and broad domains. In narrow domains, the variability among the images used in the systems is relatively lower. Examples for narrow domains are fingerprints, retina scanning systems, face recognition systems, blueprints, etc. Broad image domains are polysemic and also their semantics are not fully described, due to their 67

use in distinct aspects in different scenes. The easiest example for a broad domain is the internet, where there are images from all domains used in all possible scenes. The image retrieval architecture comprises with mainly three levels as the retrieval engine, mutidimentional indexing level and the image database and feature vector database. The image database contains the raw images, which are used to identify the features, and these extracted features are stored in the feature vector database. Using the text-annotation database and this feature vector database, the image retrieval is performed. But in CBIR, the text-annotation database is not always used, unless the system needs to be more efficient. Chakravarti and Xiannong [3] describe color characteristics of an image as the most intuitive information that can be extracted from images for comparison. The most basic form of color retrieval technique is specifying the color values of the images in the database. Another simple method is color query to associate spatial conditions to the color information. Color histogram is the most popular color based retrieval methodology used in CBIR systems, relying on the fact that images are represented as a series of pixel values, each corresponding to a specific color. IR based on color concepts must first accurately retrieve images despite of the manipulations in size, orientation and position of an image. Color histogram is an efficient and easy method to use as well. A limitation of this method is not being able to identify spatial characteristics of color in the comparison when using images with more information rather than being a simple array of pixels. Black and white images also are not well compared through algorithms that compare images solely based on color concept. Kannan and Anbazhagan [6] state that although there have many CBIR systems been developed, only few of them have been successfully commercialized. In texture based approach, the parameters are gathered on the basis of statistical approach. In classifying the texture, using statistical features of grey levels is one of the most efficient methods that are being used in commercial applications. The Grey Level Co-occurence Matrix (GLCM) can be used to extract second order statistics from an image, such as entropy, contrast, standard deviation, mean and variance which are very successful in texture feature calculation in both query image and target images. Based on these calculated values, the similar images can be retrieved. Texture is a low level feature that is used in image retrieval systems. According to [7], texture is an important feature in defining high level semantics in image retrieval as texture provides important information about the content of the real world images for image classification. Among many features, the mostly used texture features are Gabor features and wavelet features. Both Gabor filtering and wavelet transforms are designed originally for rectangular images. But in Region Based Image Retrieval (RBIR) system, there are arbitrary shapes rather than rectangular shapes. It is solved by extending the arbitrary shaped region into a rectangular shape, by padding some values outside the boundary. Vadivelet. al. [8] mention that the most simple and efficient way of texture feature extraction is Haar Wavelets. Focusing on the use of texture features along with color features on a weighted average basis provides a further efficient retrieval process. Although Haar wavelets are efficient, it also has a drawback for lossy compression. It is that, Haar wavelets tend to produce blocky image artifacts for high compression rates. For texture feature extraction, Daubechie s Wavelet Transform is also used. Although Haar Wavelets don t have sufficient sharp transition and hence become not able to separate different frequency bands properly, Daubechie s Wavelet Transforms have better frequency resolution properties, resulting from their longer filter lengths. Using texture and color features and weighted feature vectors in the retrieval process, [8] have developed a better performing system. Silakari et al. [9] defines images clustering as a data mining technique to group a set of unsupervised data based on the conceptual clustering concept; maximizing the intra-class similarity and minimizing the inter-class similarity. In Content Based Image Retrieval systems, the efficiency of retrieving process can be enhanced by integrating retrieval based low-level features with grouping images into meaningful categories. Clustering helps to reveal important information and patterns of images more effectively and efficiently. The clustering process consists of two main steps; feature extraction and grouping. In the first step the images in the database feature vectors are created and stored in a feature base and the clustering algorithm is applied over this feature base. The clustering algorithm used most frequently in the image retrieval area is k-means algorithm, which is one of the simplest unsupervised learning algorithms. According to [10], binarization process gives a great reduction in the complexity of the image retrieval process which is a very important fact in the commercial use of them. Image binarization can be defined as the conversion of a grey-scale image to a binary image. The main method of image binarization is through thresholding which is an adaptive method. Otsu s method [10] is a non-adaptive method, although it also is a powerful tool. Sauvola s method [12] is adaptive, but their parameters are hard to explain. Wang et al. [11] presents a novel algorithm which integrates color clustering with binary texture analysis. Binary texture analysis is performed on candidate binary images. Two kinds of texture features, run-length and spatial size distribution related are extracted and explored. Optimal candidate of best binarization effect can be obtained by cooperating this with LDA classifier. 3 METHODOLOGY Figure 1: Methodology of the research 3.1 Cluster Selection Clustering is used in this new algorithm in order to improve the retrieval efficiency. Clustering acts as a preprocessing step in this algorithm, as efficient indexing relies on clustering in this study. As low quality images take more time to process and the accuracy could be lower than normal and high quality image retrieval, it could be supported by clustering to make it more efficient and accurate. The most commonly used clustering mechanism is k-means clustering in image retrieval systems. It s a simple method which can be applied to systems, and get effective results. If a set of n data points are in a d- dimensional space, the problem is to determine a set of integer k points in that space, which are called as centers. They are calculated as to minimize the mean squared distance 68

from each data point to its nearest center. This mean squared distance is called the squared-error distortion. As k-means algorithm is a general variance based clustering, this study could use this method appropriately. 3.2 Compute Threshold Value Using an appropriate threshold, the binary image of the query image will be made similar to relevant database images. A simple approach to compute this threshold value is to use the percentage of black and white pixels of the binary query image. But the percentage of white pixels is needed to compute the feature vector of database images, and it depends on the query image. It s ineffective to compute feature vectors of database images at the retrieval time as it increases the response time. 3.3 Conversion to Binary Image The low-quality images differ substantially from its original. Therefore applying feature vector extraction to the original will not produce a feature vector close enough to the low-quality image. Exploiting the fact that low-quality images are almost binary in nature, before extracting features, the image will be converted to a binary image. This same conversion is done to the database images as well. But as there are also low quality color images, to convert those color images to binary images, a binarization method will be used. 3.4 Feature Vector Extraction In this proposed algorithm, feature extraction is done by a simple method, which provides the efficiency and considerably accurate results as well. In this algorithm, first, the image is normalized into a 300*300 pixel image. Then this image will be divided into 25 RGB triples, which will be used as feature vectors for image comparison step. As this method focuses on comparing similar regions of images, these 60*60 pixel image regions are considered as the extracted feature vectors, which will be used for comparison. Using these image regions, RGB color averages, mean and variance will be calculated, for each region. 3.5 Compare the Feature Vectors Previously calculated 60*60 image feature vectors are used in this step for image comparison. To compare the feature vectors, a distance classifier will be used to compute the similarity between query feature vector and database image vector. The distance classifier used in this algorithm is the minimum Euclidean Distance. The database that is used for this study contains 1000 images and the images can be clustered into several identifiable categories. Low quality images to query this database are created by taking random images from each of the identifiable 10 categories in this database and by making them of low quality. These images will be used to evaluate the new algorithm through precision and recall measures. To test and compare the results with the previous works, the same number of images that were used to query will be used in this study too. The amount is 30 low quality images, to test the retrieval efficiency and accuracy. This study mainly focuses on testing the image retrieval process through two types of low quality images. They are compressed images and brightened (high brightness) images. Thresholding of Low Quality Images In this experiment, to test the image database against low quality images, the need of defining the low quality images must be fulfilled first. This aspect has not been considered or mentioned in previous literature, as all those studies have been done for images which are generally accepted as of low quality than normal images, such as fax images, hand drawn paintings etc. Therefore in this study, the selected types of low quality images are thresholded in a customized way. The first image type is images with high brightness. These images are created by making them of 50% more brightened and 50% less brightened than the normal quality images in the database. This is done by adjusting the alpha and beta factors and converting the image. Alpha factor is the optional scale factor and beta factor is the optional delta value added to the scaled values. The second type of low quality images are, images which are highly compressed. Compression is done in many commercial applications for different purposes in different scales. In this study, the selected threshold for low quality images is quality factor 0.05f and 0.025 (50% and 75% compressed from the normal quality image). This measure and images compressed lesser than this amount will be addressed by this new algorithm. Selection of Test Query Images The image database that is used in this study has 1000 normal quality images. These can be categorized under 10 different titles. They are beach, buildings, busses, dinosaurs, elephants, flowers, food, horses, mountains and people. Each of these categories contains 100 images in the database. To test the database on each of the categories, from each category three images were randomly picked. The previous systems for low quality images were tested using only one kind of images in only one domain. But in this study, using a generic database, the algorithm is tested for a range of images. Using a random number generator, 30 images were chosen to test the system. As this should be performed twice for two types of low quality images, 60 retrievals were conducted in the testing process. Clustering the Images In the proposed algorithm, as mentioned in Methodology section, K means algorithm was used to cluster the images, prior to the images comparison component. In K means algorithm, selecting the number of clusters is the first task. This must be given by the user to the system. In this experimental design, the tests were performed using arrange of K values to identify patterns of test results according to the number of clusters used. Beginning from K=3, up to K=7, the images were clustered and tested. Analyzing the test results, until the accuracy of the algorithm shows no further increase, the cluster numbers were increased and tested. Measurements of the Experiment The experiment was conducted to measure the accuracy of the low quality image retrieval system than the existing algorithm. To measure this, the 60 low quality query images were used to test the database. Those 60 images were added to the image database and tested database consisted of 1060 images; 1000 of normal quality and 60 of low quality. The accuracy of the algorithm was measured using two well-known measures in image retrieval mechanisms. These measures have been used in previous works conducted by other 69

researchers as well. They are, 1. Precision = No. of similar images retrieved/ No. of images retrieved Figure 2: Changes in the accuracy measures over distance thresholds 2. Recall = No. of similar images retrieved/total no. of similar images in the DB For one cluster value entered by the user above set of values were recorded and the experiment was repeated for different number of clusters. The retrieval accuracy is calculated over several aspects as below. 1. No. of clusters 2. Image category 3. Low quality image type To validate the new algorithm, its accuracy should be compared with the old algorithm and shown an improvement. For this task, the precision and recall measures were also calculated for the system without image clustering phase as well. In every aspect stated above, the existing algorithm was also assessed. Validation of System Accuracy The algorithm developed for low quality image based image retrieval system, with image clustering component was tested against the existing algorithm, to prove its improved accuracy. First, the existing algorithm, which was retrieving images without the clustering phase, was used to test with the same low quality images, and the results were recorded to calculate precision and recall values. Then using the same query images and the database, the new algorithm was tested for its accuracy. To validate the outputs of the two algorithms, the method used in this study was manually counting the number of correct outputs for a given image in the database and check it against the outputs from the two algorithms. To calculate the recall values, this manually counted similar image numbers were taken into consideration. Thresholding the Distance Value in Retrieval of Images First, the distance between database image feature vectors and query image feature vector were calculated. Then, the feature vectors which show the minimum distance values are retrieved as the similar images. For this, after the distance calculation, the system should define a threshold, to retrieve images, which are below that specific value. Choosing this threshold value also is an experimental process. First, the two types of low quality images were created and they were used to query the database, setting up different threshold values. The results were checked and the best precision and recall value providing threshold distance was taken to do further experiments and compare with the existing algorithm. Following figure shows the results of precision and recall values for different threshold values. Through the above results, it can be seen that optimum threshold value which gives both precision and recall at a considerably good rate is 700, and it was used as the threshold distance value. 4 EXPERIMENT AND ANALYSIS The existing algorithm for low quality image based image retrieval system, only includes the binarizing, feature vector extraction and comparison phases. To improve the accuracy of this system, a new algorithm was developed by this study, and it has an added image clustering component. By this phase, it makes it easier and accurate to extract the feature vectors. Although the efficiency of the system may go down because of the time cost at the clustering phase, the accuracy improvement was the main objective of this study. With the added clustering phase, a sub experiment was conducted to find the optimum number of clusters that images should be clustered to obtain best results for the developed algorithm. Through this experiment, it was concluded that, the optimal number of clusters for clustering images with inappropriate brightness, is 7 or above, as it provides both precision and recall values at an optimum. For compressed images, it can be concluded that, the optimal number of clusters that provide best recall and precision values for compressed images is K = 6, because both the values show their highest points at K=6. When analyzing the experimental results, another point that is derived is that, through adding the clustering phase, although the precision has improved, recall value has stayed the same, or gone down. By selecting a higher threshold value for retrieving similar images, the recall could be increased. But it would also decrease the precision at the same time. Depending on the application of this algorithm, the decision on whether to improve precision or recall can be taken. For the existing algorithm, which doesn t include the clustering phase, the results were as follows, for images with inappropriate brightness. Table 1: Results of the Existing algorithm for Images with Inappropriate Brightness Image Type Precision Value Recall Value Images with Inappropriate Brightness 95.833% 86.666% 70

Table 2: Results for Images with Inappropriate Brightness No. of Clusters Precision Values Recall Values with with New algorithm New algorithm 3 100% 82.217% 4 99.17% 83.332% 5 99.17% 84.449% 6 99.17% 85.554% 7 99.17% 86.665% 8 99.17% 86.665% For the existing algorithm, the results were as follows, for compressed images. Table 3: Results of the Existing algorithm for Images with Inappropriate Brightness Image Type Precision Value Recall Value Compressed Images 88.786% 70.887% Table 4: Results for Compressed Images No. of Clusters Precision Values Recall Values with with New algorithm New algorithm 3 92.611% 81.11% 4 94.167% 86.665% 5 96.94% 85.56% 6 98.055% 86.665% 7 95.833% 86.666% 8 96.389% 86.666% When comparing the results for different image categories, that are available in the database and in query images, the accuracy measures show different patterns. For images with inappropriate brightness, all the image categories, except for one category, show 100% precision measures. Although the image categories include noise in different levels in different categories, in this situation, the retrieval of images is shown to be very accurate. As testing the new algorithm for different levels of compression and brightness, the system works well for 50% and 75% compression levels, although there is a drop in recall values in both compression levels. The precision value increase is lesser in 75% compression level than the 50% compression level. The system performance goes down, when the compression level increases. For images with inappropriate brightness, both the precision and recall values increase for 50% lower brightness images. Images which have 50% high brightness, only show the same recall value for the new algorithm, although the precision values is increased. Therefore it could be concluded that the system works well for low brightened images, than for high brightness images. By comparing the recall values of different image categories in both types of low quality images; images with inappropriate brightness and compressed images, there s no pattern that could be identified as to which categories show good precision and recall values specifically. But with further studies and more testing rounds, patterns among different image categories responding to the developed algorithm, could be identified, such as, the color usage in the images, different textures present, noise of the image, availability of objects in the image etc. 5 CONCLUSION Through the experiment results, it could be concluded that for images with inappropriate brightness, the precision value has increased from 95.833% to 99.17%. But the recall value hasn t changed. The maximum recall value that the new algorithm was able to provide was as the same as the existing algorithm. For compressed images, the precision values was increased from 88.786% to an optimum value of 98.055%, when the number of clusters is 6. For this type of images, the recall value was also improved from 70.887% to 86.666%. 6 LIMITATIONS OF THE STUDY As the new algorithm improves the precision measure of the image retrieval process, the reduction or the uniformity of recall measure is the main limitation of this study. As mentioned above in this chapter, the recall values can also be improved using different methods. But they have their effects on other measure too. For different image categories, the recall values are different, where for several types of images, the values are low. The selection of number of clusters for different image categories is also a limitation. As per the time limitations, a thorough analysis on image types and their optimal numbers of clusters were not correctly identified. Those values could be used in the applications of this algorithm, to improve the accuracy more, than using the same amount of clusters for every type of image. Although to increase the accuracy of the image retrieval system, a clustering phase is included in this study, it consumes more time. Therefore the efficiency of the system tends to decrease. Computers with more processing power can be used to minimize this effect on the system. But if the number of clusters each image is clustered into increases, for accuracy improvement purposes, the efficiency could be affected by that as well. 7 FUTURE WORK Both the efficiency and the accuracy of this algorithm can be improved by using a well-known algorithm for feature extraction of the images. Although the algorithm proposed in this method, doesn t take the texture features of images into consideration, by using Wavelet transform or any other appropriate algorithm, will minimize this error. To further reduce noise in images, so that the system could focus on the main object of the query images, a special pre-processing algorithm to detect plain background can be used. Focusing on the main object in the image would increase the recall value significantly, as it allows the system to identify most similar images and also images with rotations and flips as well. To increase the efficiency of the retrieval process, the images in the database can be indexed, using a suitable indexing method. As this algorithm is used for low quality image retrieval, indexing can be of help to increase the accuracy as well. REFERENCES [1] Fauzi, M. F. A., Content-based image retrieval of museum images. PhD diss., University of Southampton, 2004.W.-K. Chen, Linear Networks and Systems. Belmont, Calif.: Wadsworth, pp. 123-135, 1993. (Book style) [2] Purba, S., ed. High-performance Web databases: design, development, and deployment. CRC Press, 2010.K. Elissa, An Overview of Decision Theory,"unpublished. 71

(Unplublished manuscript) [3] Chakravarti, R., and Xiannong, M., A Study of Color Histogram Based Image Retrieval. In ITNG, pp. 1323-1328. 2009.C. J. Kaufman, Rocky Mountain Research Laboratories, Boulder, Colo., personal communication, 1992. (Personal communication) [4] Treil, N., Mallat, S., and Bajcsy, R., Image Wavelet Decomposition and Applications. Technical Reports (CIS) : 781. 1989.S.P. Bingulac, On the Compatibility of Adaptive Controllers, Proc. Fourth Ann. Allerton Conf. Circuits and Systems Theory, pp. 8-16, 1994. (Conference proceedings) [5] Jacobs, C. E., Finkelstein, A. and Salesin, D. H., Fast multiresolution image querying. In Proceedings of the 22nd annual conference on Computer graphics and interactive techniques, pp. 277-286. ACM, New York, NY, USA, 1995.J. Williams, Narrow-Band Analyzer, PhD dissertation, Dept. of Electrical Eng., Harvard Univ., Cambridge, Mass., 1993. (Thesis or dissertation) [6] Kannan, A., Mohan, V. and Anbazhagan, N., Image clustering and retrieval using image mining techniques. In 2010 IEEE International Conference on Computational Intelligence and Computing Research, Tamilnadu, India, December 2010.L. Hubert and P. Arabie, Comparing Partitions, J. Classification, vol. 2, no. 4, pp. 193-218, Apr. 1985. (Journal or magazine citation) [7] Liu, Y., Zhang, D., Lu, G. and Ma, W.Y. A survey of contentbased image retrieval with high-level semantics. Pattern Recognition 40, no. 1: 262-282. 2007. [8] Vadivel, A., Majumdar, A.K. and Sural, S., Characteristics of weighted feature vector in content-based image retrieval applications. In Intelligent Sensing and Information Processing, 2004. Proceedings of International Conference on, pp. 127-132. IEEE, 2004. [9] Silakari, S., Motwani, M. and Maheshwari, M. Color image clustering using block truncation algorithm. arxiv preprint arxiv:0910.1849. 2009. [10] Moghaddam, R.F.., and Cheriet, M. AdOtsu: An adaptive and parameterless generalization of Otsu's method for document image binarization. Pattern Recognition 45, no. 6: 2419-2431. 2012. [11] Wang, B., Li, X-F., Liu, F. and Hu, F-Q. Color text image binarization based on binary texture analysis. Pattern Recognition Letters 26, no. 11: 1650-1657. 2005. [12] J.SauvolaandM.Pietikainen.Adaptivedocumentimagebinarizati on.,pattern Recognition, 33(2):225{236,February2000. [13] STAN HORACZEK. 2013. How Many Photos Are Uploaded to The Internet Every Minute?. [ONLINE] Available at: http://www.popphoto.com/news/2013/05/how-many-photosare-uploaded-to-internet-every-minute. [Accessed 18 April 14]. 72