Comparing Images Using Color Coherence Vectors
|
|
|
- Shannon Evans
- 9 years ago
- Views:
Transcription
1 Comparing Images Using Color Coherence Vectors Greg Pass Ramin Zabih Justin Miller Computer Science Department Cornell University Ithaca, NY Abstract Color histograms are used to compare images in many applications. Their advantages are efficiency, and insensitivity to small changes in camera viewpoint. However, color histograms lack spatial information, so images with very different appearances can have similar histograms. For example, a picture of fall foliage might contain a large number of scattered red pixels; this could have a similar color histogram to a picture with a single large red object. We describe a histogram-based method for comparing images that incorporates spatial information. We classify each pixel in a given color bucket as either coherent or incoherent, based on whether or not it is part of a large similarly-colored region. A color coherence vector (CCV) stores the number of coherent versus incoherent pixels with each color. By separating coherent pixels from incoherent pixels, CCV s provide finer distinctions than color histograms. CCV s can be computed at over 5 images per second on a standard workstation. A database with 15,000 images can be queried for the images with the most similar CCV s in under 2 seconds. We show that CCV s can give superior results to color his- To whom correspondence should be addressed tograms for image retrieval. KEYWORDS: Content-based Image Retrieval, Processing, Color Histograms INTRODUCTION Many applications require simple methods for comparing pairs of images based on their overall appearance. For example, a user may wish to retrieve all images similar to a given image from a large database of images. Color histograms are a popular solution to this problem, and are used in systems like QBIC [4] and Chabot [11]. Color histograms are computationally efficient, and generally insensitive to small changes in camera position. Color histograms also have some limitations. A color histogram provides no spatial information; it merely describes which colors are present in the image, and in what quantities. In addition, color histograms are sensitive to both compression artifacts and changes in overall image brightness. In this paper we describe a color-based method for comparing images which is similar to color histograms, but which also takes spatial information into account. We begin with a 1
2 review of color histograms. We then describe color coherence vectors (CCV s) and how to compare them. Examples of CCV-based image queries demonstrate that they can give superior results to color histograms. We contrast our method with some recent algorithms [8, 14, 15, 17] that also combine spatial information with color histograms. Finally, we present some possible extensions to CCV s. COLOR HISTOGRAMS Color histograms are frequently used to compare images. Examples of their use in multimedia applications include scene break detection [1, 7, 10, 12, 22] and querying a database of images [3, 4, 11, 13]. Their popularity stems from several factors. Color histograms are computationally trivial to compute. Small changes in camera viewpoint tend not to effect color histograms. Different objects often have distinctive color histograms. Researchers in computer vision have also investigated color histograms. For example, Swain and Ballard [19] describe the use of color histograms for identifying objects. Hafner et al. [6] provide an efficient method for weighted-distance indexing of color histograms. Stricker and Swain [18] analyze the information capacity of color histograms, as well as their sensitivity. Definitions We will assume that all images are scaled to contain the same number of pixels M. We discretize the colorspace of the image such that there are n distinct (discretized) colors. A color histogram H is a vector h 1,h 2,...,h n, in which each bucket h j contains the number of pixels of color j in the image. Typically images are represented in the RGB colorspace, and a few of the most significant bits are used from each color channel. For example, Zhang [22] uses the 2 most significant bits of each color channel, for a total of n = 64 buckets in the histogram. For a given image I, the color histogram H I is a compact summary of the image. A database of images can be queried to find the most similar image to I, and can return the image I with the most similar color histogram H I. Typically color histograms are compared using the sum of squared differences (L 2 -distance) or the sum of absolute value of differences (L 1 -distance). So the most similar image to I would be the image I minimizing H I H I = for the L 2 -distance, or H I H I = n (H I [j] H I [j]) 2, j=1 n H I [j] H I [j], j=1 for the L 1 -distance. Note that we are assuming that differences are weighted evenly across different color buckets for simplicity. COLOR VECTORS COHERENCE Intuitively, we define a color s coherence as the degree to which pixels of that color are members of large similarly-colored regions. We refer to these significant regions as coherent regions, and observe that they are of significant importance in characterizing images. 2
3 Figure 1: Two images with similar color histograms For example, the images shown in figure 1 have similar color histograms, despite their rather different appearances. 1 The color red appears in both images in approximately the same quantities. In the left image the red pixels (from the flowers) are widely scattered, while in the right image the red pixels (from the golfer s shirt) form a single coherent region. Our coherence measure classifies pixels as either coherent or incoherent. Coherent pixels are a part of some sizable contiguous region, while incoherent pixels are not. A color coherence vector represents this classification for each color in the image. CCV s prevent coherent pixels in one image from matching incoherent pixels in another. This allows fine distinctions that cannot be made with color histograms. Computing CCV s The initial stage in computing a CCV is similar to the computation of a color histogram. We first blur the image slightly by replacing 1 The color images used in this paper can be found at pixel values with the average value in a small local neighborhood (currently including the 8 adjacent pixels). This eliminates small variations between neighboring pixels. We then discretize the colorspace, such that there are only n distinct colors in the image. The next step is to classify the pixels within a given color bucket as either coherent or incoherent. A coherent pixel is part of a large group of pixels of the same color, while an incoherent pixel is not. We determine the pixel groups by computing connected components. A connected component C is a maximal set of pixels such that for any two pixels p, p C, there is a path in C between p and p. (Formally, a path in C is a sequence of pixels p = p 1,p 2,...,p n = p such that each pixel p i is in C and any two sequential pixels p i,p i+1 are adjacent to each other. We consider two pixels to be adjacent if one pixel is among the eight closest neighbors of the other; in other words, we include diagonal neighbors.) Note that we only compute connected components within a given discretized color bucket. This effectively segments the image based on the discretized colorspace. Connected components can be computed in 3
4 linear time (see, for example, [20]). When this is complete, each pixel will belong to exactly one connected component. We classify pixels as either coherent or incoherent depending on the size in pixels of its connected component. A pixel is coherent if the size of its connected component exceeds a fixed value τ; otherwise, the pixel is incoherent. For a given discretized color, some of the pixels with that color will be coherent and some will be incoherent. Let us call the number of coherent pixels of the j th discretized color α j and the number of incoherent pixels β j. Clearly, the total number of pixels with that color is α j + β j, and so a color histogram would summarize an image as α 1 + β 1,...,α n +β n. Instead, for each color we compute the pair (α j,β j ) which we will call the coherence pair for the j th color. The color coherence vector for the image consists of (α 1,β 1 ),...,(α n,β n ). This is a vector of coherence pairs, one for each discretized color. In our experiments, all images were scaled to contain M =38,976 pixels, and we have used τ = 300 pixels (so a region is classified as coherent if its area is about 1% of the image). With this value of τ, an average image in our database consists of approximately 75% coherent pixels. An example CCV We next demonstrate the computation of a CCV. To keep our example small, we will let τ = 4 and assume that we are dealing with an image in which all 3 color components have the same value at every pixel (in the RGB colorspace this would represent a grayscale image). This allows us to represent a pixel s color with a single number (i.e., the pixel with R/G/B values 12/12/12 will be written as 12). Suppose that after we slightly blur the input image, the resulting intensities are as follows Let us discretize the colorspace so that bucket 1 contains intensities 10 through 19, bucket 2 contains 20 through 29, etc. Then after discretization we obtain The next step is to compute the connected components. Individual components will be labeled with letters (A,B,...) and we will need to keep a table which maintains the discretized color associated with each label, along with the number of pixels with that label. Of course, the same discretized color can be associated with different labels if multiple contiguous regions of the same color exist. The image may then become B C B B A A B B C B A A B C D B A A B B B A A E B B A A E E B B A A E E and the connected components table will be 4
5 Label A B C D E Color Size The components A, B, andehave more than τ pixels, and the components C and D less than τ pixels. Therefore the pixels in A, B and E are classified as coherent, while the pixels in C and D are classified as incoherent. The CCV for this image will be Color α β A given color bucket may thus contain only coherent pixels (as does 2), only incoherent pixels (as does 3), or a mixture of coherent and incoherent pixels (as does 1). If we assume there are only 3 possible discretized colors, the CCV can also be written (17, 3), (15, 0), (0, 1). Comparing CCV s Consider two images I and I, together with their CCV s G I and G I, and let the number of coherent pixels in color bucket j be α j (for I) andα j (for I ). Similarly, let the number of incoherent pixels be β j and β j.so G I = (α 1,β 1 ),...,(α n,β n ) and G I = (α 1,β 1),...,(α n,β n) Color histograms will compute the difference between I and I as n H = (α j + β j ) (α j + β j). (1) j=1 Our method for comparing is based on the quantity n G = (α j α j) + (β j β j). (2) j=1 From equations 1 and 2, it follows that CCV s create a finer distinction than color histograms. A given color bucket j can contain the same number of pixels in I as in I, i.e. α j + β j = α j + β j, but these pixels may be entirely coherent in I and entirely incoherent in I. In this case β j = α j = 0, and while H =0, G will be large. In general, H G. This is true even if we use squared differences instead of absolute differences in the definitions of H and G. This is because both d(x) = x and d(x) =x 2 are metrics, so they satisfy the triangle inequality d(x + y) d(x)+d(y). (3) If we rearrange the terms in equation 1 we get n H = (α j α j)+(β j β j). j=1 Applying the triangle inequality we have n (α j α j) + (β j β j) = G. H j=1 EXPERIMENTAL RESULTS We have implemented color coherence vectors, and have used them for image retrieval from a large database. Our database consists of 14,554 images, which are drawn from a variety of sources. Our largest sources include the 11,667 images used in Chabot [11], the 1,440 images used in QBIC [4], and a 1,005 image database available from Corel. In addition, we included a few groups of images in PhotoCD format. Finally, we have taken a number of MPEG videos from the Web and segmented 5
6 them into scenes using the method described in [21]. We have added one or two images from each scene to the database, totaling 349 images. The image database thus contains a wide variety of imagery. In addition, some of the imagery has undergone substantial lossy compression via JPEG or MPEG. We have compared our results with a number of color histogram variations. These include the L 1 and L 2 distances, with both 64 and 512 color buckets. In addition, we have implemented Swain s opponent axis colorspace [19], and used the color discretization scheme he describes. In each case we include a small amount of smoothing as it improves the performance of color histograms. On our database, the L 1 distance with the 64-bucket RGB colorspace gave the best results, and is used as a benchmark against CCV s. Hand examination of our database revealed 52 pairs of images which contain different views of the same scene. Examples are shown in figures 3 and 4. One image is selected as a query image, and the other represents a correct answer. In each case, we have shown where the second image ranks, when similarity is computed using color histograms versus CCV s. The color images shown are available at In 46 of the 52 cases, CCV s produced better results, while in 6 cases they produced worse results. The average change in rank due to CCV s was an improvement of just over 75 positions (note that this included the 6 cases where CCV s do worse). The average percentage change in rank was an improvement of 30%. In the 46 cases where CCV s performed better than color histograms, the average improvement in rank was 88 positions, and the average percentage improvement was 50%. In the 6 cases where color histograms performed better than CCV s, the average rank improvement was 21 positions, and the average percentage improvement was 48%. A histogram of the change in rank obtained by using CCV s is shown in figure 2. When CCV s produced worse results, it was always due to a change in overall image brightness (i.e., the two images were almost identical, except that one was brighter than the other). Because CCV s use discretized color buckets for segmentation, they are more sensitive to changes in overall image brightness than color histograms. We believe that this difficulty can be overcome by using a better colorspace than RGB, as we discuss in the extensions section of this paper. Efficiency There are two phases to the computation involved in querying an image database. First, when an image is inserted into the database, a CCV must be computed. Second, when the database is queried, some number of the most similar images must be retrieved. Most methods for content-based indexing include these distinct phases. For both color histograms and CCV s, these phases can be implemented in linear time. We ran our experiments on a 50 MHz SPARCstation 20, and provide the results from color histogramming for comparison. Color histograms can be computed at 67 images per second, while CCV s can be computed at 5 images per second. Using color histograms, 21,940 comparisons can be performed per second, while with CCV s 7,746 can be performed per second. The images used for benchmarking are Both implementations are preliminary, and the performance can definitely be improved. 6
7 Figure 2: Change in rank due to CCV s. Positive numbers indicate improved performance. RELATED WORK Recently, several authors have proposed algorithms for comparing images that combine spatial information with color histograms. Hsu et al. [8] attempts to capture the spatial arrangement of the different colors in the image, in order to perform more accurate content-based image retrieval. Rickman and Stonham [14] randomly sample the endpoints of small triangles and compare the distributions of these triplets. Smith and Chang [15] concentrate on queries that combine spatial information with color. Stricker and Dimai [17] divide the image into five partially overlapping regions and compute the first three moments of the color distributions in each image. We will discuss each approach in turn. Hsu [8] begins by selecting a set of representative colors from the image. Next, the image is partitioned into rectangular regions, where each region is predominantly a single color. The partitioning algorithm makes use of maximum entropy. The similarity between two images is the degree of overlap between regions of the same color. Hsu presents results from querying a database with 260 images, which show that the integrated approach can give better results than color histograms. While the authors do not report running times, it appears that Hsu s method requires substantially more computation than the approach we describe. A CCV can be computed in a single pass over the image, with a small number of operations per pixel. Hsu s partitioning algorithm in particular appears much more computationally intensive than our method. Hsu s approach can be extended to be independent of orientation and position, but the computation involved is quite substantial. In contrast, our method is naturally invariant to orientation and position. Rickman and Stonham [14] randomly sample pixel triples arranged in an equilateral tri- 7
8 Histogram rank: 50. CCV rank: 26. Histogram rank: 35. CCV rank: 9. Histogram rank: 367. CCV rank: 244. Histogram rank: 128. CCV rank: 32. Figure 3: Example queries with their partner images 8
9 Histogram rank: 310. CCV rank: 205. Histogram rank: 88. CCV rank: 38. Histogram rank: 13. CCV rank: 5. Histogram rank: 119. CCV rank: 43. Figure 4: Additional examples 9
10 angle with a fixed side length. They use 16 levels of color hue, with non-uniform quantization. Approximately a quarter of the pixels are selected for sampling, and their method stores 372 bits per image. They report results from a database of 100 images. Smith and Chang s algorithm also partitions the image into regions, but their approach is more elaborate than Hsu s. They allow a region to contain multiple different colors, and permit a given pixel to belong to several different regions. Their computation makes use of histogram back-projection [19] to back-project sets of colors onto the image. They then identify color sets with large connected components. Smith and Chang s image database contains 3,100 images. Again, running times are not reported, although their algorithm does speed up back-projection queries by pre-computing the back-projections of certain color sets. Their algorithm can also handle certain kinds of queries that our work does not address; for example, they can find all the images where the sun is setting in the upper left part of the image. Stricker and Dimai [17] compute moments for each channel in the HSV colorspace, where pixels close to the border have less weight. They store 45 floating point numbers per image. Their distance measure for two regions is a weighted sum of the differences in each of the three moments. The distance measure for a pair of images is the sum of the distance between the center regions, plus (for each of the 4 side regions) the minimum distance of that region to the corresponding region in the other image, when rotated by 0, 90, 180 or 270 degrees. Because the regions overlap, their method is insensitive to small rotations or translations. Because they explicitly handle rotations of 0, 90, 180 or 270 degrees, their method is not affected by these particular rotations. Their database contains over 11,000 images, but the performance of their method is only illustrated on 3 example queries. Like Smith and Chang, their method is designed to handle certain kinds of more complex queries that we do not consider. EXTENSIONS Thereareanumberofwaysinwhichouralgorithm could be extended and improved. One extension involves generalizing CCV s, while another centers on improving the choice of colorspace. Histogram refinement and effective feature extraction Our approach can be generalized to features other than color and coherence. We are investigating histogram refinement, in which the pixels of a given color bucket are subdivided based on particular features (coherence, texture, etc.). CCV s can be viewed as a simple form of histogram refinement, in which histogram buckets are split in two based on coherence. Histogram refinement also permits further distinctions to be imposed upon a CCV. In the same way that we distinguish between pixels of similar color by coherence, for example, we can distinguish between pixels of similar coherence by some additional feature. We can apply this method repeatedly; each split imposes an additional constraint on what it means for two pixels to be similar. If the initial histogram is a color histogram, and it is only refined once based on coherence, then the resulting refined histogram is a CCV. But there is no requirement that the initial histogram be based on color, or that the initial split be based on coherence. Consider the example shown in figure 5. We begin with a color histogram and divide the 10
11 Color Red Blue Green Coherence Incoherent Adjacency Coherent Edge Information Red Blue Green Edge Adjacency Not an edge Red Blue Green Figure 5: An example of histogram refinement. 11
12 pixels in each bucket based on coherence (i.e., generate a CCV). The incoherent pixels can be refined by tallying the colors of pixels adjacent to the incoherent pixels in a particular bin. The coherent pixels, in turn, can be further refined into those that lie on edges and those which do not. The coherent pixels on edges can then be further refined by recording what colors the coherent regions are adjacent to. The leaves of the tree, therefore, represent those images that satisfy these successive constraints. This set of refinements defines a compact summary of the image, and remains an efficient structure to compute and compare. The best system of constraints to impose on the image is an open issue. Any combination of features might give effective results, and there are many possible features to choose from. However, it is possible to take advantage of the temporal structure of a successively refined histogram. One feature might serve as a filter for another feature, by ensuring that the second feature is only computed on pixels which already possess the first feature. For example, the perimeter-to-area ratio can be used to classify the relative shapes of color regions. If we used this ratio as an initial refinement on color histograms, incoherent pixels would result in statistical outliers, and thus give questionable results. This feature is better employed after the coherent pixels have been segregated. We call the choice of a sensible structure of constraints effective feature extraction. Refining a histogram not only makes finer distinctions between pixels, but functions as a statistical filter for successive refinements. Choice of colorspace Many systems based on color histograms spend considerable effort on selecting a good set of colors. Hsu [8], for example, assumes that the colors in the center of the image are more important than those at the periphery, while Smith and Chang [15] use several different thresholds to extract colors and regions. A wide variety of different colorspaces have also been investigated. For example, Swain and Ballard [19] make use of the opponent-axis colorspace, while QBIC [4] uses the Munsell colorspace. The choice of colorspace is a particularly significant issue for CCV s, since they use the discretized color buckets to segment the image. A perceptually uniform colorspace, such as CIE Lab, should result in better segmentations and improve the performance of CCV s. A related issue is the color constancy problem [9], which causes objects of the same color to appear rather differently depending upon the lighting conditions. The simplest effect of color constancy is a change in overall image brightness, which is responsible for the negative examples obtained in our experiments. Standard histogramming methods are sensitive to image gain. More sophisticated methods, such as cumulative histograms [16] or color ratio histograms [5], might alleviate this problem. In fact, most proposed improvements to color histograms can also be applied to CCV s. This includes improvements beyond the selection of better colorspaces. For example, some authors [17] suggest that color moments be used in lieu of histograms. Color moments could be computed separately for coherent and incoherent pixels. CONCLUSIONS We have described a new method for comparing pairs of images that combines color histograms with spatial information. Most research in content-based image retrieval has focused on query by example (where the system automatically finds images similar to an input image). However, other types of queries are also important. For example, it is often useful 12
13 to search for images in which a subset of another image (e.g. a particular object) appears. This would be particularly useful for queries on a database of videos. One approach to this problem might be to generalize histogram back-projection [19] to separate pixels based on spatial coherence. It is clear that larger and larger image databases will demand more complex similarity measures. This added time complexity can be offset by using efficient, coarse measures that prune the search space by removing images which are clearly not the desired answer. Measures which are less efficient but more effective can then be applied to the remaining images. Baker and Nayar [2] have begun to investigate similar ideas for pattern recognition problems. To effectively handle large image databases will require a balance between increasingly fine measures (such as histogram refinement) and efficient coarse measures. Acknowledgements We wish to thank Virginia Ogle for giving us access to the Chabot imagery, and Thorsten von Eicken for supplying additional images. Greg Pass has been supported by Cornell s Alumni-Sponsored Undergraduate Research Program, and Justin Miller has been supported by a grant from the GTE Research Foundation. References [1] Farshid Arman, Arding Hsu, and Ming- Yee Chiu. Image processing on compressed data for large video databases. In ACM Multimedia Conference, pages , [2] Simon Baker and Shree Nayar. Pattern rejection. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pages , [3] M.G.Brown,J.T.Foote,G.J.F.Jones, K. Sparck Jones, and S. J. Young. Automatic content-based retrieval of broadcast news. In ACM Multimedia Conference, [4] M. Flickner et al. Query by image and video content: The QBIC system. IEEE Computer, 28(9):23 32, September [5] Brian V. Funt and Graham D. Finlayson. Color constant color indexing. IEEE Transactions on Pattern Analysis and Machine Intelligence, 17(5): , May [6] J. Hafner, H. Sawhney, W. Equitz, M. Flickner, and W. Niblack. Efficient color histogram indexing for quadratic form distance functions. IEEE Transactions on Pattern Analysis and Machine Intelligence, 17(7): , July [7] Arun Hampapur, Ramesh Jain, and Terry Weymouth. Production model based digital video segmentation. Journal of Multimedia Tools and Applications, 1:1 38, March [8] Wynne Hsu, T. S. Chua, and H. K. Pung. An integrated color-spatial approach to content-based image retrieval. In ACM Multimedia Conference, pages , [9] E. H. Land and J. J. McCann. Lightness and Retinex theory. Journal of the Optical Society of America, 61(1):1 11, [10] Akio Nagasaka and Yuzuru Tanaka. Automatic video indexing and full-video search for object appearances. In 2nd Working Conference on Visual Database Systems, October
14 [11] Virginia Ogle and Michael Stonebraker. Chabot: Retrieval from a relational database of images. IEEE Computer, 28(9):40 48, September [12] K. Otsuji and Y. Tonomura. Projectiondetecting filter for video cut detection. Multimedia Systems, 1: , [13] Alex Pentland, Rosalind Picard, and Stan Sclaroff. Photobook: Content-based manipulation of image databases. International Journal of Computer Vision, 18(3): , June [21] Ramin Zabih, Justin Miller, and Kevin Mai. A feature-based algorithm for detecting and classifying scene breaks. In ACM Multimedia Conference, pages , November [22] HongJiang Zhang, Atreyi Kankanhalli, and Stephen William Smoliar. Automatic partitioning of full-motion video. Multimedia Systems, 1:10 28, [14] Rick Rickman and John Stonham. Content-based image retrieval using color tuple histograms. SPIE proceedings, 2670:2 7, February [15] John Smith and Shih-Fu Chang. Tools and techniques for color image retrieval. SPIE proceedings, 2670: , February [16] Thomas M. Strat. Natural Object Recognition. Springer-Verlag, [17] Markus Stricker and Alexander Dimai. Color indexing with weak spatial constraints. SPIE proceedings, 2670:29 40, February [18] Markus Stricker and Michael Swain. The capacity of color histogram indexing. In Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pages , [19] Michael Swain and Dana Ballard. Color indexing. International Journal of Computer Vision, 7(1):11 32, [20] Patrick Winston and Berthold Horn. Lisp. Addison-Wesley, second edition,
Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall
Automatic Photo Quality Assessment Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Estimating i the photorealism of images: Distinguishing i i paintings from photographs h Florin
A World Wide Web Based Image Search Engine Using Text and Image Content Features
A World Wide Web Based Image Search Engine Using Text and Image Content Features Bo Luo, Xiaogang Wang, and Xiaoou Tang Department of Information Engineering The Chinese University of Hong Kong ABSTRACT
Colour Image Segmentation Technique for Screen Printing
60 R.U. Hewage and D.U.J. Sonnadara Department of Physics, University of Colombo, Sri Lanka ABSTRACT Screen-printing is an industry with a large number of applications ranging from printing mobile phone
Module II: Multimedia Data Mining
ALMA MATER STUDIORUM - UNIVERSITÀ DI BOLOGNA Module II: Multimedia Data Mining Laurea Magistrale in Ingegneria Informatica University of Bologna Multimedia Data Retrieval Home page: http://www-db.disi.unibo.it/courses/dm/
HSI BASED COLOUR IMAGE EQUALIZATION USING ITERATIVE n th ROOT AND n th POWER
HSI BASED COLOUR IMAGE EQUALIZATION USING ITERATIVE n th ROOT AND n th POWER Gholamreza Anbarjafari icv Group, IMS Lab, Institute of Technology, University of Tartu, Tartu 50411, Estonia [email protected]
Image Segmentation and Registration
Image Segmentation and Registration Dr. Christine Tanner ([email protected]) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation
A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow
, pp.233-237 http://dx.doi.org/10.14257/astl.2014.51.53 A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow Giwoo Kim 1, Hye-Youn Lim 1 and Dae-Seong Kang 1, 1 Department of electronices
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Metric Spaces. Chapter 7. 7.1. Metrics
Chapter 7 Metric Spaces A metric space is a set X that has a notion of the distance d(x, y) between every pair of points x, y X. The purpose of this chapter is to introduce metric spaces and give some
Simultaneous Gamma Correction and Registration in the Frequency Domain
Simultaneous Gamma Correction and Registration in the Frequency Domain Alexander Wong [email protected] William Bishop [email protected] Department of Electrical and Computer Engineering University
AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION
AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION Saurabh Asija 1, Rakesh Singh 2 1 Research Scholar (Computer Engineering Department), Punjabi University, Patiala. 2 Asst.
Robert Collins CSE598G. More on Mean-shift. R.Collins, CSE, PSU CSE598G Spring 2006
More on Mean-shift R.Collins, CSE, PSU Spring 2006 Recall: Kernel Density Estimation Given a set of data samples x i ; i=1...n Convolve with a kernel function H to generate a smooth function f(x) Equivalent
Mean-Shift Tracking with Random Sampling
1 Mean-Shift Tracking with Random Sampling Alex Po Leung, Shaogang Gong Department of Computer Science Queen Mary, University of London, London, E1 4NS Abstract In this work, boosting the efficiency of
Tracking and Recognition in Sports Videos
Tracking and Recognition in Sports Videos Mustafa Teke a, Masoud Sattari b a Graduate School of Informatics, Middle East Technical University, Ankara, Turkey [email protected] b Department of Computer
Segmentation of building models from dense 3D point-clouds
Segmentation of building models from dense 3D point-clouds Joachim Bauer, Konrad Karner, Konrad Schindler, Andreas Klaus, Christopher Zach VRVis Research Center for Virtual Reality and Visualization, Institute
JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions
Edith Cowan University Research Online ECU Publications Pre. JPEG compression of monochrome D-barcode images using DCT coefficient distributions Keng Teong Tan Hong Kong Baptist University Douglas Chai
Canny Edge Detection
Canny Edge Detection 09gr820 March 23, 2009 1 Introduction The purpose of edge detection in general is to significantly reduce the amount of data in an image, while preserving the structural properties
Natural Language Querying for Content Based Image Retrieval System
Natural Language Querying for Content Based Image Retrieval System Sreena P. H. 1, David Solomon George 2 M.Tech Student, Department of ECE, Rajiv Gandhi Institute of Technology, Kottayam, India 1, Asst.
UNIVERSITY OF CENTRAL FLORIDA AT TRECVID 2003. Yun Zhai, Zeeshan Rasheed, Mubarak Shah
UNIVERSITY OF CENTRAL FLORIDA AT TRECVID 2003 Yun Zhai, Zeeshan Rasheed, Mubarak Shah Computer Vision Laboratory School of Computer Science University of Central Florida, Orlando, Florida ABSTRACT In this
Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability
Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller
Open issues and research trends in Content-based Image Retrieval
Open issues and research trends in Content-based Image Retrieval Raimondo Schettini DISCo Universita di Milano Bicocca [email protected] www.disco.unimib.it/schettini/ IEEE Signal Processing Society
Content-Based Image Retrieval
Content-Based Image Retrieval Selim Aksoy Department of Computer Engineering Bilkent University [email protected] Image retrieval Searching a large database for images that match a query: What kind
Choosing a digital camera for your microscope John C. Russ, Materials Science and Engineering Dept., North Carolina State Univ.
Choosing a digital camera for your microscope John C. Russ, Materials Science and Engineering Dept., North Carolina State Univ., Raleigh, NC One vital step is to choose a transfer lens matched to your
Video compression: Performance of available codec software
Video compression: Performance of available codec software Introduction. Digital Video A digital video is a collection of images presented sequentially to produce the effect of continuous motion. It takes
Cees Snoek. Machine. Humans. Multimedia Archives. Euvision Technologies The Netherlands. University of Amsterdam The Netherlands. Tree.
Visual search: what's next? Cees Snoek University of Amsterdam The Netherlands Euvision Technologies The Netherlands Problem statement US flag Tree Aircraft Humans Dog Smoking Building Basketball Table
DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS
DATA MINING CLUSTER ANALYSIS: BASIC CONCEPTS 1 AND ALGORITHMS Chiara Renso KDD-LAB ISTI- CNR, Pisa, Italy WHAT IS CLUSTER ANALYSIS? Finding groups of objects such that the objects in a group will be similar
A new cut detection algorithm with constant false-alarm ratio for video segmentation
J. Vis. Commun. Image R. 5 (004) 44 www.elsevier.com/locate/jvci A new cut detection algorithm with constant false-alarm ratio for video segmentation Tie-Yan Liu, a Kwok-Tung Lo, b, * Xu-Dong Zhang, a
Scanners and How to Use Them
Written by Jonathan Sachs Copyright 1996-1999 Digital Light & Color Introduction A scanner is a device that converts images to a digital file you can use with your computer. There are many different types
MVA ENS Cachan. Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos [email protected]
Machine Learning for Computer Vision 1 MVA ENS Cachan Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos [email protected] Department of Applied Mathematics Ecole Centrale Paris Galen
Probabilistic Latent Semantic Analysis (plsa)
Probabilistic Latent Semantic Analysis (plsa) SS 2008 Bayesian Networks Multimedia Computing, Universität Augsburg [email protected] www.multimedia-computing.{de,org} References
Topographic Change Detection Using CloudCompare Version 1.0
Topographic Change Detection Using CloudCompare Version 1.0 Emily Kleber, Arizona State University Edwin Nissen, Colorado School of Mines J Ramón Arrowsmith, Arizona State University Introduction CloudCompare
Part-Based Recognition
Part-Based Recognition Benedict Brown CS597D, Fall 2003 Princeton University CS 597D, Part-Based Recognition p. 1/32 Introduction Many objects are made up of parts It s presumably easier to identify simple
Machine vision systems - 2
Machine vision systems Problem definition Image acquisition Image segmentation Connected component analysis Machine vision systems - 1 Problem definition Design a vision system to see a flat world Page
Classifying Manipulation Primitives from Visual Data
Classifying Manipulation Primitives from Visual Data Sandy Huang and Dylan Hadfield-Menell Abstract One approach to learning from demonstrations in robotics is to make use of a classifier to predict if
Comparison of different image compression formats. ECE 533 Project Report Paula Aguilera
Comparison of different image compression formats ECE 533 Project Report Paula Aguilera Introduction: Images are very important documents nowadays; to work with them in some applications they need to be
Tracking of Small Unmanned Aerial Vehicles
Tracking of Small Unmanned Aerial Vehicles Steven Krukowski Adrien Perkins Aeronautics and Astronautics Stanford University Stanford, CA 94305 Email: [email protected] Aeronautics and Astronautics Stanford
Visibility optimization for data visualization: A Survey of Issues and Techniques
Visibility optimization for data visualization: A Survey of Issues and Techniques Ch Harika, Dr.Supreethi K.P Student, M.Tech, Assistant Professor College of Engineering, Jawaharlal Nehru Technological
Implementation of Canny Edge Detector of color images on CELL/B.E. Architecture.
Implementation of Canny Edge Detector of color images on CELL/B.E. Architecture. Chirag Gupta,Sumod Mohan K [email protected], [email protected] Abstract In this project we propose a method to improve
Image Normalization for Illumination Compensation in Facial Images
Image Normalization for Illumination Compensation in Facial Images by Martin D. Levine, Maulin R. Gandhi, Jisnu Bhattacharyya Department of Electrical & Computer Engineering & Center for Intelligent Machines
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 269 Class Project Report
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 69 Class Project Report Junhua Mao and Lunbo Xu University of California, Los Angeles [email protected] and lunbo
Object Recognition and Template Matching
Object Recognition and Template Matching Template Matching A template is a small image (sub-image) The goal is to find occurrences of this template in a larger image That is, you want to find matches of
EECS 556 Image Processing W 09. Interpolation. Interpolation techniques B splines
EECS 556 Image Processing W 09 Interpolation Interpolation techniques B splines What is image processing? Image processing is the application of 2D signal processing methods to images Image representation
An Iterative Image Registration Technique with an Application to Stereo Vision
An Iterative Image Registration Technique with an Application to Stereo Vision Bruce D. Lucas Takeo Kanade Computer Science Department Carnegie-Mellon University Pittsburgh, Pennsylvania 15213 Abstract
Introduction. Chapter 1
1 Chapter 1 Introduction Robotics and automation have undergone an outstanding development in the manufacturing industry over the last decades owing to the increasing demand for higher levels of productivity
Multivariate data visualization using shadow
Proceedings of the IIEEJ Ima and Visual Computing Wor Kuching, Malaysia, Novembe Multivariate data visualization using shadow Zhongxiang ZHENG Suguru SAITO Tokyo Institute of Technology ABSTRACT When visualizing
A Short Introduction to Computer Graphics
A Short Introduction to Computer Graphics Frédo Durand MIT Laboratory for Computer Science 1 Introduction Chapter I: Basics Although computer graphics is a vast field that encompasses almost any graphical
LOCAL SURFACE PATCH BASED TIME ATTENDANCE SYSTEM USING FACE. [email protected]
LOCAL SURFACE PATCH BASED TIME ATTENDANCE SYSTEM USING FACE 1 S.Manikandan, 2 S.Abirami, 2 R.Indumathi, 2 R.Nandhini, 2 T.Nanthini 1 Assistant Professor, VSA group of institution, Salem. 2 BE(ECE), VSA
In mathematics, there are four attainment targets: using and applying mathematics; number and algebra; shape, space and measures, and handling data.
MATHEMATICS: THE LEVEL DESCRIPTIONS In mathematics, there are four attainment targets: using and applying mathematics; number and algebra; shape, space and measures, and handling data. Attainment target
Digital Image Fundamentals. Selim Aksoy Department of Computer Engineering Bilkent University [email protected]
Digital Image Fundamentals Selim Aksoy Department of Computer Engineering Bilkent University [email protected] Imaging process Light reaches surfaces in 3D. Surfaces reflect. Sensor element receives
CBIR: Colour Representation. COMPSCI.708.S1.C A/P Georgy Gimel farb
CBIR: Colour Representation COMPSCI.708.S1.C A/P Georgy Gimel farb Colour Representation Colour is the most widely used visual feature in multimedia context CBIR systems are not aware of the difference
Session 7 Bivariate Data and Analysis
Session 7 Bivariate Data and Analysis Key Terms for This Session Previously Introduced mean standard deviation New in This Session association bivariate analysis contingency table co-variation least squares
Data Storage 3.1. Foundations of Computer Science Cengage Learning
3 Data Storage 3.1 Foundations of Computer Science Cengage Learning Objectives After studying this chapter, the student should be able to: List five different data types used in a computer. Describe how
How To Filter Spam Image From A Picture By Color Or Color
Image Content-Based Email Spam Image Filtering Jianyi Wang and Kazuki Katagishi Abstract With the population of Internet around the world, email has become one of the main methods of communication among
Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data
CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear
A Genetic Algorithm-Evolved 3D Point Cloud Descriptor
A Genetic Algorithm-Evolved 3D Point Cloud Descriptor Dominik Wȩgrzyn and Luís A. Alexandre IT - Instituto de Telecomunicações Dept. of Computer Science, Univ. Beira Interior, 6200-001 Covilhã, Portugal
Environmental Remote Sensing GEOG 2021
Environmental Remote Sensing GEOG 2021 Lecture 4 Image classification 2 Purpose categorising data data abstraction / simplification data interpretation mapping for land cover mapping use land cover class
. Learn the number of classes and the structure of each class using similarity between unlabeled training patterns
Outline Part 1: of data clustering Non-Supervised Learning and Clustering : Problem formulation cluster analysis : Taxonomies of Clustering Techniques : Data types and Proximity Measures : Difficulties
Character Image Patterns as Big Data
22 International Conference on Frontiers in Handwriting Recognition Character Image Patterns as Big Data Seiichi Uchida, Ryosuke Ishida, Akira Yoshida, Wenjie Cai, Yaokai Feng Kyushu University, Fukuoka,
Data Storage. Chapter 3. Objectives. 3-1 Data Types. Data Inside the Computer. After studying this chapter, students should be able to:
Chapter 3 Data Storage Objectives After studying this chapter, students should be able to: List five different data types used in a computer. Describe how integers are stored in a computer. Describe how
Computer Vision. Color image processing. 25 August 2014
Computer Vision Color image processing 25 August 2014 Copyright 2001 2014 by NHL Hogeschool and Van de Loosdrecht Machine Vision BV All rights reserved [email protected], [email protected] Color image
Edge tracking for motion segmentation and depth ordering
Edge tracking for motion segmentation and depth ordering P. Smith, T. Drummond and R. Cipolla Department of Engineering University of Cambridge Cambridge CB2 1PZ,UK {pas1001 twd20 cipolla}@eng.cam.ac.uk
Speed Performance Improvement of Vehicle Blob Tracking System
Speed Performance Improvement of Vehicle Blob Tracking System Sung Chun Lee and Ram Nevatia University of Southern California, Los Angeles, CA 90089, USA [email protected], [email protected] Abstract. A speed
Clustering. Adrian Groza. Department of Computer Science Technical University of Cluj-Napoca
Clustering Adrian Groza Department of Computer Science Technical University of Cluj-Napoca Outline 1 Cluster Analysis What is Datamining? Cluster Analysis 2 K-means 3 Hierarchical Clustering What is Datamining?
CONTENT-BASED IMAGE RETRIEVAL FOR ASSET MANAGEMENT BASED ON WEIGHTED FEATURE AND K-MEANS CLUSTERING
CONTENT-BASED IMAGE RETRIEVAL FOR ASSET MANAGEMENT BASED ON WEIGHTED FEATURE AND K-MEANS CLUSTERING JUMI¹, AGUS HARJOKO 2, AHMAD ASHARI 3 1,2,3 Department of Computer Science and Electronics, Faculty of
Determining optimal window size for texture feature extraction methods
IX Spanish Symposium on Pattern Recognition and Image Analysis, Castellon, Spain, May 2001, vol.2, 237-242, ISBN: 84-8021-351-5. Determining optimal window size for texture feature extraction methods Domènec
An Order-Invariant Time Series Distance Measure [Position on Recent Developments in Time Series Analysis]
An Order-Invariant Time Series Distance Measure [Position on Recent Developments in Time Series Analysis] Stephan Spiegel and Sahin Albayrak DAI-Lab, Technische Universität Berlin, Ernst-Reuter-Platz 7,
Signature verification using Kolmogorov-Smirnov. statistic
Signature verification using Kolmogorov-Smirnov statistic Harish Srinivasan, Sargur N.Srihari and Matthew J Beal University at Buffalo, the State University of New York, Buffalo USA {srihari,hs32}@cedar.buffalo.edu,[email protected]
Local features and matching. Image classification & object localization
Overview Instance level search Local features and matching Efficient visual recognition Image classification & object localization Category recognition Image classification: assigning a class label to
CS 534: Computer Vision 3D Model-based recognition
CS 534: Computer Vision 3D Model-based recognition Ahmed Elgammal Dept of Computer Science CS 534 3D Model-based Vision - 1 High Level Vision Object Recognition: What it means? Two main recognition tasks:!
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
Texture. Chapter 7. 7.1 Introduction
Chapter 7 Texture 7.1 Introduction Texture plays an important role in many machine vision tasks such as surface inspection, scene classification, and surface orientation and shape determination. For example,
Descriptive statistics Statistical inference statistical inference, statistical induction and inferential statistics
Descriptive statistics is the discipline of quantitatively describing the main features of a collection of data. Descriptive statistics are distinguished from inferential statistics (or inductive statistics),
COLOR-BASED PRINTED CIRCUIT BOARD SOLDER SEGMENTATION
COLOR-BASED PRINTED CIRCUIT BOARD SOLDER SEGMENTATION Tz-Sheng Peng ( 彭 志 昇 ), Chiou-Shann Fuh ( 傅 楸 善 ) Dept. of Computer Science and Information Engineering, National Taiwan University E-mail: [email protected]
Palmprint Recognition. By Sree Rama Murthy kora Praveen Verma Yashwant Kashyap
Palmprint Recognition By Sree Rama Murthy kora Praveen Verma Yashwant Kashyap Palm print Palm Patterns are utilized in many applications: 1. To correlate palm patterns with medical disorders, e.g. genetic
Highlight Removal by Illumination-Constrained Inpainting
Highlight Removal by Illumination-Constrained Inpainting Ping Tan Stephen Lin Long Quan Heung-Yeung Shum Microsoft Research, Asia Hong Kong University of Science and Technology Abstract We present a single-image
CS231M Project Report - Automated Real-Time Face Tracking and Blending
CS231M Project Report - Automated Real-Time Face Tracking and Blending Steven Lee, [email protected] June 6, 2015 1 Introduction Summary statement: The goal of this project is to create an Android
Open Access A Facial Expression Recognition Algorithm Based on Local Binary Pattern and Empirical Mode Decomposition
Send Orders for Reprints to [email protected] The Open Electrical & Electronic Engineering Journal, 2014, 8, 599-604 599 Open Access A Facial Expression Recognition Algorithm Based on Local Binary
Vision based Vehicle Tracking using a high angle camera
Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu [email protected] [email protected] Abstract A vehicle tracking and grouping algorithm is presented in this work
Neovision2 Performance Evaluation Protocol
Neovision2 Performance Evaluation Protocol Version 3.0 4/16/2012 Public Release Prepared by Rajmadhan Ekambaram [email protected] Dmitry Goldgof, Ph.D. [email protected] Rangachar Kasturi, Ph.D.
International Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 3, May-Jun 2014
RESEARCH ARTICLE OPEN ACCESS A Survey of Data Mining: Concepts with Applications and its Future Scope Dr. Zubair Khan 1, Ashish Kumar 2, Sunny Kumar 3 M.Tech Research Scholar 2. Department of Computer
Detection and Restoration of Vertical Non-linear Scratches in Digitized Film Sequences
Detection and Restoration of Vertical Non-linear Scratches in Digitized Film Sequences Byoung-moon You 1, Kyung-tack Jung 2, Sang-kook Kim 2, and Doo-sung Hwang 3 1 L&Y Vision Technologies, Inc., Daejeon,
Face Model Fitting on Low Resolution Images
Face Model Fitting on Low Resolution Images Xiaoming Liu Peter H. Tu Frederick W. Wheeler Visualization and Computer Vision Lab General Electric Global Research Center Niskayuna, NY, 1239, USA {liux,tu,wheeler}@research.ge.com
Supporting Online Material for
www.sciencemag.org/cgi/content/full/313/5786/504/dc1 Supporting Online Material for Reducing the Dimensionality of Data with Neural Networks G. E. Hinton* and R. R. Salakhutdinov *To whom correspondence
Grade 5 Math Content 1
Grade 5 Math Content 1 Number and Operations: Whole Numbers Multiplication and Division In Grade 5, students consolidate their understanding of the computational strategies they use for multiplication.
Template-based Eye and Mouth Detection for 3D Video Conferencing
Template-based Eye and Mouth Detection for 3D Video Conferencing Jürgen Rurainsky and Peter Eisert Fraunhofer Institute for Telecommunications - Heinrich-Hertz-Institute, Image Processing Department, Einsteinufer
FPGA Implementation of Human Behavior Analysis Using Facial Image
RESEARCH ARTICLE OPEN ACCESS FPGA Implementation of Human Behavior Analysis Using Facial Image A.J Ezhil, K. Adalarasu Department of Electronics & Communication Engineering PSNA College of Engineering
