Using a Markov Random Field for Image Re-ranking Based on Visual and Textual Features
|
|
|
- Paul Park
- 10 years ago
- Views:
Transcription
1 Using a Markov Random Field for Image Re-ranking Based on Visual and Textual Features Utilizando un Campo Aleatorio de Markov para el Reordenamiento de Imágenes Basado en Atributos Visuales y Textuales R. Omar Chávez, Manuel Montes and L. Enrique Sucar Coordinación de Ciencias Computacionales, Instituto Nacional de Astrofísica Óptica y Electrónica, Puebla, México {romarcg,mmontesg,esucar}@ccc.inaoep.mx Article received on January 15, 2010; accepted on June 29, 2010 Abstract. We propose a novel method to re-order the list of images returned by an image retrieval system (IRS). The method combines the original order obtained by the IRS, the similarity between images obtained with visual and textual features, and a relevance feedback approach, all of them with the purpose of separating relevant from irrelevant images, and thus, obtaining a more appropriate order. The method is based on a Markov random field (MRF) model, in which each image in the list is represented as a random variable that could be relevant or irrelevant. The energy function proposed for the MRF combines two factors: the similarity between the images in the list (internal similarity); and information obtained from the original order and the similarity of each image with the query (external similarity). Experiments were conducted with resources from the Image CLEF 2008 forum for the photo retrieval track, taking into account textual and visual features. The results show that the proposed method improves, according to the MAP measure, the order of the original list up to 63% (in the textual case) and up to 55% (in the visual case); and suggest future work using a combination of both kind of features. Keywords: Image Re-ranking, Image Retrieval, Markov Random Field, Relevance Feedback. Resumen. En este trabajo proponemos un método novedoso para re-ordenar una lista de imágenes recuperadas por un sistema de recuperación de imágenes (SRI). El método combina el orden original obtenido por el SRI, la similitud entre imágenes, obtenida con las características visuales y textuales, y un enfoque de retroalimentación de relevancia, todos ellos con el propósito de separar las imágenes relevantes de las irrelevantes, y así, obtener un orden más apropiado. El método está basado en el modelo de un campo aleatorio de Markov (CAM), en el que cada imagen en la lista fue representada como una variable aleatoria con dos posibles valores: relevante o irrelevante. La función de energía propuesta para el campo aleatorio de Markov combina dos factores: la similitud entre imágenes en la lista (similitud interna); y la información obtenida del orden original y la similitud de cada imagen con la consulta (similitud externa). Los experimentos fueron realizados con los recursos del foro Image CLEF 2008 para la tarea de recuperación de fotografías, tomando en cuenta los atributos textuales y visuales. Los resultados mostraron que el método propuesto mejora, de acuerdo con la medida MAP, el orden de la lista original hasta en un 63% (en el caso textual) y hasta un 55% (en el caso visual); y sugieren como trabajo a futuro el utilizar una combinación de ambos tipos de atributos. Palabras clave: Re-ordenamiento de Imágenes, Recuperación de Imágenes, Campos Aleatorios de Markov, Retroalimentación de Relevancia. 1 Introduction An image retrieval system (IRS) receives a query from a user, as keywords or sample images, and it is expected to return an ordered list of images that satisfies the user's request. Ideally, the IRS should return a list with all relevant images (Datta et al. 2008) ordered according to the user's request, so that the top images in the list are the ones closer to the user's expectations. The proper order in the list is important because it is easier and faster for the user to find images relevant to the query (Deselaers et al. 2008; Cui et al. 2008; Marakakis et al. 2008). Current IRS, in general, tend to include several relevant images in the retrieved list. However, the images are not ordered properly, so there could be relevant images that are at the bottom of the list, and
2 394 R. Omar Chávez, Manuel Montes and L. Enrique Sucar many of the top images are not relevant. That is, IRS have a relatively good performance in terms of recall, but poor in terms of precision, in particular precision in the first 5, 10 and 20 images. Where precision indicates the percentage of relevant retrieved items, and recall the percentage of retrieved relevant elements from the total relevant items (Mani, 2001). Therefore an alternative to improve image retrieval is to re-order the list of images, so that the relevant ones are in the top of the list. One way to improve the order of the results of an IRS is to use relevance feedback. That is, once the results are obtained by the IRS, a subset of relevant images is selected manually (by the user) or automatically. This subset is used to further refine the list obtained (Deselaers et al. 2008; Marakakis et al. 2008; Cui et al. 2008). There are several approaches for relevance feedback. Some attempt to enrich the query to perform a new retrieval and obtain better results (Datta et al. 2008). This approach tends to be computationally expensive because it must re-use retrieval mechanisms to obtain the relevant images from the entire collection. This motivates to use only the retrieved list and reorder the images on the assumption that this list has relevant images, but not necessarily in the first positions. Previous work (Lin et al. 2003; Cui et al. 2008; Marakakis et al. 2008) that tries to improve the order of a list of retrieved images do not use all the information available. Some just use a relevance feedback approach to locate images similar to the selected images as feedback. Other methods generate models that depend on the collection of images, assuming a certain number of search intentions. Some methods omit the use of information provided by the IRS, for example the original order. We consider that all the available information the original order, the subset obtained via relevance feedback, the original query, and the entire list of retrieved images is useful to improve the list order, and propose a re-ranking method that combines all this information. This paper proposes a method that combines the original order obtained by a IRS, the similarity between images obtained with textual and visual features and a relevance feedback approach, all of them with the purpose of separating the relevant images from does that are not relevant, and thus obtain a more appropriate order for the results generated by the base IRS. The method is based on a Markov random field (MRF) model, in which each image in the list is represented as a random variable that could be relevant or irrelevant. The relevance feedback is incorporated in the initialization of the model, making these images relevant. The energy function of the MRF combines two factors: the similarity between the images in the list (internal similarity); and information obtained from the original order and the similarity of each image with the query (external similarity). Taking these factors into account and assigning a weight to each, the MRF is solved (obtaining the more probable configuration) separating the relevant images from the rest. Based on this result, the list of images is reordered placing in the top positions the images selected as feedback and the images marked as relevant by the MRF, and at bottom the rest of images. In this paper we consider an image collection that includes, for each query, three sample images and a textual description. Each image in the collection has an associated text. We consider for measuring similarity the textual and visual components. Experiments were conducted using the resources of the forum ImageCLEF 2008 for the photo retrieval track (Arni et al. 2008). We used one of the IRS that participated in this forum as the base retriever to obtain the initial lists, and tested our method with each one of the 39 queries used in this competition. Each of the results obtained by our method improved the original list order; according to the MAP measure, the improvements obtained are up to 63% using textual features and 55% using visual features. The rest of the document is organized as follows. Section 2 provides a brief review of related work. Section 3 describes the proposed method. Section 4 describes the textual and visual features used. Section 5 presents the experiments and the results. Finally, Section 6 gives the conclusions and suggests future work. 2 Related Work The task of image retrieval consists of, given a user query, retrieve all the relevant images from an image collection. The query can be visual (one or more sample images), textual (a set of keyword or sentences) or a combination of both. The IRS returns a list of relevant images that should be ordered according to the user's request.
3 Using a Markov Random Field for Image Re-ranking Based on Visual and Textual Features 395 An IRS may use visual or textual features, or a combination of them, to index and retrieve images. The results obtained by the IRS depend largely on the features used to describe the images, so that most of the time the order of the results are not appropriate for all the queries. Some methods try to improve the order obtained by enriching the query and searching the image again, but because they have to rerun retrieval mechanisms this solution is computationally expensive. In general, the lists returned by IRS have several relevant images, but these are not placed in the top of the list. As mentioned before, it is important that the relevant images are in the first positions in the retrieved list. One way to improve the order of the results of an IRS is to use relevance feedback (Clough & Sanderson, 2004). That is, once the results are obtained by the IRS, a subset of relevant images is selected and used to further refine the given list. There are different approaches for image reranking that can be classified as follows: 1. According to the features used to describe the image content: a) Visual. Use visual image features such as color, texture, shape, SIFT (Scale Invariant Feature Transform), SURF (Speeded Up Robust Features) among others to describe the visual content of each image (Lowe, 2004; Cui et al. 2008; Bay et al. 2006; Berg, 2009; Jianjiang et al. 2007; Jégou et al. 2010). It is still an unsolved problem to find the visual features needed to fully represent the visual content of an image. b) Textual. Use descriptive text to represent the image content (Datta et al. 2008; Lin et al. 2003; Chong et al. 2009; Choochaiwattana & Spring, 2009). It is expected that the textual description of the image must be both specific and general to cover all search intentions. c) Multimodal. Combine visual and textual features to describe image content, trying to cover more search intentions (Awang Iskanda et al. 2006; Gong & Liu, 2009; Richter et al. 2010). An open problem is how to combine both types of features to take advantage of the selected features. 2. Based on the type of relevance feedback used: a) Manual. Ask a user to select some of the retrieved images that he considers relevant to his search intention (Datta et al. 2008; Jianjiang et al. 2007; Choochaiwattana & Spring, 2009). The greater the number of selected images, the IRS will have more information about the images relevant to the user, but it also implies more additional time and effort for the user. b) Automatic. Make assumptions about the user's search intention to automatically select some sample images (Cui et al. 2008; Marakakis et al. 2008; Chong et al. 2009; Bihong et al. 2007). It is a difficult task making assumptions that cover all search intentions. This approach reduces the user's effort. 3. According to the technique used to sort images: a) Classifiers.Select some relevant and irrelevant images to build a classifier and label the remaining images as relevant or irrelevant (Martin et al. 2004; Deselaers et al. 2008). These methods compare images from the collection with those selected as relevant (positive examples) without taking into account other external information or images labeled as relevant by the classifier itself. b) Similarity functions. Use similarity functions to determine the images that are more related with pre-established categories (Bihong et al. 2007; Cui et al. 2008; Choochaiwattana & Spring, 2009). These methods depend on the type of images and the number of categories in the collection. c) Probabilistic models. Use probabilistic models to order the retrieved images, Gaussian mixtures models to combine features, graph based models to describe the relation between images or spatial configuration models to improve the object retrieval precision (Ke et al. 2008; Lin et al. 2003; Richter et al. 2010; Gong & Liu, 2009). These methods need to process the entire image collection to get a more accurate model. If the images collection is changing constantly, is necessary to rebuild the model. d) List fusion. Merge result lists from several IRSs for the same query, in order to get a list to improve the order of any of the base lists taken into account (Escalante et al. 2008). Choose what lists and how merged these is the main problem of this approach. Previous methods do not take into account all the available information to produce an adequate order for the retrieved list of images. In contrast, the method proposed in this paper, combines the original order obtained by an IRS, the similarity
4 396 R. Omar Chávez, Manuel Montes and L. Enrique Sucar between images based on textual or visual features, and a manual relevance feedback approach; all of them with the purpose of separating the relevant images from those that are not relevant, and thus obtaining a more appropriate order than that generated by the base IRS. The proposed method does not require accessing again the entire collection, considering only the list provided by IRS. The method is described in detail in the following sections. 3 Proposed Method A general outline of the proposed method is given in Fig. 1. Given a query, the IRS retrieves from a given collection of images (that includes text captions) a list of files sorted according to a relevance criteria. From this list, some relevant images are selected based on a relevance feedback approach. For each image in the list, textual and visual features are extracted. The textual and visual feature description of each image in the list, the query given by the user, and a subset of images selected via relevance feedback, are combined to produce a re-ordered list. This re-ranking is obtained based on a Markov random field (MRF) model that separates the relevant images from irrelevant ones, generating a new list by positioning the relevant images first, and the others after. Next we give a brief review of MRFs, and then we describe in detail each component of the proposed method. 3.1 Markov Random Fields Markov Random Fields (Li, 2004) are probabilistic models which combine a priori knowledge given by some observations and knowledge given by the interaction with neighbors. Let be random variables on a set S, where each can take a value in a set of labels L. This F is called a random field, and the instantiation of each of these as an is what is called a configuration of F, so, the probability that a random variable takes the value is denoted by, and the joint probability is denoted as. A random field is said to be an MRF if it has the property of locality, i.e., if the field satisfies the following property: (1) where represents the set S without the element,, and is the set of neighboring nodes of the node. The joint probability can be expressed as: where Z is called the partition function or normalizing constant, and is called the energy function. The optimal configuration is found by minimizing the energy function, obtaining a value for every random variable in F. Fig. 1. Block diagram of the proposed method 3.2 Model In our case we consider a MRF in which each node corresponds to a document (image + text caption) in the list. Each document is represented as a random variable with 2 possible values: relevant and irrelevant. We consider a fully connected graph, such that each node (document) is connected to all other nodes in the field; that is, we defined a neighborhood scheme in which each variable is adjacent to all the others. Given that the number of images in the list is relatively low (100 in the experiments), to consider a complete graph is not a problem computationally, and allows us to consider the relations between all documents in the list. For representing the images we consider visual features from the image or textual features from the caption. To describe the images, in the textual case, we used a binary bag of words representation, in which each vector element represents a textual word (2)
5 Using a Markov Random Field for Image Re-ranking Based on Visual and Textual Features 397 from the collection vocabulary; and the query is represented in the same manner. In visual case, the retrieved images, like the query images, are described by SIFT (Scale-Invariant Feature Transform) features. The internal and external similarities are considered via the energy function described next. 3.3 Energy Function The energy function of the MRF combines two factors: the similarity between the images in the list (internal similarity); and external information obtained from the original order and the similarity of each image with the query (external similarity). The internal similarities correspond to the interaction potentials and the external similarities to the observation potentials. The proposed energy function takes into account both aspects and is defined as follows: Where is the interaction potential and it considers the similarity between random variable and its neighbors, representing the support that neighboring variables give to. is the observation potential and represents the influence of external information on variable. The weight factor λ favors ( < 1), ( > 1), or both ( = 1). is defined as: Where ( ) represents the average similarity between variable and its neighbors with irrelevant value. ( ) represents the average similarity between variable and its neighbors with relevant value. Where n+m is the total of images in the original list. is defined as follows: (3) (4) (5) The potential is obtained by combing two factors. The first indicates how similar,, or different, is the variable with the query q. The second is a function that converts the position in the list given by a base IRS to a real value. The function returns the position of the image in the original list, returns the inverse position of the variable in this list. The initial configuration of the MRF is obtained by relevance feedback. That is, the subset of images selected via relevance feedback are initialized as relevant, and all other images as irrelevant. Then, the MRF configuration of minimum energy (MAP) is obtained via stochastic simulation using the ICM algorithm. We experimented using also Simulated Annealing with similar results (Chellappa, 1993). At the end of this optimization process, each variable (image) has a value of relevant or irrelevant. Based on these values, a new re-ordered list is produced, by positioning first the relevant images according to the MRF, and then the irrelevant ones. 4 Features Extraction Each image is represented by its textual or visual features; for each of them a word based representation is used: textual words and visual words. In the case of textual features we used the words of the textual description of the image, representing each image by a word vector. Each element of the vector indicates the absence (0) or presence (1) value of a textual word in the image description. To obtain the textual description of each image we follow the next steps: 1. Stopword removal. Stopwords were removed from the description of the retrieved images and the query. 2. Vocabulary extraction. We obtained the vocabulary from all textual descriptions of images and query. 3. Vector construction. We identified the occurrence or absence of the vocabulary words in the description of each of the images and the query and the vector was constructed for each image. In the case of visual features, SIFT features were used to represent objects within the image. These features are taken as visual words and, as in the case of textual features, we built a representation with these words. To find the SIFT features in an
6 398 R. Omar Chávez, Manuel Montes and L. Enrique Sucar image we performed the following steps (Lowe, 2004): 1. Scale-space extrema detection. The first stage of computation searches over all scales and image locations. 2. Keypoint localization. At each candidate location, a detailed model is fit to determine location and scale. 3. Orientation assignment. One or more orientations are assigned to each keypoint location based on local image gradient directions. 4. Keypoint descriptor. The local image gradients are measured at the selected scale in the region around each keypoint. The SIFT features are invariant to image scale and rotation, change in 3D viewpoint, additive noise, and change in illumination. To calculate the SIFT points and common points between images we used the implementations proposed in (Lowe, 2004). Fig. 2 shows the textual words selected from the descriptive fields and the SIFT keypoints of an image in the collection. Only the textual words showed in the Fig. 2 have the value 1 in the textual vector from this image. The SIFT keypoints are specified by 4 floating point numbers giving subpixel row and column location, scale, and orientation, the invariant descriptor vector for the keypoint is given as a 128 integers in the range [0,255] (Lowe, 2004). includes for each image, a textual description of its visual content. < machu, picchu, rear, view, llama, terraces, ruins, bald, mountain, range, clouds, background > Fig. 2. SIFT keypoints and textual words obtained after feature extraction of an image in the collection 5 Experimental Evaluation We conducted a series of experiments with the following objectives: (i) to test the results of the proposed method compared with the original list, (ii) to asses the level of improvement when using textual or visual features to measure similarity, (iii) to evaluate the sensitivity of the method to the model parameters and (iv) to compare the improvement between the use of visual or textual features. 5.1 Experimental setup To perform the experiments we used the resources of the forum Image CLEF 2008, which consists of the image collection IAPR TC-12 (Arni et al. 2008), a set of queries for the photo retrieval track and a list of results from one of the participants (TIA-TXTIMG). This collection was chosen because it has relevance judgments for each query, this allowed a comparison with results obtained by the proposed method; it also <title> church with more than two towers</title> <narr> Relevant images will show a church, cathedral or a mosque with three or more towers.</narr> Fig. 3. Query example for the photo retrieval track from the Image CLEF 2008 forum. Each query includes, among other fields, a field title summarizing the query objective, and a field narrative specifying textually which images are relevant to the query The TIA-TXTIMG SRI retrieves a list of images by combining the results of several retrieval methods (Escalante et al. 2008). We considered the best results obtained by this group as the input for our method.
7 Using a Markov Random Field for Image Re-ranking Based on Visual and Textual Features 399 The proposed method was tested with the 39 queries from Image CLEF 2008 forum. These queries have 3 sample images and a textual description that includes a narrative about the images relevant to the query. For experiments using only textual information, the sample images were not considered. For experiments using visual information the proposed method used only the three sample images for each query. Fig. 3 shows a sample query for the photo retrieval track. Each of the images in the IAPR TC-12 collection has assigned a set of descriptive fields, we included the words in the title and in the textual description to represent the images. Fig. 4 shows an example of an image of the IAPR TC-12 collection and its descriptive fields. As input data for the proposed method, we selected the first 100 images retrieved by TIA- TXTIMG IRS for each of the 39 queries. For evaluation we used MAP and Precision at the first N retrieved items (Mani, 2001). MAP is defined as follows: For the similarity based on visual features, the following similarity functions were proposed: The similarity between neighboring variables was measured using, where and are variables represented by its SIFT features. The similarity function is defined as: (9) The function finds the common number of SIFT features between the variable and. Function calculates the number of SIFTS features found for image. Because the query is composed by 3 example images, the similarity function between an image and query Q was defined as: (10) (6) Where is the precision at the first r documents, is a binary function which indicates if document at position r is relevant or not for the query i; n is the total number of relevant documents for the query i, m is the number of relevant documents retrieved and Q is the set of all queries. Precision at N is defined as the percentage of retrieved relevant items at the first N positions of the result list. Experiments were conducted using only textual or visual features. For this, similarity functions were defined for the potentials and. For the similarity based on textual features, the following similarity functions were proposed: The similarity function used to measure the similarity between variables is defined as:, where: The similarity function between an image and the query q was defined as, where: (7) (8) <TITLE>The Plaza de Armas</TITLE> <DESCRIPTION>a yellow building with white columns in the background; two palm trees in front of the house; cars parkedin front of the house; a woman and a child are walking over the square</description> Fig. 4. Example of an image from the IAPR TC-12 collection and its set of descriptive fields: title and description The features for textual and visual similarity functions can be defined based on 2 general operations. The first is the intersection of either textual or visual words, and determines the words in
8 Baseline Textual features Visual features 400 R. Omar Chávez, Manuel Montes and L. Enrique Sucar common between two images. The second is the coverage, and determines how many words in an image are contained in the other. The function used to map the order from the original list is, the intuitive idea of this function is such that it first increases slowly so that the top images have a small potential, and then it increases exponentially to amplify the potential for those images in the bottom of the list. 5.2 Experimental results Five experiments were conducted for each type of feature, visual or textual, varying λ, see Table 1. Each of the 5 experiments were made taking into account 1, 3, 5, 8 and 10 images as relevance feedback. A simulated user feedback technique was used to perform the experiments. The collection used contains, in addition to the queries and images, relevance judgments indicating which images are relevant to each of the proposed queries, given that it is known beforehand which images are relevant in the retrieved list, hence some of this images are taken as feedback. This type of feedback is known as simulated user feedback. When the MRF converges, the images selected by the relevance feedback are placed in the top of the new list, then are placed those with relevant value respecting the order in the original list. Images with irrelevant value are placed after placing all the images with relevant value. Table.1..Meaning of each experiment conducted to evaluate the proposed method based on the value of λ Value of λ λ = 1.5 λ = 1 λ = 0.5,0.3 λ = 0 λ = Description More importance to Va Equal importance to Vc and Va More importance to Vc Cancel the contribution of Va Cancel the contribution of Vc A comparison between the results of the original list retrieved by the TIA-TXTIMG IRS and the results obtained with the proposed method (using visual or textual features) for the different configurations of parameters and features are shown in Fig. 5, where the results indicated the average of the MAP values obtained for the 39 queries. The graph shows that better values are obtained with consistently for different number of images selected as feedback. Note that all variants of the proposed method showed in Fig. 5 improved the results of the original list. When the value of λ is small (eg. ) the proposed method yields the best results. So it seems that, at least for this collection, the information from the neighbors is more valuable than the information from the original order and the similarity with the query. Table 2 shows a comparison between the best results obtained by the proposed method and the results obtained by the TIA-TXTIMG IRS. This table also shows the values for precision measure at the first 5, 10 and 20 retrieved images, where the proposed method also overcomes the TIA-TXTIMG IRS. Table 2. A comparison between the best results obtained by some variants of the proposed method and the results obtained by the TIA-TXTIMG IRS. The number after the letter F indicates the number of images taken for relevance feedback, the number following the letter L indicates the value of λ. P5, P10 and P20 indicate the precision to 5, 10 and 20 retrieved images respectively. MAP is the Mean Average Precision Experiment P5 P10 P20 MAP TIA TXT- IMG F1-L F3-L F5-L F8-L F10-L F1-L F3-L F5-L F8-L F10-L The results show that for simulated user feedback and considering only textual features to measure the similarity of images, an improvement of up to 63% in the MAP is obtained selecting 10 images as feedback, and an improvement of 23% in the MAP when only one image is selected as feedback. As more images are given as feedback, the performance improves. The results for visual features also show that proposed method improved by to 55% the MAP the
9 Using a Markov Random Field for Image Re-ranking Based on Visual and Textual Features 401 original list when 10 images were selected as feedback, and by 9% when only one image was used as feedback. To show that the results obtained by the proposed method are significant with respect to the results from the base IRS, we used the paired t- student test, and showed that using a confidence level the results are statistically significant (Kanji, 1993; Dietterich, 1998). a) Fig. 5. Comparison between the results obtained by the TIA-TXTIMG IRS and the results obtained by the proposed method, using visual or textual features, for simulated user feedback varying the value of λ and the number of selected images as feedback (the bottom line shows the MAP for the original list) Fig. 6 shows the first 20 images from the original list obtained by the IRS for the query: straight road in the USA as well as the list obtained after the application of the proposed method. In this example the proposed method, using textual or visual features, placed in the top of the list some images that were not in the original list, in fact all images placed in the first 10 positions are relevant. For this query, textual features were sufficient to properly order (with the proposed method) the recovered images. Experiments showed cases where the textual description is insufficient to locate relevant images. Fig. 7 shows the 20 first images obtained by the IRS and the first 20 obtained by our method, using textual features only, for the query: church with more than two towers. For this example, few of the images at the top are relevant, because in its textual description is not mentioned information about the number of towers in the building, which is an important factor that determines the relevance of the images to this query. In addition, some relevant images that were not selected by the method using textual features only, were identified using visual features. These results motivate the use of a combination of features that allows to use the advantage of both. c) Fig. 6. First 20 images in the list obtained by the IRS (a) and the list sorted by our method using textual (b) or visual (c) features for the query straight road in the USA. The relevant images to the query are indicated with a red dot in the upper left corner b)
10 402 R. Omar Chávez, Manuel Montes and L. Enrique Sucar 6 Conclusions and Future Work This paper proposed a method for improving the order of a list of images retrieved by an IRS. Based on relevance feedback, the proposed method integrates, via a MRF, to separate the relevant and irrelevant images from the original list: the similarity between the images in the list (internal similarity); and information obtained from the original order and the query (external similarity). a) b) differentiate relevant and irrelevant images. From the results of the experiments we identified some cases in which a feature type provides more information than the other. Experiments were conducted using the resources of the forum ImageCLEF 2008 for the photo retrieval track. The results showed that, in the best case, the proposed method improved the MAP up to 63% compared with the original list selecting 10 images as feedback, and 23% selecting only one image as feedback when using textual features, and up to 55 % selecting 10 images as feedback and 9% when using one image as feedback using visual features only. These differences are statistically significant according to the paired t-student test at a confidence level. The best results are obtained by giving greater importance to information from neighbors, obtained from the textual or visual similarity between images. We are currently experimenting with other IRS to analyze the level of improvement of the proposed method. In summary, the main contributions of this work are: (i) a novel image re-ranking method based on a MRF that integrates a relevance feedback approach, the similarity between the images in the list and external information obtained from the original order and the query, (ii) a common representation for visual and textual features and the corresponding similarity measures. For some queries the textual or visual description does not provide relevant information about the visual content of the image, therefore it proposed as future work to include a combination of textual and visual features to exploit the advantages of both. References c) Fig. 7. First 20 images in the list obtained by the IRS (a) and the list sorted by the MRF using textual (b) or visual (c) features for query church with more than two towers. The relevant images to the query are indicated with a red dot in the upper left corner Also we proposed similarity measures based on textual and visual features that allowed to 1. Arni, T., Clough, P., Sanderson, M., & Grubinger, M. (2008). Overview of the ImageCLEFphoto 2008 Photographic Retrieval Task. 9th Cross-language evaluation forum conference on Evaluating systems for multilingual and multimodal information access (CLEF 2008), Aarhus, Denmark, Awang, D. N. F., Pehcevski, J., Thom, J. A., & Tahaghoghi, S. M.M. (2006). Combining image and structured text retrieval. Advances in XML Information Retrieval and Evaluation. Lecture Notes in Computer Science, 3977, Bay, H., Tuytelaars, T., & Van Gool, L. (2006). Surf: Speeded up robust features. 9th European Conference on Computer Vision, Graz, Austria, Berg, T. L. (2009). Finding Iconic Images. IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Miami, FL, USA, 2009, 1-8.
11 Using a Markov Random Field for Image Re-ranking Based on Visual and Textual Features Bihong, G., Bo, P., & Xiaoming, L. (2007). A Personalized Re-ranking Algorithm Based on Relevance Feedback. Advances in Web and Network Technologies, and Information Management. Lecture Notes in Computer Science, 4537, Chellappa, R., & Jain, A. (1993). Markov Random Fields: Theory and Application. Boston: Academic Press. 7. Chong, T., Yanxiang, H., Donghong, J., Guimin, L., & Zhewei, M. (2009). A Study on Pseudo Labeled Document Constructed for Document Re-ranking. International Conference on Artificial Intelligence and Computational Intelligence, Shanghai, China, Choochaiwattana, W., & Spring, M. B. (2009). Applying Social Annotations to Retrieve and Re-rank Web Resources. International Conference on Information Management and Engineering, Kuala Lumpur, Malaysia, Clough, P., & Sanderson, M. (2004). Relevance Feedback for Cross Language Image Retrieval. Advances in Information Retrieval: 26th European Conference on IR Research, ECIR Lecture Notes in Computer Science, 2997, Cui, J., Wen, F., & Tang, X. (2008). Real time google and live image search re-ranking. 16th ACM international conference on multimedia MM '08, Vancouver, Canada, Datta, R., Joshi, D., Li, J., & Wang, J. Z. (2008). Image retrieval: Ideas, influences, and trends of the new age. ACM Computer Surveys, 40 (2), Deselaers, T., Paredes, R., Vidal E., & V, E. (2008). Learning Weighted Distances for Relevance Feedback in Image Retrieval. 19 th International Conference on Pattern Recognition, Tampa, Florida, USA, Dietterich, T. G. (1998). Approximate Statistical Tests for Comparing Supervised Classification Learning Algorithms. Neural Computation, 10 (7), Escalante, H. J., Gonzalez, J. A., Hernandez, C., Lopez, A., Montes, M., Morales, E., et al. (2008). TIA-INAOE's Participation at ImageCLEF2008. Working Notes of the CLEF 2008 Workshop, Aarhus, Denmark. 15. Escalante, H. J., Hernandez, C. A., Sucar, L. E., & Montes, M. (2008). Late fusion of heterogeneous methods for multimedia image Retrieval. 1st ACM international conference on Multimedia information retrieval (MIR '08), Vancouver, Canada, Gong, Z., & Liu, Q. (2009). Improving keyword based web image search with visual feature distribution and term expansion. Knowledge and Information Systems, 21 (1), Jégou, H., Douze, M., & Schmid, C. (2010). Improving bag-of-features for large scale image search. International Journal of Computer Vision, 87 (3), Jianjiang, L., Zhenghui, X., Ran, L., Yafei, Z., & Jiabao, W. (2009). A Framework of CBIR System Based on Relevance Feedback,. Third International Symposium on Intelligent Information Technology Application (IITA 2009), Kanji, G. K. (1993). 100 statistical tests. Newbury Park, California: Sage Publications. 20. Gao, K., Lin, S., Zhang, Y., & Tang, S. (2008). Objectbased Image Retrieval with Attention Analysis and Spatial Re-ranking. IFIP Advances in Information and Communication Technology, 288, Li, S.Z. (1994). Markov random field models in computer vision. Computer Vision ECCV '94, Lecture Notes in Computer science, 801, Lin, W. H., Jin, R., & Hauptmann, A. (2003). Web Image Retrieval Re-Ranking with Relevance Model. IEEE/WIC International Conference on Web Intelligence (WI '03), Halifax, Canada, Lowe, D. G. (2004). Distinctive Image Features from Scale-Invariant Keypoints. International Journal of Computer Vision, 60 (2), Mani, I. (2001). Automatic Summarization. Amsterdam ; Philadelphia: John Benjamins Publishing Co. 25. Marakakis, A., Galatsanos, N., Likas, A., & Stafylopatis, A. (2008). Application of Relevance Feedback in Content Based Image Retrieval Using Gaussian Mixture Models: 20th IEEE International Conference on Tools with Artificial Intelligence (ICTAI '08), Dayton, Ohio, USA, Martin, B., Dirk, P., & Josiah, P. (2004). Application of Machine Learning Techniques to the Re-ranking of Search Results KI 2004: Advances in Artificial Intelligence. Lecture Notes in Computer Science, 3238, Richter, F., Romberg, S., Horster, E., & Lienhart, R. (2010). Multimodal ranking for image search on community databases. International conference on multimedia information retrieval (MIR '10), Philadelphia, Pennsylvania, USA, R. Omar Chávez He is a Master in Computational Sciences student at the National Institute of Astrophysics, Optics and Electronics, located in Puebla, Mexico. He obtained his bachelor degree in Computer Engeneering from the Universidad Tecnológica de la Mixteca. His main research interests include probalistics models, computer vision, information retrieval and information re-ranking.
12 404 R. Omar Chávez, Manuel Montes and L. Enrique Sucar Manuel Montes y Gómez Obtained his Ph.D. in Computer Science from the Computing Research Center of the National Polytechnic Institute of Mexico. Currently, he is a full-time lecturer at the Computational Sciences Department of the National Institute of Astrophysics, Optics and Electronics, located in Puebla, Mexico. His research is on automatic text processing. He is author of more than 100 international papers in the fields of information retrieval, question answering, information extraction and text mining. Dr. Montes has been visiting professor at the Polytechnic University of Valencia (Spain), at the University of Geneva (Italy), and from August 2010 he is visiting professor at the University of Alabama (USA). In addition, Dr. Montes is member of the Mexican National System of Researchers, the Mexican Society of Artificial Intelligence, the Mexican Association for Natural Language Processing and the International Web Intelligence Consortium. L. Enrique Sucar Has a Ph.D in computing from Imperial College, London, UK, 1992; a M.Sc. in electrical engineering from Stanford University, California, USA, 1982; and a B.Sc. in electronics and communications engineering from ITESM, Monterrey, México, He is currently a Senior Researcher at INAOE, Puebla, Mexico. Dr. Sucar is Member of the National Research System, the Mexican Science Academy, AAAI, SMIA and Senior Member of the IEEE. He has served as president of the Mexican AI Society, has been member of the Advisory Board of IJCAI, and is Associate Editor of the journals Computación y Sistemas and Revista Iberoamericana de Inteligencia Artificial. His main research interests are in graphical models and probabilistic reasoning, and their applications in computer vision, robotics and biomedicine.
Approaches of Using a Word-Image Ontology and an Annotated Image Corpus as Intermedia for Cross-Language Image Retrieval
Approaches of Using a Word-Image Ontology and an Annotated Image Corpus as Intermedia for Cross-Language Image Retrieval Yih-Chen Chang and Hsin-Hsi Chen Department of Computer Science and Information
MIRACLE at VideoCLEF 2008: Classification of Multilingual Speech Transcripts
MIRACLE at VideoCLEF 2008: Classification of Multilingual Speech Transcripts Julio Villena-Román 1,3, Sara Lana-Serrano 2,3 1 Universidad Carlos III de Madrid 2 Universidad Politécnica de Madrid 3 DAEDALUS
Multiscale Object-Based Classification of Satellite Images Merging Multispectral Information with Panchromatic Textural Features
Remote Sensing and Geoinformation Lena Halounová, Editor not only for Scientific Cooperation EARSeL, 2011 Multiscale Object-Based Classification of Satellite Images Merging Multispectral Information with
LABERINTO at ImageCLEF 2011 Medical Image Retrieval Task
LABERINTO at ImageCLEF 2011 Medical Image Retrieval Task Jacinto Mata, Mariano Crespo, Manuel J. Maña Dpto. de Tecnologías de la Información. Universidad de Huelva Ctra. Huelva - Palos de la Frontera s/n.
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
siftservice.com - Turning a Computer Vision algorithm into a World Wide Web Service
siftservice.com - Turning a Computer Vision algorithm into a World Wide Web Service Ahmad Pahlavan Tafti 1, Hamid Hassannia 2, and Zeyun Yu 1 1 Department of Computer Science, University of Wisconsin -Milwaukee,
Probabilistic Latent Semantic Analysis (plsa)
Probabilistic Latent Semantic Analysis (plsa) SS 2008 Bayesian Networks Multimedia Computing, Universität Augsburg [email protected] www.multimedia-computing.{de,org} References
A Genetic Algorithm-Evolved 3D Point Cloud Descriptor
A Genetic Algorithm-Evolved 3D Point Cloud Descriptor Dominik Wȩgrzyn and Luís A. Alexandre IT - Instituto de Telecomunicações Dept. of Computer Science, Univ. Beira Interior, 6200-001 Covilhã, Portugal
Blog Post Extraction Using Title Finding
Blog Post Extraction Using Title Finding Linhai Song 1, 2, Xueqi Cheng 1, Yan Guo 1, Bo Wu 1, 2, Yu Wang 1, 2 1 Institute of Computing Technology, Chinese Academy of Sciences, Beijing 2 Graduate School
A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow
, pp.233-237 http://dx.doi.org/10.14257/astl.2014.51.53 A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow Giwoo Kim 1, Hye-Youn Lim 1 and Dae-Seong Kang 1, 1 Department of electronices
Fast Matching of Binary Features
Fast Matching of Binary Features Marius Muja and David G. Lowe Laboratory for Computational Intelligence University of British Columbia, Vancouver, Canada {mariusm,lowe}@cs.ubc.ca Abstract There has been
TouchPaper - An Augmented Reality Application with Cloud-Based Image Recognition Service
TouchPaper - An Augmented Reality Application with Cloud-Based Image Recognition Service Feng Tang, Daniel R. Tretter, Qian Lin HP Laboratories HPL-2012-131R1 Keyword(s): image recognition; cloud service;
Cees Snoek. Machine. Humans. Multimedia Archives. Euvision Technologies The Netherlands. University of Amsterdam The Netherlands. Tree.
Visual search: what's next? Cees Snoek University of Amsterdam The Netherlands Euvision Technologies The Netherlands Problem statement US flag Tree Aircraft Humans Dog Smoking Building Basketball Table
SZTAKI @ ImageCLEF 2011
SZTAKI @ ImageCLEF 2011 Bálint Daróczy Róbert Pethes András A. Benczúr Data Mining and Web search Research Group, Informatics Laboratory Computer and Automation Research Institute of the Hungarian Academy
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
Categorical Data Visualization and Clustering Using Subjective Factors
Categorical Data Visualization and Clustering Using Subjective Factors Chia-Hui Chang and Zhi-Kai Ding Department of Computer Science and Information Engineering, National Central University, Chung-Li,
Florida International University - University of Miami TRECVID 2014
Florida International University - University of Miami TRECVID 2014 Miguel Gavidia 3, Tarek Sayed 1, Yilin Yan 1, Quisha Zhu 1, Mei-Ling Shyu 1, Shu-Ching Chen 2, Hsin-Yu Ha 2, Ming Ma 1, Winnie Chen 4,
DIFFERENTIATIONS OF OBJECTS IN DIFFUSE DATABASES DIFERENCIACIONES DE OBJETOS EN BASES DE DATOS DIFUSAS
Recibido: 01 de agosto de 2012 Aceptado: 23 de octubre de 2012 DIFFERENTIATIONS OF OBJECTS IN DIFFUSE DATABASES DIFERENCIACIONES DE OBJETOS EN BASES DE DATOS DIFUSAS PhD. Amaury Caballero*; PhD. Gabriel
3D Model based Object Class Detection in An Arbitrary View
3D Model based Object Class Detection in An Arbitrary View Pingkun Yan, Saad M. Khan, Mubarak Shah School of Electrical Engineering and Computer Science University of Central Florida http://www.eecs.ucf.edu/
Clustering Technique in Data Mining for Text Documents
Clustering Technique in Data Mining for Text Documents Ms.J.Sathya Priya Assistant Professor Dept Of Information Technology. Velammal Engineering College. Chennai. Ms.S.Priyadharshini Assistant Professor
Building an Advanced Invariant Real-Time Human Tracking System
UDC 004.41 Building an Advanced Invariant Real-Time Human Tracking System Fayez Idris 1, Mazen Abu_Zaher 2, Rashad J. Rasras 3, and Ibrahiem M. M. El Emary 4 1 School of Informatics and Computing, German-Jordanian
An Order-Invariant Time Series Distance Measure [Position on Recent Developments in Time Series Analysis]
An Order-Invariant Time Series Distance Measure [Position on Recent Developments in Time Series Analysis] Stephan Spiegel and Sahin Albayrak DAI-Lab, Technische Universität Berlin, Ernst-Reuter-Platz 7,
AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION
AN IMPROVED DOUBLE CODING LOCAL BINARY PATTERN ALGORITHM FOR FACE RECOGNITION Saurabh Asija 1, Rakesh Singh 2 1 Research Scholar (Computer Engineering Department), Punjabi University, Patiala. 2 Asst.
Research of Postal Data mining system based on big data
3rd International Conference on Mechatronics, Robotics and Automation (ICMRA 2015) Research of Postal Data mining system based on big data Xia Hu 1, Yanfeng Jin 1, Fan Wang 1 1 Shi Jiazhuang Post & Telecommunication
MusicGuide: Album Reviews on the Go Serdar Sali
MusicGuide: Album Reviews on the Go Serdar Sali Abstract The cameras on mobile phones have untapped potential as input devices. In this paper, we present MusicGuide, an application that can be used to
Content-Based Image Retrieval
Content-Based Image Retrieval Selim Aksoy Department of Computer Engineering Bilkent University [email protected] Image retrieval Searching a large database for images that match a query: What kind
The Visual Internet of Things System Based on Depth Camera
The Visual Internet of Things System Based on Depth Camera Xucong Zhang 1, Xiaoyun Wang and Yingmin Jia Abstract The Visual Internet of Things is an important part of information technology. It is proposed
Open issues and research trends in Content-based Image Retrieval
Open issues and research trends in Content-based Image Retrieval Raimondo Schettini DISCo Universita di Milano Bicocca [email protected] www.disco.unimib.it/schettini/ IEEE Signal Processing Society
Build Panoramas on Android Phones
Build Panoramas on Android Phones Tao Chu, Bowen Meng, Zixuan Wang Stanford University, Stanford CA Abstract The purpose of this work is to implement panorama stitching from a sequence of photos taken
A Comparative Study between SIFT- Particle and SURF-Particle Video Tracking Algorithms
A Comparative Study between SIFT- Particle and SURF-Particle Video Tracking Algorithms H. Kandil and A. Atwan Information Technology Department, Faculty of Computer and Information Sciences, Mansoura University,El-Gomhoria
Simultaneous Gamma Correction and Registration in the Frequency Domain
Simultaneous Gamma Correction and Registration in the Frequency Domain Alexander Wong [email protected] William Bishop [email protected] Department of Electrical and Computer Engineering University
An Automatic and Accurate Segmentation for High Resolution Satellite Image S.Saumya 1, D.V.Jiji Thanka Ligoshia 2
An Automatic and Accurate Segmentation for High Resolution Satellite Image S.Saumya 1, D.V.Jiji Thanka Ligoshia 2 Assistant Professor, Dept of ECE, Bethlahem Institute of Engineering, Karungal, Tamilnadu,
Susana Sanduvete-Chaves, Salvador Chacón-Moscoso, Milagrosa Sánchez- Martín y José Antonio Pérez-Gil ( )
ACCIÓN PSICOLÓGICA, junio 2014, vol. 10, n. o 2, 3-20. ISSN: 1578-908X 19 THE REVISED OSTERLIND INDEX. A COMPARATIVE ANALYSIS IN CONTENT VALIDITY STUDIES 1 EL ÍNDICE DE OSTERLIND REVISADO. UN ANÁLISIS
Domain Classification of Technical Terms Using the Web
Systems and Computers in Japan, Vol. 38, No. 14, 2007 Translated from Denshi Joho Tsushin Gakkai Ronbunshi, Vol. J89-D, No. 11, November 2006, pp. 2470 2482 Domain Classification of Technical Terms Using
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
Segmentation of building models from dense 3D point-clouds
Segmentation of building models from dense 3D point-clouds Joachim Bauer, Konrad Karner, Konrad Schindler, Andreas Klaus, Christopher Zach VRVis Research Center for Virtual Reality and Visualization, Institute
Tracking and Recognition in Sports Videos
Tracking and Recognition in Sports Videos Mustafa Teke a, Masoud Sattari b a Graduate School of Informatics, Middle East Technical University, Ankara, Turkey [email protected] b Department of Computer
Dublin City University at CLEF 2004: Experiments with the ImageCLEF St Andrew s Collection
Dublin City University at CLEF 2004: Experiments with the ImageCLEF St Andrew s Collection Gareth J. F. Jones, Declan Groves, Anna Khasin, Adenike Lam-Adesina, Bart Mellebeek. Andy Way School of Computing,
Data Mining in Web Search Engine Optimization and User Assisted Rank Results
Data Mining in Web Search Engine Optimization and User Assisted Rank Results Minky Jindal Institute of Technology and Management Gurgaon 122017, Haryana, India Nisha kharb Institute of Technology and Management
Evaluation of an exercise for measuring impact in e-learning: Case study of learning a second language
Evaluation of an exercise for measuring impact in e-learning: Case study of learning a second language J.M. Sánchez-Torres Universidad Nacional de Colombia Bogotá D.C., Colombia [email protected]
Search Result Optimization using Annotators
Search Result Optimization using Annotators Vishal A. Kamble 1, Amit B. Chougule 2 1 Department of Computer Science and Engineering, D Y Patil College of engineering, Kolhapur, Maharashtra, India 2 Professor,
A Time Efficient Algorithm for Web Log Analysis
A Time Efficient Algorithm for Web Log Analysis Santosh Shakya Anju Singh Divakar Singh Student [M.Tech.6 th sem (CSE)] Asst.Proff, Dept. of CSE BU HOD (CSE), BUIT, BUIT,BU Bhopal Barkatullah University,
Specific Usage of Visual Data Analysis Techniques
Specific Usage of Visual Data Analysis Techniques Snezana Savoska 1 and Suzana Loskovska 2 1 Faculty of Administration and Management of Information systems, Partizanska bb, 7000, Bitola, Republic of Macedonia
Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability
Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller
A Cognitive Approach to Vision for a Mobile Robot
A Cognitive Approach to Vision for a Mobile Robot D. Paul Benjamin Christopher Funk Pace University, 1 Pace Plaza, New York, New York 10038, 212-346-1012 [email protected] Damian Lyons Fordham University,
Web 3.0 image search: a World First
Web 3.0 image search: a World First The digital age has provided a virtually free worldwide digital distribution infrastructure through the internet. Many areas of commerce, government and academia have
Neovision2 Performance Evaluation Protocol
Neovision2 Performance Evaluation Protocol Version 3.0 4/16/2012 Public Release Prepared by Rajmadhan Ekambaram [email protected] Dmitry Goldgof, Ph.D. [email protected] Rangachar Kasturi, Ph.D.
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 269 Class Project Report
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 69 Class Project Report Junhua Mao and Lunbo Xu University of California, Los Angeles [email protected] and lunbo
A Survey on Product Aspect Ranking
A Survey on Product Aspect Ranking Charushila Patil 1, Prof. P. M. Chawan 2, Priyamvada Chauhan 3, Sonali Wankhede 4 M. Tech Student, Department of Computer Engineering and IT, VJTI College, Mumbai, Maharashtra,
Adaptación de MoProSoft para la producción de software en instituciones académicas
Adaptación de MoProSoft para la producción de software en instituciones académicas Adaptation of MoProSoft for software production in academic institutions Gabriela Alejandra Martínez Cárdenas Instituto
Distributed forests for MapReduce-based machine learning
Distributed forests for MapReduce-based machine learning Ryoji Wakayama, Ryuei Murata, Akisato Kimura, Takayoshi Yamashita, Yuji Yamauchi, Hironobu Fujiyoshi Chubu University, Japan. NTT Communication
Chapter 6. The stacking ensemble approach
82 This chapter proposes the stacking ensemble approach for combining different data mining classifiers to get better performance. Other combination techniques like voting, bagging etc are also described
A comparative study of two models for the seismic analysis of buildings
INGENIERÍA E INVESTIGACIÓN VOL. No., DECEMBER 0 (-) A comparative study of two models for the seismic analysis of buildings Estudio comparativo de dos modelos para análisis sísmico de edificios A. Luévanos
An Introduction to Data Mining
An Introduction to Intel Beijing [email protected] January 17, 2014 Outline 1 DW Overview What is Notable Application of Conference, Software and Applications Major Process in 2 Major Tasks in Detail
Image Segmentation and Registration
Image Segmentation and Registration Dr. Christine Tanner ([email protected]) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation
Open Access A Facial Expression Recognition Algorithm Based on Local Binary Pattern and Empirical Mode Decomposition
Send Orders for Reprints to [email protected] The Open Electrical & Electronic Engineering Journal, 2014, 8, 599-604 599 Open Access A Facial Expression Recognition Algorithm Based on Local Binary
Teaching in School of Electronic, Information and Electrical Engineering
Introduction to Teaching in School of Electronic, Information and Electrical Engineering Shanghai Jiao Tong University Outline Organization of SEIEE Faculty Enrollments Undergraduate Programs Sample Curricula
Comparing Tag Clouds, Term Histograms, and Term Lists for Enhancing Personalized Web Search
Comparing Tag Clouds, Term Histograms, and Term Lists for Enhancing Personalized Web Search Orland Hoeber and Hanze Liu Department of Computer Science, Memorial University St. John s, NL, Canada A1B 3X5
Term extraction for user profiling: evaluation by the user
Term extraction for user profiling: evaluation by the user Suzan Verberne 1, Maya Sappelli 1,2, Wessel Kraaij 1,2 1 Institute for Computing and Information Sciences, Radboud University Nijmegen 2 TNO,
Computer Forensics Application. ebay-uab Collaborative Research: Product Image Analysis for Authorship Identification
Computer Forensics Application ebay-uab Collaborative Research: Product Image Analysis for Authorship Identification Project Overview A new framework that provides additional clues extracted from images
Classifying Manipulation Primitives from Visual Data
Classifying Manipulation Primitives from Visual Data Sandy Huang and Dylan Hadfield-Menell Abstract One approach to learning from demonstrations in robotics is to make use of a classifier to predict if
A Framework of User-Driven Data Analytics in the Cloud for Course Management
A Framework of User-Driven Data Analytics in the Cloud for Course Management Jie ZHANG 1, William Chandra TJHI 2, Bu Sung LEE 1, Kee Khoon LEE 2, Julita VASSILEVA 3 & Chee Kit LOOI 4 1 School of Computer
Recognition. Sanja Fidler CSC420: Intro to Image Understanding 1 / 28
Recognition Topics that we will try to cover: Indexing for fast retrieval (we still owe this one) History of recognition techniques Object classification Bag-of-words Spatial pyramids Neural Networks Object
Document Image Retrieval using Signatures as Queries
Document Image Retrieval using Signatures as Queries Sargur N. Srihari, Shravya Shetty, Siyuan Chen, Harish Srinivasan, Chen Huang CEDAR, University at Buffalo(SUNY) Amherst, New York 14228 Gady Agam and
Signature Segmentation from Machine Printed Documents using Conditional Random Field
2011 International Conference on Document Analysis and Recognition Signature Segmentation from Machine Printed Documents using Conditional Random Field Ranju Mandal Computer Vision and Pattern Recognition
Local features and matching. Image classification & object localization
Overview Instance level search Local features and matching Efficient visual recognition Image classification & object localization Category recognition Image classification: assigning a class label to
Offline sorting buffers on Line
Offline sorting buffers on Line Rohit Khandekar 1 and Vinayaka Pandit 2 1 University of Waterloo, ON, Canada. email: [email protected] 2 IBM India Research Lab, New Delhi. email: [email protected]
Android Ros Application
Android Ros Application Advanced Practical course : Sensor-enabled Intelligent Environments 2011/2012 Presentation by: Rim Zahir Supervisor: Dejan Pangercic SIFT Matching Objects Android Camera Topic :
Face Recognition in Low-resolution Images by Using Local Zernike Moments
Proceedings of the International Conference on Machine Vision and Machine Learning Prague, Czech Republic, August14-15, 014 Paper No. 15 Face Recognition in Low-resolution Images by Using Local Zernie
Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall
Automatic Photo Quality Assessment Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Estimating i the photorealism of images: Distinguishing i i paintings from photographs h Florin
A Dynamic Approach to Extract Texts and Captions from Videos
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 4, April 2014,
Path Selection Methods for Localized Quality of Service Routing
Path Selection Methods for Localized Quality of Service Routing Xin Yuan and Arif Saifee Department of Computer Science, Florida State University, Tallahassee, FL Abstract Localized Quality of Service
MapReduce Approach to Collective Classification for Networks
MapReduce Approach to Collective Classification for Networks Wojciech Indyk 1, Tomasz Kajdanowicz 1, Przemyslaw Kazienko 1, and Slawomir Plamowski 1 Wroclaw University of Technology, Wroclaw, Poland Faculty
Implementation of kalman filter for the indoor location system of a lego nxt mobile robot. Abstract
Implementation of kalman filter for the indoor location system of a lego nxt mobile robot Leidy López Osorio * Giovanni Bermúdez Bohórquez ** Miguel Pérez Pereira *** submitted date: March 2013 received
Design call center management system of e-commerce based on BP neural network and multifractal
Available online www.jocpr.com Journal of Chemical and Pharmaceutical Research, 2014, 6(6):951-956 Research Article ISSN : 0975-7384 CODEN(USA) : JCPRC5 Design call center management system of e-commerce
How To Write A Book On The Digital Age Of Science
New Trends in Digital University Libraries Peter Schirmbacher * Introduction When speaking about trends in university digital libraries it is necessary to concentrate on the main points and discuss only
Frsq: A Binary Image Coding Method
Frsq: A Binary Image Coding Method Peter L. Stanchev, William I. Grosky, John G. Geske Kettering University, Flint, MI 4854, {pstanche, jgeske}@kettering.edu University of Michigan-Dearborn, Dearborn,
Recognizing Cats and Dogs with Shape and Appearance based Models. Group Member: Chu Wang, Landu Jiang
Recognizing Cats and Dogs with Shape and Appearance based Models Group Member: Chu Wang, Landu Jiang Abstract Recognizing cats and dogs from images is a challenging competition raised by Kaggle platform
Introducing diversity among the models of multi-label classification ensemble
Introducing diversity among the models of multi-label classification ensemble Lena Chekina, Lior Rokach and Bracha Shapira Ben-Gurion University of the Negev Dept. of Information Systems Engineering and
UNSUPERVISED COSEGMENTATION BASED ON SUPERPIXEL MATCHING AND FASTGRABCUT. Hongkai Yu and Xiaojun Qi
UNSUPERVISED COSEGMENTATION BASED ON SUPERPIXEL MATCHING AND FASTGRABCUT Hongkai Yu and Xiaojun Qi Department of Computer Science, Utah State University, Logan, UT 84322-4205 [email protected]
Part-Based Recognition
Part-Based Recognition Benedict Brown CS597D, Fall 2003 Princeton University CS 597D, Part-Based Recognition p. 1/32 Introduction Many objects are made up of parts It s presumably easier to identify simple
Mining Text Data: An Introduction
Bölüm 10. Metin ve WEB Madenciliği http://ceng.gazi.edu.tr/~ozdemir Mining Text Data: An Introduction Data Mining / Knowledge Discovery Structured Data Multimedia Free Text Hypertext HomeLoan ( Frank Rizzo
FACE RECOGNITION BASED ATTENDANCE MARKING SYSTEM
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 2, February 2014,
Optimization of Image Search from Photo Sharing Websites Using Personal Data
Optimization of Image Search from Photo Sharing Websites Using Personal Data Mr. Naeem Naik Walchand Institute of Technology, Solapur, India Abstract The present research aims at optimizing the image search
An Approach for Utility Pole Recognition in Real Conditions
6th Pacific-Rim Symposium on Image and Video Technology 1st PSIVT Workshop on Quality Assessment and Control by Image and Video Analysis An Approach for Utility Pole Recognition in Real Conditions Barranco
Statistical Models in Data Mining
Statistical Models in Data Mining Sargur N. Srihari University at Buffalo The State University of New York Department of Computer Science and Engineering Department of Biostatistics 1 Srihari Flood of
STATISTICA. Financial Institutions. Case Study: Credit Scoring. and
Financial Institutions and STATISTICA Case Study: Credit Scoring STATISTICA Solutions for Business Intelligence, Data Mining, Quality Control, and Web-based Analytics Table of Contents INTRODUCTION: WHAT
Large-Scale Data Sets Clustering Based on MapReduce and Hadoop
Journal of Computational Information Systems 7: 16 (2011) 5956-5963 Available at http://www.jofcis.com Large-Scale Data Sets Clustering Based on MapReduce and Hadoop Ping ZHOU, Jingsheng LEI, Wenjun YE
Guidelines for Designing Web Maps - An Academic Experience
Guidelines for Designing Web Maps - An Academic Experience Luz Angela ROCHA SALAMANCA, Colombia Key words: web map, map production, GIS on line, visualization, web cartography SUMMARY Nowadays Internet
Hyperspectral images retrieval with Support Vector Machines (SVM)
Hyperspectral images retrieval with Support Vector Machines (SVM) Miguel A. Veganzones Grupo Inteligencia Computacional Universidad del País Vasco (Grupo Inteligencia SVM-retrieval Computacional Universidad
