FREAK: Fast Retina Keypoint
|
|
|
- Rosalind Harvey
- 10 years ago
- Views:
Transcription
1 FREAK: Fast Retina Keypoint Alexandre Alahi, Raphael Ortiz, Pierre Vandergheynst Ecole Polytechnique Fédérale de Lausanne (EPFL), Switzerland Abstract A large number of vision applications rely on matching keypoints across images. The last decade featured an arms-race towards faster and more robust keypoints and association algorithms: Scale Invariant Feature Transform (SIFT)[17], Speed-up Robust Feature (SURF)[4], and more recently Binary Robust Invariant Scalable Keypoints (BRISK)[16] to name a few. These days, the deployment of vision algorithms on smart phones and embedded devices with low memory and computation complexity has even upped the ante: the goal is to make descriptors faster to compute, more compact while remaining robust to scale, rotation and noise. To best address the current requirements, we propose a novel keypoint descriptor inspired by the human visual system and more precisely the retina, coined Fast Retina Keypoint (FREAK). A cascade of binary strings is computed by efficiently comparing image intensities over a retinal sampling pattern. Our experiments show that FREAKs are in general faster to compute with lower memory load and also more robust than SIFT, SURF or BRISK. They are thus competitive alternatives to existing keypoints in particular for embedded applications. 1. Introduction Visual correspondence, object matching, and many other vision applications rely on representing images with sparse number of keypoints. A real challenge is to efficiently describe keypoints, i.e. image patches, with stable, compact and robust representations invariant to scale, rotation, affine transformation, and noise. The past decades witnessed key players to efficiently describe keypoints and match them. The most popular descriptor is the histogram of oriented gradient proposed by Lowe [17] to describe the Scale Invariant Feature Transform (SIFT) keypoints. Most of the efforts in the last years was to perform as good as SIFT [14] with lower computational complexity. The Speeded up Robust Feature (SURF) by Bay et al. [4] is a good example. It has similar matching rates with much faster performance Figure 1: llustration of our FREAK descriptor. A series of Difference of Gaussians (DoG) over a retinal pattern are 1 bit quantized. by describing keypoints with the responses of few Haar-like filters. In general, Alahi et al. show in [2] that a grid of descriptors, similar to SIFT and SURF, is better than a single one to match an image region. Typically, a grid of covariance matrices [30] attains high detection rate but remains computationally too expensive for real-time applications. The deployment of cameras on every phone coupled with the growing computing power of mobile devices has enabled a new trend: vision algorithms need to run on mobile devices with low computing power and memory capacity. Images obtained by smart phones can be used to perform structure from motion [27], image retrieval [22], or object recognition [15]. As a result, new algorithms are needed where fixed-point operations and low memory load are preferred. The Binary Robust Independent Elementary Feature (BRIEF) [5], the Oriented Fast and Rotated BRIEF (ORB)[26], and the Binary Robust Invariant Scalable Keypoints[16] (BRISK) are good examples. In the next section, we will briefly present these descriptors. Their stimulating contribution is that a binary string obtained by simply comparing pairs of image intensities can efficiently describe a keypoint, i.e. an image patch. However, several problems remain: how to efficiently select the ideal pairs within an image patch? How to match them? Interestingly, such trend is inline with the models of the nature to describe complex observations with simple rules. We propose to address such unknowns by designing a descriptor inspired by the Human Visual System, and more precisely the retina. We propose the Fast Retina Keypoint (FREAK) as a fast, 1
2 compact and robust keypoint descriptor. A cascade of binary strings is computed by efficiently comparing pairs of image intensities over a retinal sampling pattern. Interestingly, selecting pairs to reduce the dimensionality of the descriptor yields a highly structured pattern that mimics the saccadic search of the human eyes. 2. Related work Keypoint descriptors are often coupled with their detection. Tuytelaar et al. in [29] and Gauglitz et al. in [11] presented a detailed survey. We briefly present state-of-the-art detectors and mainly focus on descriptors Keypoint detectors A first solution is to consider corners as keypoints. Harris and Stephen in [12] proposed the Harris corner detector. Mikolajczyk and Schmid made it scale invariant in [20]. Another solution is to use local extrema of the responses of certain filters as potential keypoints. Lowe in [17] filtered the image with differences of Gaussians. Bay et al. in [4] used a Fast Hessian detector. Agrawal et al. in [1] proposed simplified center-surround filters to approximate the Laplacian.. Ebrahimi and Mayol-Cuevas in [7] accelerated the process by skipping the computation of the filter response if the response for the previous pixel is very low. Rostenand and Drummond proposed in [25] the FAST criterion for corner detection, improved by Mair et al. in [18] with their AGAST detector. The latter is a fast algorithm to locate keypoints. The detector used in BRISK by Leutenegger et al. in [16] is a multi-scale AGAST. They search for maxima in scale-space using the FAST score as a measure of saliency. We use the same detector for our evaluation of FREAK SIFT-like descriptors Once keypoints are located, we are interested in describing the image patch with a robust feature vector. The most well-known descriptor is SIFT [17]. A 128-dimensional vector is obtained from a grid of histograms of oriented gradient. Its high descriptive power and robustness to illumination change have ranked it as the reference keypoint descriptor for the past decade. A family of SIFT-like descriptor has emerged in the past years. The PCA-SIFT [14] reduces the description vector from 128 to 36 dimension using principal component analysis. The matching time is reduced, but the time to build the descriptor is increased leading to a small gain in speed and a loss of distinctiveness. The GLOH descriptor [21] is an extension of the SIFT descriptor that is more distinctive, but also more expensive to compute. The robustness to change of viewpoint is improved in [31] by simulating multiple deformations to the descriptive patch. Good compromises between performances and the number of simulated patches lead to an algorithm twice slower than SIFT. Ambai and Yoshida proposed a Compact And Realtime Descriptors (CARD) in [3] to extract the histogram of oriented gradient from the grid binning of SIFT or the logpolar binning of GLOH. The computation of the histograms is simplified by using lookup tables. One of the widely used keypoints at the moment is clearly SURF [4]. It has similar matching performances as SIFT but is much faster. It also relies on local gradient histograms. The Haar-wavelet responses are efficiently computed with integral images leading to 64 or 128-dimensional vectors. However, the dimensionality of the feature vector is still too high for large-scale applications such as image retrieval or 3D reconstruction. Often, Principal Component Analysis (PCA), or hashing functions are used to reduce the dimensionality of the descriptors [24]. Such steps involve time-consuming computation and hence affect the real-time performance Binary descriptors Calonder et al. in [5] showed that it is possible to shortcut the dimensionality reduction step by directly building a short binary descriptor in which each bits are independent, called BRIEF. A clear advantage of binary descriptors is that the Hamming distance (bitwise XOR followed by a bit count) can replace the usual Euclidean distance. The descriptor vector is obtained by comparing the intensity of 512 pairs of pixels after applying a Gaussian smoothing to reduce the noise sensitivity. The positions of the pixels are pre-selected randomly according to a Gaussian distribution around the patch center. The obtained descriptor is not invariant to scale and rotation changes unless coupled with detector providing it. Calonder et al. also highlighted in their work that usually orientation detection reduces the recognition rate and should therefore be avoided when it is not required by the target application. Rublee et al. in [26] proposed the Oriented Fast and Rotated BRIEF (ORB) descriptor. Their binary descriptor is invariant to rotation and robust to noise. Similarly, Leutenegger et al. in [16] proposed a binary descriptor invariant to scale and rotation called BRISK. To build the descriptor bit-stream, a limited number of points in a specific sampling pattern is used. Each point contributes to many pairs. The pairs are divided in short-distance and long-distance subsets. The long-distance subset is used to estimate the direction of the keypoint while the short-distance subset is used to build binary descriptor after rotating the sampling pattern. In Section 5, we compare our proposed FREAK descriptor with the above presented descriptors. But first, we present a possible intuition on why these trendy binary descriptors can work based on the study of the human retina.
3 / "?4::7" ;<=>:81="?4::7" FGH<="" I8781=" /894:7"!" # $ ""!" # % ""!" # & ""!" # ' ""!" # * ""!" # + ""!" #, "" D8=4<3" (" )" - $.$$..$" (" )" E1=-:8=4<3" B8=<3C"7238=>" (a) Density of ganglion cells over the retina [10]. (b) Retina areas [13] Figure 3: Illustration of the distribution of ganglion cells over the retina. The density is clustered into four areas: (a) the foveola, (b) fovea, (c) parafoveal, and (d) perifoveal. Figure 2: From human retina to computer vision: the biological pathways leading to action potentials is emulated by simple binary tests over pixel regions. [Upper part of the image is a courtesy of the book Avian Visual Cognition by R: Cook]. 3. Human retina 3.1. Motivations In the presented literature, we have seen that recent progress in image representation has shown that simple intensity comparison of several pairs of pixels can be good enough to describe and match image patches [5, 26, 16]. However, there exist some open interrogations on the ideal selection of pairs. How should we sample them and compare them? How to be robust to noise? Should we smooth with a single Gaussian kernel? In this work, we show how to gain performance by selecting a solution inspired by the human retina, while enforcing low computational complexity. Neuroscience has made lots of progress in understanding the visual system and how images are transmitted to the brain [8]. It is believed that the human retina extracts details from images using Difference of Gaussians (DoG) of various sizes and encodes such differences with action potentials. The topology of the retina plays an important role. We propose to mimic the same strategy to design our image descriptor Analogy: from retinal photoreceptors to pixels The topology and spatial encoding of the retina is quite fascinating. First, several photoreceptors influence a ganglion cell. The region where light influences the response of a ganglion cell is the receptive field. Its size and dendritic field increases with radial distance from the foveola (Figure 3). The spatial distribution of ganglion cells reduces exponentially with the distance to the foveal. They are segmented into four areas: foveal, fovea, parafoveal, and perifoveal. Each area plays an interesting role in the process of detecting and recognizing objects since higher resolution is captured in the fovea whereas a low acuity image is formed in the perifoveal. One can interpret the decrease of resolution as a body resource optimization. Let us now turn these insights into an actual keypoint descriptor. Figure 2 presents the proposed analogy. 4. FREAK 4.1. Retinal sampling pattern Many sampling grids are possible to compare pairs of pixel intensities. BRIEF and ORB use random pairs. BRISK uses a circular pattern where points are equally spaced on circles concentric, similar to DAISY [28]. We propose to use the retinal sampling grid which is also circular with the difference of having higher density of points near the center. The density of points drops exponentially as can be seen in Figure 3.. Each sample point needs to be smoothed to be less sensitive to noise. BRIEF and ORB use the same kernel for all points in the patch. To match the retina model, we use different kernels size for every sample points similar to BRISK. The difference with BRISK is the exponential change in size and the overlapping receptive fields. Figure 4 illustrates the topology of the receptive fields. Each circle represents the standard deviations of the Gaussian kernels applied to the corresponding sampling points. We have experimentally observed that changing the size of the Gaussian kernels with respect to the log-polar retinal pattern leads to better performance. In addition, overlapping the receptive fields also increases the performance. A possible reason is that with the presented overlap in Figure 4, more information is captured. We add redundancy that brings more discriminative power. Let s consider the intensities I i measured at the receptive fields A, B, and C where: I A > I B, I B > I C, and I A > I C. (1) If the fields do not have overlap, then the last test I A > I C is not adding any discriminant information. However,
4 Figure 4. We use 43 receptive fields leading to approximately one thousand pairs. 2. We compute the mean of each column. In order to have a discriminant feature, high variance is desired. A mean of 0.5 leads to the highest variance of a binary distribution. 3. We order the columns with respect to the highest variance. Figure 4: Illustration of the FREAK sampling pattern similar to the retinal ganglion cells distribution with their corresponding receptive fields. Each circle represents a receptive field where the image is smoothed with its corresponding Gaussian kernel. if the fields overlap, partially new information can be encoded. In general, adding redundancy allow us to use less receptive fields which is a known strategy employed in compressed sensing or dictionary learning [6]. According to Olshausen and Field in [23], such redundancy also exists in the receptive fields of the retina Coarse-to-fine descriptor We construct our binary descriptor F by thresholding the difference between pairs of receptive fields with their corresponding Gaussian kernel. In other words, F is a binary string formed by a sequence of one-bit Difference of Gaussians (DoG): F = 2 a T (P a ), (2) 0 a<n where P a is a pair of receptive fields, N is the desired size of the descriptor, and T (P a ) = { 1 if (I(P r 1 a ) I(P r2 a ) > 0, 0 otherwise, with I(P r1 a ) is the smoothed intensity of the first receptive field of the pair P a. With few dozen of receptive fields, thousands of pairs are possible leading to a large descriptor. However, many of the pairs might not be useful to efficiently describe an image. A possible strategy can be to select pairs given their spatial distance similar to BRISK. However, the selected pairs can be highly correlated and not discriminant. Consequently, we run an algorithm similar to ORB [26] to learn the best pairs from training data: 1. We create a matrix D of nearly fifty thousands of extracted keypoints. Each row corresponds to a keypoint represented with its large descriptor made of all possible pairs in the retina sampling pattern illustrated in 4. We keep the best column (mean of 0.5) and iteratively add remaining columns having low correlation with the selected columns. Strikingly, there is a structure in the selected pairs. A coarse-to-fine ordering of the difference of Gaussians (the pairs) is automatically prefered. Figure 5 illustrates the pairs selected by grouping them into four clusters (128 pairs per group). We experimentally observed that the first 512 pairs are the most relevant and adding more pairs is not increasing the performance. A symmetric scheme is captured due to the orientation of the pattern along the global gradient. Interestingly, the first cluster involves mainly peripheral receptive fields whereas the last ones implicates highly centered fields. It appears to be reminiscent of the behavior of the human eye. We first use the perifoveal receptive fields to estimate the location of an object of interest. Then, the validation is performed with the more densely distributed receptive fields in the fovea area. Although the used feature selection algorithm is a heuristic, it seems to match our understanding of the model of the human retina. Our matching step takes advantage of the coarse-to-fine structured of FREAK descriptor. Note that in the last decades, coarse-to-fine strategy has often been explored in detecting and matching objects [9, 2] Saccadic search Humans do not look at a scene in fixed steadiness. Their eyes move around with discontinuous individual movements called saccades. The presented cells topology in the retina is one reason for such movements. As explained previously, the fovea captures high-resolution information thanks to its high-density photoreceptors. Hence, it plays a critical role in recognizing and matching objects. The perifoveal area captures less detailed information, i.e. lowfrequency observation. Consequently, they are used to compile first estimates of the locations of our objects of interest. We propose to mimic the saccadic search by parsing our descriptor in several steps. We start by searching with the first 16 bytes of the FREAK descriptor representing coarse information. If the distance is smaller than a threshold, we further continue the comparison with the next bytes to analyze finer information. As a result, a cascade of comparisons is performed accelerating even further the matching
5 (a) An object of interest de- (b) Best matches after 1st (c) Distance map with first (d) Distance map with last (e) Matched objects after last scribed by 1 FREAK cascade cascade (in color jet) cascade (in color jet) CHAPTER 4. INTEREST cascade POINT DESCRIPTOR We illustrate the saccadic search in Figure 4.6. For visualization purposes, we describe an object of interest with a single FREAK descriptor of the size of its bounding circle (Figure 4.6 (a)). Then, we search for the same object in a new image. All candidate image regions (a regular grid of key-points at 3 different scales) are also described with a single descriptor of the size of the candidate region. The first cascade (first 16 bytes) discards many candidates and select very few of them to compare with the remaining bytes. In Figure 4.6 (e), the last cascade has correctly selected the locations of our object of interest despite the changes of illuminations and viewpoints. Figure 6: Illustration of the cascade approach Orientation assignment 4.4. Orientation In most detector/descriptor frameworks, to achieve invariance to rotation, an orientation is assigned to each image patch at an intermediate stage between the detection and description steps. In our case, the orientation can be estimated from the FREAK In order to estimate rotation of our we sum pattern. The orientation isthe determined by the local gradients estimatedkeypoint, similarly to BRISK [2] from a selection of pairs. the estimated local gradients over selected pairs similar to p p 1 G= (I(p, σ ) I(p, σ )), (4.3) N p p BRISK [16].whereThe latter is using long pairs to compute the N is the number of pairs and p is the 2D vector of the spatial coordinates of the center of receptive field r. The key for good orientation estimation is the rotational global orientation whereas we mainly select pairs with symsymmetry of the pattern and a wise choice of pairs. Mainly pairs of points symmetric with respect to the center of the pattern are selected. This condition has to be particumetric receptive fields with respect tointhe larly enforced when the distance between the points is short. the case center of BRISK, the (see Figure pattern is not perfectly symmetric and only long pairs are used. Those defects are coped by using a considerable number of pairs. BRISK is using about 870 pairs while 7. in FREAK, the orientation is estimated from 45 quality pairs as depicted on Figure i, j<n rj j ri i rj ri rj ri ri i Figure 4.7: Illustration of the pairs used to estimate the orientation with the FREAK pattern Figure 7: Illustration of the pairs selected to computed the orientation. In addition, our retinal pattern has larger receptive fields in the perifoveal area than Figure 5: Illustration of the coarse-to-fine analysis. The first cluster involves mainly perifoveal receptive fields and the last ones fovea. 35/64 Let G be the set of all the pairs used to compute the local gradients: O= step. More than 90% of the candidates are discarded with the first 16 bytes of our FREAK descriptor. Note that we have chosen 16 bytes for the first cascade to match hardware requirements. To compare 1 byte or 16 bytes is almost equivalent with Single Instruction and Multiple Data (SIMD) instructions on Intel processors since operations are performed in parallel. We illustrate the saccadic search in Figure 6. For visualization purposes, we describe an object of interest with a single FREAK descriptor of the size of its bounding circle (Figure 6 (a)). Then, we search for the same object in a new image. All candidate image regions are also described with a single descriptor of the size of the candidate region. The first cascade (first 16 bytes) discards many candidates and select very few of them to compare with the remaining bytes. In Figure 6 (e), the last cascade has correctly selected the locations of our object of interest despite the changes of illuminations and viewpoints. 1 X P r1 Por2, (I(Por1 ) I(Por2 )) or1 M kpo Por2 k (3) Po G where M is the number of pairs in G and Pori is the 2D vector of the spatial coordinates of the center of receptive field. We select 45 pairs as opposed to the few hundreds pairs of BRISK. In addition, our retinal pattern has larger receptive fields in the perifoveal area than BRISK allowing more error in the orientation estimation. We therefore discretize the space of orientations in much bigger steps leading to more than 5 times smaller memory load (about 7 MB against 40 MB). 5. Performance evaluation The evaluation of keypoints has been exhaustively studied in the literature. Two testing environments are used: first, the well-known dataset introduced by Mikolajczyk and Schmid [21]. We present the recall (number of correct matches / number of correspondences) vs 1-precision curve
6 Time per keypoint SIFT SURF BRISK FREAK Description in [ms] Matching time in [ns] Table 1: Computation time on 800x600 images where approximately 1500 keypoints are detected per image. The computation times correspond to the description and matching of all keypoints. (number of false matches / number of matches). We also evaluate FREAK on a more task specific framework similar to the one proposed on-line 1. The graf1 image from the previous dataset is transformed to precisely evaluate the rotation, change of scale, change of viewpoint, blur (gaussian), and brightness change. We can continuously measure the impact of an image deformation, which is less visible in the first testing environment. Figures 8 to 10 present the quantitative results. The performances of the descriptors are highly related to the combination detector/descriptor. Some descriptors are more discriminant for blobs than corners. Nevertheless, we noticed that the global ranking of their matching performance remain the same regardless of the selected detector. As a result, we present our tests using the multi-scale AGAST detector introduced by BRISK. Both testing environments rank FREAK as the most robust to all the tested image deformation. Surprisingly, SIFT is the worst descriptor in the first testing environment (8) similar to what has been shown in BRISK [16]. SIFT extracts several descriptors when the estimated orientation has multiple good candidates. Therefore, the number of possible correspondences is over estimated. Nevertheless, the second testing environment (9) is not affected by the additional descriptors created with SIFT. Consequently, SIFT is more competitive and often ranks second. Figure 10 present the performance of descriptors that are not invariant to rotation and scale invariant. We compare FREAK while disabling its scale and orientation invariance. It is either as good as other descriptor or slightly better. Table 1 compares the computation time of the scale and rotation invariant descriptors. All algorithms are running on an Intel Duo core of 2.2 GHZ using a single core. FREAK is even faster than BRISK although the latter is two orders of magnitude faster than SIFT and SURF. 6. Conclusions We have presented a retina-inspired keypoint descriptor to enhance the performance of current image descriptors. It outperforms recent state-of-the-art keypoint descriptors while remaining simple (faster with lower memory load). We do not claim any biological significance but find it remarkable that the used learning stage to identify the most 1 feature-descriptor-comparison-report/ Figure 8: Performance evaluation on the dataset introduced by Mikolajczyk and Schmid [21]. relevant Difference of Gaussians could match one possible understanding of the resource optimization of the human visual system. In fact, as a future work, we want to investigate more on the selection of such relevant pairs for high level applications such as object recognition. References [1] M. Agrawal, K. Konolige, and M. Blas. Censure: Center surround extremas for realtime feature detection and matching. Computer Vision ECCV 2008, pages , [2] A. Alahi, P. Vandergheynst, M. Bierlaire, and M. Kunt. Cascade of descriptors to detect and track objects across any net-
7 Figure 9: Global evaluation on the second testing environment. The view point changes are simulated by an affine matrix transformation where the shear parameter goes from to 0.65 in both x and y directions. work of cameras. Computer Vision and Image Understanding, 114(6): , , 4 [3] M. Ambai and Y. Yoshida. Card: Compact and real-time descriptors. In Computer Vision, 2011 IEEE Computer Society Conference on. IEEE, [4] H. Bay, T. Tuytelaars, and L. Van Gool. Surf: Speeded up robust features. Computer Vision ECCV 2006, pages , , 2 [5] M. Calonder, V. Lepetit, C. Strecha, and P. Fua. Brief: Binary robust independent elementary features. Computer Vision ECCV 2010, pages , , 2, 3 [6] E. Candes, Y. Eldar, D. Needell, and P. Randall. Compressed sensing with coherent and redundant dictionaries. Applied and Computational Harmonic Analysis, 31(1):59 73, [7] M. Ebrahimi and W. Mayol-Cuevas. Susure: Speeded up surround extrema feature detector and descriptor for realtime applications [8] G. Field, J. Gauthier, A. Sher, M. Greschner, T. Machado, L. Jepson, J. Shlens, D. Gunning, K. Mathieson, W. Dabrowski, et al. Functional connectivity in the retina at the resolution of photoreceptors. Nature, 467(7316): , [9] F. Fleuret and D. Geman. Coarse-to-fine face detection. International Journal of Computer Vision, 41(1):85 107, [10] D. Garway-Heath, J. Caprioli, F. Fitzke, and R. Hitchings. Scaling the hill of vision: the physiological relationship between light sensitivity and ganglion cell numbers. Investigative Ophthalmology & Visual Science, 41(7): , [11] S. Gauglitz, T. Höllerer, and M. Turk. Evaluation of interest point detectors and feature descriptors for visual tracking. International Journal of Computer Vision, pages 1 26, [12] C. Harris and M. Stephens. A combined corner and edge detector. In Alvey vision conference, volume 15, page 50. Manchester, UK, [13] M. Hogan and J. JA Weddell. Histology of the human eye: an atlas and textbook [14] Y. Ke and R. Sukthankar. Pca-sift: A more distinctive representation for local image descriptors , 2 [15] S. Lazebnik, C. Schmid, and J. Ponce. Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories. In Computer Vision and Pattern Recognition, 2006 IEEE Computer Society Conference on, volume 2, pages Ieee, [16] S. Leutenegger, M. Chli, and R. Siegwart. Brisk: Binary robust invariant scalable keypoints , 2, 3, 5, 6 [17] D. Lowe. Object recognition from local scale-invariant features. In Computer Vision, The Proceedings of the Seventh IEEE International Conference on, volume 2, pages Ieee, , 2
8 Figure 10: Global evaluation on the second testing environment: (S: single scale; U: unrotated) BRIEF64, S-BRISK, SU-BRISK, S-FREAK, SU-FREAK descriptors are extracted on keypoints detected by FAST. [18] E. Mair, G. Hager, D. Burschka, M. Suppa, and G. Hirzinger. Adaptive and generic corner detection based on the accelerated segment test. Computer Vision ECCV 2010, pages , [19] E. Marc and H. Stuart. Lateral information processing by spiking neurons: A theoretical model of the neural correlate of consciousness. Computational Intelligence and Neuroscience, 2011, [20] K. Mikolajczyk and C. Schmid. Indexing based on scale invariant interest points [21] K. Mikolajczyk and C. Schmid. A performance evaluation of local descriptors. IEEE transactions on pattern analysis and machine intelligence, pages , , 5, 6 [22] D. Nister and H. Stewenius. Scalable recognition with a vocabulary tree. In Computer Vision and Pattern Recognition, 2006 IEEE Computer Society Conference on, volume 2, pages IEEE, [23] B. Olshausen and D. Field. What is the other 85% of v1 doing. Problems in Systems Neuroscience, pages , [24] J. Philbin, O. Chum, M. Isard, J. Sivic, and A. Zisserman. Object retrieval with large vocabularies and fast spatial matching. In 2007 IEEE Conference on Computer Vision and Pattern Recognition, pages 1 8. IEEE, [25] E. Rosten and T. Drummond. Machine learning for highspeed corner detection. Computer Vision ECCV 2006, pages , [26] E. Rublee, V. Rabaud, K. Konolige, and G. Bradski. Orb: an efficient alternative to sift or surf , 2, 3, 4 [27] N. Snavely, S. Seitz, and R. Szeliski. Photo tourism: exploring photo collections in 3d. In ACM Transactions on Graphics (TOG), volume 25, pages ACM, [28] E. Tola, V. Lepetit, and P. Fua. Daisy: An efficient dense descriptor applied to wide-baseline stereo. Pattern Analysis and Machine Intelligence, IEEE Transactions on, 32(5): , [29] T. Tuytelaars and K. Mikolajczyk. Local invariant feature detectors: a survey. Foundations and Trends in Computer Graphics and Vision, 3(3): , [30] O. Tuzel, F. Porikli, and P. Meer. Region covariance: A fast descriptor for detection and classification. Computer Vision ECCV 2006, pages , [31] G. Yu and J. Morel. A fully affine invariant image comparison method. In Acoustics, Speech and Signal Processing, ICASSP IEEE International Conference on, pages IEEE,
Fast Matching of Binary Features
Fast Matching of Binary Features Marius Muja and David G. Lowe Laboratory for Computational Intelligence University of British Columbia, Vancouver, Canada {mariusm,lowe}@cs.ubc.ca Abstract There has been
BRIEF: Binary Robust Independent Elementary Features
BRIEF: Binary Robust Independent Elementary Features Michael Calonder, Vincent Lepetit, Christoph Strecha, and Pascal Fua CVLab, EPFL, Lausanne, Switzerland e-mail: [email protected] Abstract.
CODING MODE DECISION ALGORITHM FOR BINARY DESCRIPTOR CODING
CODING MODE DECISION ALGORITHM FOR BINARY DESCRIPTOR CODING Pedro Monteiro and João Ascenso Instituto Superior Técnico - Instituto de Telecomunicações ABSTRACT In visual sensor networks, local feature
A Comparison of Keypoint Descriptors in the Context of Pedestrian Detection: FREAK vs. SURF vs. BRISK
A Comparison of Keypoint Descriptors in the Context of Pedestrian Detection: FREAK vs. SURF vs. BRISK Cameron Schaeffer Stanford University CS Department [email protected] Abstract The subject of keypoint
A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow
, pp.233-237 http://dx.doi.org/10.14257/astl.2014.51.53 A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow Giwoo Kim 1, Hye-Youn Lim 1 and Dae-Seong Kang 1, 1 Department of electronices
Randomized Trees for Real-Time Keypoint Recognition
Randomized Trees for Real-Time Keypoint Recognition Vincent Lepetit Pascal Lagger Pascal Fua Computer Vision Laboratory École Polytechnique Fédérale de Lausanne (EPFL) 1015 Lausanne, Switzerland Email:
siftservice.com - Turning a Computer Vision algorithm into a World Wide Web Service
siftservice.com - Turning a Computer Vision algorithm into a World Wide Web Service Ahmad Pahlavan Tafti 1, Hamid Hassannia 2, and Zeyun Yu 1 1 Department of Computer Science, University of Wisconsin -Milwaukee,
TouchPaper - An Augmented Reality Application with Cloud-Based Image Recognition Service
TouchPaper - An Augmented Reality Application with Cloud-Based Image Recognition Service Feng Tang, Daniel R. Tretter, Qian Lin HP Laboratories HPL-2012-131R1 Keyword(s): image recognition; cloud service;
Augmented Reality Tic-Tac-Toe
Augmented Reality Tic-Tac-Toe Joe Maguire, David Saltzman Department of Electrical Engineering [email protected], [email protected] Abstract: This project implements an augmented reality version
The use of computer vision technologies to augment human monitoring of secure computing facilities
The use of computer vision technologies to augment human monitoring of secure computing facilities Marius Potgieter School of Information and Communication Technology Nelson Mandela Metropolitan University
Recognition. Sanja Fidler CSC420: Intro to Image Understanding 1 / 28
Recognition Topics that we will try to cover: Indexing for fast retrieval (we still owe this one) History of recognition techniques Object classification Bag-of-words Spatial pyramids Neural Networks Object
Probabilistic Latent Semantic Analysis (plsa)
Probabilistic Latent Semantic Analysis (plsa) SS 2008 Bayesian Networks Multimedia Computing, Universität Augsburg [email protected] www.multimedia-computing.{de,org} References
A Comparative Study between SIFT- Particle and SURF-Particle Video Tracking Algorithms
A Comparative Study between SIFT- Particle and SURF-Particle Video Tracking Algorithms H. Kandil and A. Atwan Information Technology Department, Faculty of Computer and Information Sciences, Mansoura University,El-Gomhoria
How To Write A Fast Binary Descriptor Based On Brief
ORB: an efficient alternative to SIFT or SURF Ethan Rublee Vincent Rabaud Kurt Konolige Gary Bradski Willow Garage, Menlo Park, California {erublee}{vrabaud}{konolige}{bradski}@willowgarage.com Abstract
Data, Measurements, Features
Data, Measurements, Features Middle East Technical University Dep. of Computer Engineering 2009 compiled by V. Atalay What do you think of when someone says Data? We might abstract the idea that data are
Local features and matching. Image classification & object localization
Overview Instance level search Local features and matching Efficient visual recognition Image classification & object localization Category recognition Image classification: assigning a class label to
3D Model based Object Class Detection in An Arbitrary View
3D Model based Object Class Detection in An Arbitrary View Pingkun Yan, Saad M. Khan, Mubarak Shah School of Electrical Engineering and Computer Science University of Central Florida http://www.eecs.ucf.edu/
Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall
Automatic Photo Quality Assessment Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Estimating i the photorealism of images: Distinguishing i i paintings from photographs h Florin
MIFT: A Mirror Reflection Invariant Feature Descriptor
MIFT: A Mirror Reflection Invariant Feature Descriptor Xiaojie Guo, Xiaochun Cao, Jiawan Zhang, and Xuewei Li School of Computer Science and Technology Tianjin University, China {xguo,xcao,jwzhang,lixuewei}@tju.edu.cn
Image Segmentation and Registration
Image Segmentation and Registration Dr. Christine Tanner ([email protected]) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation
CATEGORIZATION OF SIMILAR OBJECTS USING BAG OF VISUAL WORDS AND k NEAREST NEIGHBOUR CLASSIFIER
TECHNICAL SCIENCES Abbrev.: Techn. Sc., No 15(2), Y 2012 CATEGORIZATION OF SIMILAR OBJECTS USING BAG OF VISUAL WORDS AND k NEAREST NEIGHBOUR CLASSIFIER Piotr Artiemjew, Przemysław Górecki, Krzysztof Sopyła
Feature Tracking and Optical Flow
02/09/12 Feature Tracking and Optical Flow Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem Many slides adapted from Lana Lazebnik, Silvio Saverse, who in turn adapted slides from Steve
Mean-Shift Tracking with Random Sampling
1 Mean-Shift Tracking with Random Sampling Alex Po Leung, Shaogang Gong Department of Computer Science Queen Mary, University of London, London, E1 4NS Abstract In this work, boosting the efficiency of
Build Panoramas on Android Phones
Build Panoramas on Android Phones Tao Chu, Bowen Meng, Zixuan Wang Stanford University, Stanford CA Abstract The purpose of this work is to implement panorama stitching from a sequence of photos taken
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Point Matching as a Classification Problem for Fast and Robust Object Pose Estimation
Point Matching as a Classification Problem for Fast and Robust Object Pose Estimation Vincent Lepetit Julien Pilet Pascal Fua Computer Vision Laboratory Swiss Federal Institute of Technology (EPFL) 1015
Bags of Binary Words for Fast Place Recognition in Image Sequences
IEEE TRANSACTIONS ON ROBOTICS, VOL., NO., MONTH, YEAR. SHORT PAPER 1 Bags of Binary Words for Fast Place Recognition in Image Sequences Dorian Gálvez-López and Juan D. Tardós, Member, IEEE Abstract We
Cloud based object recognition: A system proposal
Cloud based object recognition: A system proposal Daniel LORENČÍK 1, Peter SINČÁK 2 Abstract In this chapter, we will present a proposal for the cloud based object recognition system. The system will extract
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 269 Class Project Report
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 69 Class Project Report Junhua Mao and Lunbo Xu University of California, Los Angeles [email protected] and lunbo
SURF: Speeded Up Robust Features
SURF: Speeded Up Robust Features Herbert Bay 1, Tinne Tuytelaars 2, and Luc Van Gool 12 1 ETH Zurich {bay, vangool}@vision.ee.ethz.ch 2 Katholieke Universiteit Leuven {Tinne.Tuytelaars, Luc.Vangool}@esat.kuleuven.be
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
Recognizing Cats and Dogs with Shape and Appearance based Models. Group Member: Chu Wang, Landu Jiang
Recognizing Cats and Dogs with Shape and Appearance based Models Group Member: Chu Wang, Landu Jiang Abstract Recognizing cats and dogs from images is a challenging competition raised by Kaggle platform
FastKeypointRecognitioninTenLinesofCode
FastKeypointRecognitioninTenLinesofCode Mustafa Özuysal Pascal Fua Vincent Lepetit Computer Vision Laboratory École Polytechnique Fédérale de Lausanne(EPFL) 115 Lausanne, Switzerland Email: {Mustafa.Oezuysal,
Group Sparse Coding. Fernando Pereira Google Mountain View, CA [email protected]. Dennis Strelow Google Mountain View, CA strelow@google.
Group Sparse Coding Samy Bengio Google Mountain View, CA [email protected] Fernando Pereira Google Mountain View, CA [email protected] Yoram Singer Google Mountain View, CA [email protected] Dennis Strelow
Automatic georeferencing of imagery from high-resolution, low-altitude, low-cost aerial platforms
Automatic georeferencing of imagery from high-resolution, low-altitude, low-cost aerial platforms Amanda Geniviva, Jason Faulring and Carl Salvaggio Rochester Institute of Technology, 54 Lomb Memorial
Distinctive Image Features from Scale-Invariant Keypoints
Distinctive Image Features from Scale-Invariant Keypoints David G. Lowe Computer Science Department University of British Columbia Vancouver, B.C., Canada [email protected] January 5, 2004 Abstract This paper
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set
EM Clustering Approach for Multi-Dimensional Analysis of Big Data Set Amhmed A. Bhih School of Electrical and Electronic Engineering Princy Johnson School of Electrical and Electronic Engineering Martin
Android Ros Application
Android Ros Application Advanced Practical course : Sensor-enabled Intelligent Environments 2011/2012 Presentation by: Rim Zahir Supervisor: Dejan Pangercic SIFT Matching Objects Android Camera Topic :
Cees Snoek. Machine. Humans. Multimedia Archives. Euvision Technologies The Netherlands. University of Amsterdam The Netherlands. Tree.
Visual search: what's next? Cees Snoek University of Amsterdam The Netherlands Euvision Technologies The Netherlands Problem statement US flag Tree Aircraft Humans Dog Smoking Building Basketball Table
AN EFFICIENT HYBRID REAL TIME FACE RECOGNITION ALGORITHM IN JAVA ENVIRONMENT ABSTRACT
AN EFFICIENT HYBRID REAL TIME FACE RECOGNITION ALGORITHM IN JAVA ENVIRONMENT M. A. Abdou*, M. H. Fathy** *Informatics Research Institute, City for Scientific Research and Technology Applications (SRTA-City),
Surgical Tools Recognition and Pupil Segmentation for Cataract Surgical Process Modeling
Surgical Tools Recognition and Pupil Segmentation for Cataract Surgical Process Modeling David Bouget, Florent Lalys, Pierre Jannin To cite this version: David Bouget, Florent Lalys, Pierre Jannin. Surgical
FAST APPROXIMATE NEAREST NEIGHBORS WITH AUTOMATIC ALGORITHM CONFIGURATION
FAST APPROXIMATE NEAREST NEIGHBORS WITH AUTOMATIC ALGORITHM CONFIGURATION Marius Muja, David G. Lowe Computer Science Department, University of British Columbia, Vancouver, B.C., Canada [email protected],
Classifying Manipulation Primitives from Visual Data
Classifying Manipulation Primitives from Visual Data Sandy Huang and Dylan Hadfield-Menell Abstract One approach to learning from demonstrations in robotics is to make use of a classifier to predict if
How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm
IJSTE - International Journal of Science Technology & Engineering Volume 1 Issue 10 April 2015 ISSN (online): 2349-784X Image Estimation Algorithm for Out of Focus and Blur Images to Retrieve the Barcode
Clustering & Visualization
Chapter 5 Clustering & Visualization Clustering in high-dimensional databases is an important problem and there are a number of different clustering paradigms which are applicable to high-dimensional data.
A Genetic Algorithm-Evolved 3D Point Cloud Descriptor
A Genetic Algorithm-Evolved 3D Point Cloud Descriptor Dominik Wȩgrzyn and Luís A. Alexandre IT - Instituto de Telecomunicações Dept. of Computer Science, Univ. Beira Interior, 6200-001 Covilhã, Portugal
Face Recognition in Low-resolution Images by Using Local Zernike Moments
Proceedings of the International Conference on Machine Vision and Machine Learning Prague, Czech Republic, August14-15, 014 Paper No. 15 Face Recognition in Low-resolution Images by Using Local Zernie
CS231M Project Report - Automated Real-Time Face Tracking and Blending
CS231M Project Report - Automated Real-Time Face Tracking and Blending Steven Lee, [email protected] June 6, 2015 1 Introduction Summary statement: The goal of this project is to create an Android
Convolution. 1D Formula: 2D Formula: Example on the web: http://www.jhu.edu/~signals/convolve/
Basic Filters (7) Convolution/correlation/Linear filtering Gaussian filters Smoothing and noise reduction First derivatives of Gaussian Second derivative of Gaussian: Laplacian Oriented Gaussian filters
Incremental PCA: An Alternative Approach for Novelty Detection
Incremental PCA: An Alternative Approach for Detection Hugo Vieira Neto and Ulrich Nehmzow Department of Computer Science University of Essex Wivenhoe Park Colchester CO4 3SQ {hvieir, udfn}@essex.ac.uk
14.10.2014. Overview. Swarms in nature. Fish, birds, ants, termites, Introduction to swarm intelligence principles Particle Swarm Optimization (PSO)
Overview Kyrre Glette kyrrehg@ifi INF3490 Swarm Intelligence Particle Swarm Optimization Introduction to swarm intelligence principles Particle Swarm Optimization (PSO) 3 Swarms in nature Fish, birds,
Part-Based Recognition
Part-Based Recognition Benedict Brown CS597D, Fall 2003 Princeton University CS 597D, Part-Based Recognition p. 1/32 Introduction Many objects are made up of parts It s presumably easier to identify simple
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
Practical Tour of Visual tracking. David Fleet and Allan Jepson January, 2006
Practical Tour of Visual tracking David Fleet and Allan Jepson January, 2006 Designing a Visual Tracker: What is the state? pose and motion (position, velocity, acceleration, ) shape (size, deformation,
Next Generation Artificial Vision Systems
Next Generation Artificial Vision Systems Reverse Engineering the Human Visual System Anil Bharath Maria Petrou Imperial College London ARTECH H O U S E BOSTON LONDON artechhouse.com Contents Preface xiii
Similarity Search in a Very Large Scale Using Hadoop and HBase
http:/cedric.cnam.fr Similarity Search in a Very Large Scale Using Hadoop and HBase Rapport de recherche CEDRIC 2012 http://cedric.cnam.fr Stanislav Barton (CNAM), Vlastislav Dohnal (Masaryk University),
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
How To Program With Adaptive Vision Studio
Studio 4 intuitive powerful adaptable software for machine vision engineers Introduction Adaptive Vision Studio Adaptive Vision Studio software is the most powerful graphical environment for machine vision
FACE RECOGNITION BASED ATTENDANCE MARKING SYSTEM
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 2, February 2014,
Supervised Feature Selection & Unsupervised Dimensionality Reduction
Supervised Feature Selection & Unsupervised Dimensionality Reduction Feature Subset Selection Supervised: class labels are given Select a subset of the problem features Why? Redundant features much or
Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data
CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear
Feature Matching for Aligning Historical and Modern Images
188 IJCA, Vol. 21, No. 3, Sept. 2014 Feature Matching for Aligning Historical and Modern Images Heider K. Ali * Carleton University, Ottawa, ON, K1S 5B8, CANADA Anthony Whitehead Carleton University, Ottawa,
Character Image Patterns as Big Data
22 International Conference on Frontiers in Handwriting Recognition Character Image Patterns as Big Data Seiichi Uchida, Ryosuke Ishida, Akira Yoshida, Wenjie Cai, Yaokai Feng Kyushu University, Fukuoka,
The Visual Internet of Things System Based on Depth Camera
The Visual Internet of Things System Based on Depth Camera Xucong Zhang 1, Xiaoyun Wang and Yingmin Jia Abstract The Visual Internet of Things is an important part of information technology. It is proposed
MVA ENS Cachan. Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos [email protected]
Machine Learning for Computer Vision 1 MVA ENS Cachan Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos [email protected] Department of Applied Mathematics Ecole Centrale Paris Galen
MusicGuide: Album Reviews on the Go Serdar Sali
MusicGuide: Album Reviews on the Go Serdar Sali Abstract The cameras on mobile phones have untapped potential as input devices. In this paper, we present MusicGuide, an application that can be used to
Parallel Data Selection Based on Neurodynamic Optimization in the Era of Big Data
Parallel Data Selection Based on Neurodynamic Optimization in the Era of Big Data Jun Wang Department of Mechanical and Automation Engineering The Chinese University of Hong Kong Shatin, New Territories,
Scale-Invariant Object Categorization using a Scale-Adaptive Mean-Shift Search
in DAGM 04 Pattern Recognition Symposium, Tübingen, Germany, Aug 2004. Scale-Invariant Object Categorization using a Scale-Adaptive Mean-Shift Search Bastian Leibe ½ and Bernt Schiele ½¾ ½ Perceptual Computing
Canny Edge Detection
Canny Edge Detection 09gr820 March 23, 2009 1 Introduction The purpose of edge detection in general is to significantly reduce the amount of data in an image, while preserving the structural properties
Robert Collins CSE598G. More on Mean-shift. R.Collins, CSE, PSU CSE598G Spring 2006
More on Mean-shift R.Collins, CSE, PSU Spring 2006 Recall: Kernel Density Estimation Given a set of data samples x i ; i=1...n Convolve with a kernel function H to generate a smooth function f(x) Equivalent
SZTAKI @ ImageCLEF 2011
SZTAKI @ ImageCLEF 2011 Bálint Daróczy Róbert Pethes András A. Benczúr Data Mining and Web search Research Group, Informatics Laboratory Computer and Automation Research Institute of the Hungarian Academy
Environmental Remote Sensing GEOG 2021
Environmental Remote Sensing GEOG 2021 Lecture 4 Image classification 2 Purpose categorising data data abstraction / simplification data interpretation mapping for land cover mapping use land cover class
Object tracking & Motion detection in video sequences
Introduction Object tracking & Motion detection in video sequences Recomended link: http://cmp.felk.cvut.cz/~hlavac/teachpresen/17compvision3d/41imagemotion.pdf 1 2 DYNAMIC SCENE ANALYSIS The input to
Similarity Search in a Very Large Scale Using Hadoop and HBase
Similarity Search in a Very Large Scale Using Hadoop and HBase Stanislav Barton, Vlastislav Dohnal, Philippe Rigaux LAMSADE - Universite Paris Dauphine, France Internet Memory Foundation, Paris, France
Edge tracking for motion segmentation and depth ordering
Edge tracking for motion segmentation and depth ordering P. Smith, T. Drummond and R. Cipolla Department of Engineering University of Cambridge Cambridge CB2 1PZ,UK {pas1001 twd20 cipolla}@eng.cam.ac.uk
A Guided User Experience Using Subtle Gaze Direction
A Guided User Experience Using Subtle Gaze Direction Eli Ben-Joseph and Eric Greenstein Stanford University {ebj, ecgreens}@stanford.edu 1 Abstract This paper demonstrates how illumination modulation can
A Comparative Study of SIFT and its Variants
10.2478/msr-2013-0021 MEASUREMENT SCIENCE REVIEW, Volume 13, No. 3, 2013 A Comparative Study of SIFT and its Variants Jian Wu 1, Zhiming Cui 1, Victor S. Sheng 2, Pengpeng Zhao 1, Dongliang Su 1, Shengrong
Multi-View Object Class Detection with a 3D Geometric Model
Multi-View Object Class Detection with a 3D Geometric Model Joerg Liebelt IW-SI, EADS Innovation Works D-81663 Munich, Germany [email protected] Cordelia Schmid LEAR, INRIA Grenoble F-38330 Montbonnot,
Euler Vector: A Combinatorial Signature for Gray-Tone Images
Euler Vector: A Combinatorial Signature for Gray-Tone Images Arijit Bishnu, Bhargab B. Bhattacharya y, Malay K. Kundu, C. A. Murthy fbishnu t, bhargab, malay, [email protected] Indian Statistical Institute,
The Delicate Art of Flower Classification
The Delicate Art of Flower Classification Paul Vicol Simon Fraser University University Burnaby, BC [email protected] Note: The following is my contribution to a group project for a graduate machine learning
How To Cluster
Data Clustering Dec 2nd, 2013 Kyrylo Bessonov Talk outline Introduction to clustering Types of clustering Supervised Unsupervised Similarity measures Main clustering algorithms k-means Hierarchical Main
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS
VEHICLE LOCALISATION AND CLASSIFICATION IN URBAN CCTV STREAMS Norbert Buch 1, Mark Cracknell 2, James Orwell 1 and Sergio A. Velastin 1 1. Kingston University, Penrhyn Road, Kingston upon Thames, KT1 2EE,
Tracking in flussi video 3D. Ing. Samuele Salti
Seminari XXIII ciclo Tracking in flussi video 3D Ing. Tutors: Prof. Tullio Salmon Cinotti Prof. Luigi Di Stefano The Tracking problem Detection Object model, Track initiation, Track termination, Tracking
Intelligent Flexible Automation
Intelligent Flexible Automation David Peters Chief Executive Officer Universal Robotics February 20-22, 2013 Orlando World Marriott Center Orlando, Florida USA Trends in AI and Computing Power Convergence
Image Classification for Dogs and Cats
Image Classification for Dogs and Cats Bang Liu, Yan Liu Department of Electrical and Computer Engineering {bang3,yan10}@ualberta.ca Kai Zhou Department of Computing Science [email protected] Abstract
Scalable Object Detection by Filter Compression with Regularized Sparse Coding
Scalable Object Detection by Filter Compression with Regularized Sparse Coding Ting-Hsuan Chao, Yen-Liang Lin, Yin-Hsi Kuo, and Winston H Hsu National Taiwan University, Taipei, Taiwan Abstract For practical
Feature description based on Center-Symmetric Local Mapped Patterns
Feature description based on Center-Symmetric Local Mapped Patterns Carolina Toledo Ferraz, Osmando Pereira Jr., Adilson Gonzaga Department of Electrical Engineering, University of São Paulo Av. Trabalhador
Image Compression through DCT and Huffman Coding Technique
International Journal of Current Engineering and Technology E-ISSN 2277 4106, P-ISSN 2347 5161 2015 INPRESSCO, All Rights Reserved Available at http://inpressco.com/category/ijcet Research Article Rahul
Unsupervised Data Mining (Clustering)
Unsupervised Data Mining (Clustering) Javier Béjar KEMLG December 01 Javier Béjar (KEMLG) Unsupervised Data Mining (Clustering) December 01 1 / 51 Introduction Clustering in KDD One of the main tasks in
An Iterative Image Registration Technique with an Application to Stereo Vision
An Iterative Image Registration Technique with an Application to Stereo Vision Bruce D. Lucas Takeo Kanade Computer Science Department Carnegie-Mellon University Pittsburgh, Pennsylvania 15213 Abstract
MetropoGIS: A City Modeling System DI Dr. Konrad KARNER, DI Andreas KLAUS, DI Joachim BAUER, DI Christopher ZACH
MetropoGIS: A City Modeling System DI Dr. Konrad KARNER, DI Andreas KLAUS, DI Joachim BAUER, DI Christopher ZACH VRVis Research Center for Virtual Reality and Visualization, Virtual Habitat, Inffeldgasse
Visual area MT responds to local motion. Visual area MST responds to optic flow. Visual area STS responds to biological motion. Macaque visual areas
Visual area responds to local motion MST a Visual area MST responds to optic flow MST a Visual area STS responds to biological motion STS Macaque visual areas Flattening the brain What is a visual area?
