Fast Matting Using Large Kernel Matting Laplacian Matrices
|
|
|
- Aldous Ryan
- 9 years ago
- Views:
Transcription
1 Fast Matting Using Large Kernel Matting Laplacian Matrices Kaiming He 1 1 Department of Information Engineering The Chinese University of Hong Kong Jian Sun 2 2 Microsoft Research Asia Xiaoou Tang 1,3 3 Shenzhen Institute of Advanced Technology Chinese Academy of Sciences Abstract Image matting is of great importance in both computer vision and graphics applications. Most existing state-of-the-art techniques rely on large sparse matrices such as the matting Laplacian [12]. However, solving these linear systems is often time-consuming, which is unfavored for the user interaction. In this paper, we propose a fast method for high quality matting. We first derive an efficient algorithm to solve a large kernel matting Laplacian. A large kernel propagates information more quickly and may improve the matte quality. To further reduce running time, we also use adaptive kernel sizes by a KD-tree trimap segmentation technique. A variety of experiments show that our algorithm provides high quality results and is 5 to 20 times faster than previous methods. 1. Introduction Image matting refers to the problem of softly extracting the foreground object from a single input image. Denoting a pixel s foreground color, background color, and foreground opacity (alpha matte) as F, B, and α respectively, the pixel s color I is a convex combination of F and B: I = Fα+B(1 α). (1) Image matting has a number of important applications, such as image/video segmentation, layer extraction, new view synthesis, interactive image editing, and film making. Since the problem is highly ill-posed, a trimap (or strokes) indicating definite foreground, definite background, and unknown regions is usually given by the user. To infer the alpha matte in the unknown regions, Bayesian Matting [4] uses a progressively moving window marching inward from the known regions. In [2], the alpha value is calculated by the geodesic distance from a pixel to the known regions. Affinity-matrix-based methods propagate constraints from the known regions to the unknown regions: Poisson matting [19] solves a homogenous Laplacian matrix; Random Walk matting [7] defines the affinity matrix by a Gaussian function; in the closed-form matting [12], a matting Laplacian matrix is proposed under a color line assumption. The matting Laplacian can also be combined (a) (c) Figure 1. (a) Image ( pixels). (b) Trimap (180k unknown pixels). (c) Result of closed-form matting [12]. The running time is 28 seconds using a coarse-to-fine solver in [12]. (d) Our result. The running time is 2.5 seconds. Notice the gap between the toys. with different data weights or priors [24, 14, 15]. New appearance models [18] and a learning based technique [25] are proposed to generalize the color line assumption. Currently, matting-laplacian-based methods produce the best quality results [16, 23]. However, efficiency is also very important for image matting, especially when we are faced with multi-megapixel images produced by today s digital cameras. The progressively moving window in Bayesian Matting [4] is computationally expensive. The geodesic framework [2] presents a linear time algorithm. But the model itself may not be able to handel complex cases like hair. Poisson Matting [19] and Random Walk Matting can be accelerated by a multigrid solver or a GPU implementation, but lack satisfactory quality [16]. The methods using the matting Laplacian [12, 24, 14, 15] or other sophisticated matrices [18, 25] have to solve a large linear system - a very time consuming operation. Soft Scissor [21] locally solves a small linear system while the user is drawing the trimap. But it is still slow for mega-pixel images and not applicable for predrawn trimaps. A more efficient method for high quality matting is still desired. (b) (d) 1
2 In this paper, we propose a fast algorithm for high quality image matting using large kernel matting Laplacian matrices. The algorithm is based on an efficient method to solve the linear system with the large kernel matting Laplacian. Using a large kernel can accelerate the constraint propagation, reduce the time of the linear solver for convergence, and improve the matting quality. To further speed-up the algorithm, we use a KD-tree based technique to decompose the trimap so that we can assign an adaptive kernel size to each sub-trimap. Thus, the number of iterations can be fixed beforehand and the running time of the whole algorithm is only linear to the number of the unknown pixels. We demonstrate that our algorithm is about 5 to 20 times faster than previous methods while achieving high quality (Fig. 1). Our algorithm is also useful for other applications using the matting Laplacian, like haze removal [9], spatially variant white balance [10], and intrinsic images [3]. 2. Background The matting Laplacian matrix proposed in [12] is an affinity matrix specially designed for image matting. The key assumption of this method is the color line model: the foreground (or background) colors in a local window lie on a single line in the RGB color space. It can be proved that α is a linear transformation ofiin the local window: α i = a T I i +b, i ω, (2) Hereiis a pixel index,i i andaare3 1 vectors, andaand b are assumed to be constant in the local windowω. Accordingly, a cost function J(α, a, b) can be defined to encourage the alpha obeying this model: J(α,a,b) = ( (α i a T k I i b k ) 2 +εa T k a k), (3) k I i ω k whereω k is the window centered at pixelk, andεis a regularization parameter. By minimizing the cost function w.r.t. (a,b), a quadratic function ofαcan be obtained: J(α) = α T Lα. (4) Hereαis denoted as an N 1 vector, where N is the number of unknowns. And L is called matting Laplacian. It is an N N symmetric matrix whose (i, j) element is: k (i,j) ω k (δ ij 1 ω k (1+(I i μ k ) T (Σ k + ε ω k U) 1 (I j μ k ))), (5) where δ ij is the Kronecker delta, μ k and Σ k are the mean and covariance matrix of the colors in window ω k, ω k is the number of pixels inω k, anduis a 3 3 identity matrix. Combining this cost function with the user-specified constraints (trimap), the whole cost function is defined as: E(α) = α T Lα+λ(α β) T D(α β), (6) ω k j ω k k ω i i Figure 2. According to (5), the (i, j) element of L is non-zero only if (i,j) ω k. If the radius of the window ω is r, the radius of the kernel centered at i is 2r. The kernel is a rectangle including all the gray pixels in this figure. where β is the trimap, D is a diagonal matrix whose elements are one for constraint pixels and zero otherwise, and λ is a large number. A data term specified by color sampling confidence [24, 14] or a sparsity prior [15] can also be incorporated. The cost function (6) can be optimized by solving a sparse linear system: r ω i 2r (L+λD)α = λdβ. (7) The techniques of haze removal [9], spatially variant white balance [10], and intrinsic images [3] also involve this linear system. Given the linear system, we need to use an appropriate solver to recover α. Non-iterative methods like LU decomposition are not effective to handle this system in large scale due to the high memory cost. So iterative methods are usually applied, like the Jacobi method or Conjugate Gradient () [17]. The information of the known pixels is propagated into the unknown region by iteratively multiplying the matrix. However, iterative methods are often time consuming. For example, the time complexity of is O(N 1.5 ) for N unknowns. A medium size image ( ) may take hundreds or thousands of iterations. Another drawback of the iterative methods is that the time needed is hard to predict because the iteration number depends on the number of unknowns, the image content, and the trimap shape. This is not favored for interactive image matting. In [12], a coarse-to-fine scheme is proposed to speedup the linear system solver. But the down-sampling process may degrades the matte quality, especially for the foreground with very thin structures. Moreover, in the fine resolution it also suffers from the drawbacks of the iterative methods. Locally adapted hierarchical basis preconditioning in [20] is not directly applicable because the elements of the matting Laplacian matrix (5) are not first-order smooth. 3. Large Kernel Matting Laplacian Existing methods [12, 24, 14, 15] usually use a small window because the matrix will be less sparse with a larger window. Conventional wisdoms hold that solving a less
3 sparse system is even slower. But we point out it is not necessarily true. A less sparse system needs fewer iterations to converge; the only bottle neck is the increased computational burden in each iteration. In this section, we propose an algorithm to greatly reduce the time needed for each iteration, so the solver is in fact faster if we use a larger window. The kernel size of a matting Laplacian is defined as the the number of the non-zero elements in a row of L. Denote the window radius of ω in (5) by r. The kernel size is (4r +1) 2 (see Fig. 2). Existing methods mainly use r = 1 because L will become less sparse when r is larger. Both the memory and the time needed to solve the linear system will increase tremendously. Take for example. The solver iteratively multiplies the conjugate vector by the Laplacian matrix (please see [17] for detailed description of ). In each iteration, the matrix product Lp dominates the computation cost. Here p is the conjugate vector of the previous iteration. In the view of signal processing, the matrix product Lp is the response of a spatially variant filter L onp, whosei th element is: (Lp) i = j L ij p j. (8) Computing Lp using (5) and (8) involves spatially variant convolution, whose time and memory complexity is O(Nr 2 ). This is not affordable whenr gets larger. We notice that in each iteration, a pixel can influence another pixel that is 2r away. So the information is propagated by 2r pixels. The solver will converge in fewer iterations if the kernel size is larger. This motivates us to envision that the use of a large kernel may reduce the solver s running time, if we can significantly increase the speed of the convolution in (8). Toward this goal, we present an O(N) time (independent of r) algorithm to compute the productlp in each iteration. Instead of computingl s elements and the convolution explicitly, we calculate the product Lp as a whole by the following algorithm: Algorithm for calculating Lp Given a conjugate vector p, Lp can be calculated through the following three steps: a k = Δ 1 k ( 1 ω i ω k I i p i μ k p k ), (9) b k = p k a kt μ k, (10) (Lp) i q i = ω p i (( a k )T I i +( b k )), (11) k ω i k ω i wherea k is a 3 1 vector for each pixelk, p k is the mean of p inω k, Δ k = Σ k + ε ω k U, and we denote(lp) i asq i. Theorem: The q given by the above algorithm is identical to thelp calculated by (5) and (8). Proof: Written in matrix notation, from (9) there is an affine transform: a = Ap, (12) where A is a coefficient matrix only dependent on I. If we put (9) into (10), we can see that b is also p s affine transform: b = Bp. Similarly, q is p s affine transform: q = Qp. Consequently, in order to show q = Lp, we only need to prove q i / p j = L(i,j). Putting (10) into (11) and eliminatingb, we obtain: q i = ω δ ij ( p k + a T k (I i μ k )) (13) p j p j p j k ω i Here, we have: p k = 1 p j ω n ω k p n p j = 1 ω δ j ω k = 1 ω δ k ω j (14) where δ j ωk is 1 if j ω k, and is 0 otherwise. According to (9) we also have: a k p j = Δ 1 k ( 1 p i I i p k μ k ) ω p i ω j p j k = Δ 1 k ( 1 ω I j 1 ω μ k)δ k ωj (15) Putting (14) and (15) into (13), we obtain: q i = ω δ ij 1 p j ω k ω i,k ω j (1+(I j μ k ) T Δ 1 k (I i μ k )) This is exactlyl(i,j) in (5). The proposed algorithm also has intuitive interpretations: Equations (9) and (10) indeed are linear regression solutions and the regression model isp i a kt I i +b k, i ω k. Further, (11) can be rewritten as: (Lp) i = (p i (a kt I i +b k )). (16) k ω i For any pixel I i, the term (p i (a kt I i +b k )) is the error betweenp i and its linear prediction. AsI i is involved in all the regression processes satisfyingk ω i, equation (16) is the sum of errors in all windows aroundi. All the summations in (9) and (11) can be very efficiently computed using the integral image technique [6]. Using the integral image, the sum of the any window can be obtained in constant time (four operations). Therefore, the time complexity for computinglp in each iteration is O(N ) O(N), where N is the size of the bounding box of the unknown region. Note that the time complexity is independent of the kernel size.
4 (a) (b) (c) (d) (a) (b) (c) (d) log 10 err r=1 r=5 r=10 r=20 iteration Figure 3. (a) Cropped image (98 130). (b) Trimap. (c) r=1, iteration 519. (d) r=20, iteration 44. On the bottom are the curves of error vs. iteration numbers. (a) (b) (c) (d) Figure 4. (a) Cropped image ( ). (b) Trimap. (c) r=1. (d) r=20. Discussion on Kernel Size Using the above algorithm, the running time of the solver can be greatly reduced if we use a large kernel - the solver converges in fewer iterations. In Fig. 3, we experiment several different kernel sizes on the same image. The solver converges much faster in the large kernel case. Our algorithm s running time is 1.07s (r=1, converges in 519 iterations) and 0.090s (r=20, converges in 44 iterations), while the running time of brute force using (5) and (8) is 0.95s (r=1) and 22s (r=20). In this example, the resulting mattes are almost visually identical (Fig. 3(c)(d)). A large kernel may improve the quality because a large window may cover disconnected regions of the foreground/background. This property is particularly favored when the foreground object has holes. In Fig. 4(c), r = 1 is too small to cover the known background and the holes. We can obtain a better matte if we use a larger window size r = 20 (Fig. 4(d)). Our large kernel method is typically appropriate for high resolution images. In Fig. 5(c), a small window is not sufficient to describe the color lines and results in the loss of the fine structures. However, a large window collects enough color samples and gives a better result (Fig. 5(d)). A large window also covers more known pixels near the boundary of unknowns and provides more stable constraints. The drawback of a large kernel is that when the foreground/background is complicated, a larger window leads Figure 5. (a) A cropped image ( ) from a image. (b) Trimap. (c) r=1. (d) r=60. We recommend viewing the electronic version of this paper for the fine structures. to a higher probability of breaking the color line assumption. In the zoomed-in patches of Fig. 6, a large window covers more than one color clusters of the background, so the colors do not lie in the same line. Consequently, the matte is not as accurate as using a small window. We observe that in practice, the user draws a wide band of unknowns near a fuzzy object or an object with holes, as it is difficult to provide a precise boundary; the band of unknowns near a solid boundary is relatively narrow. We can see this fact in Fig. 1(b) and Fig. 6(b). Besides, a high resolution image may also result in a wider band of unknowns. So we prefer to use a large kernel to efficiently propagate constraints in a wide band. For a narrow band, a small kernel is favored to avoid breaking the color line assumption. To achieve this goal, in the next section, we adaptively set the kernel size based on a trimap segmentation technique. 4. KD-tree Based Trimap Segmentation In this section, we propose a trimap segmentation method to further improve both the quality and efficiency. This allows us to assign different window sizes to different regions. Given a trimap, we first calculate the unknown region s barycenter (x c,y c ), x variance σx 2 = 1 n U (x c x) 2 and y variance σy 2, where U is the set of unknown pixels, and n is the size of U. Then we use a line that passes through (x c,y c ) to divide the map into two regions. This line is vertical to the axis (x or y) with a larger variance. A trimap is recursively divided and a 2D KD-tree is built accordingly. Next we discuss the conditions when the recursive process stops. We observe that if a cropped image covers enough foreground and background constraints, a reasonably good matte can be obtained by only considering this cropped image independently, as demonstrated in Fig. 3 and Fig. 4. Therefore, we hope the segments are small enough, while as many of them as possible have both foreground (F) and background (B) constraints. So the recursive division stops until one of these conditions is satisfied: (a) the segment only has F and U; (b) the segment only has B and U; (c) if we divide the segment, one of its children only has F and U, and the other only has B and U; (d) The segment s min(width, height) is smaller than a threshold (32 in our experiments). Fig. 7 gives an example of a segmented trimap. Notice that the band width of the unknowns in each segment
5 (a) (b) (c) (d) Figure 6. (a) A image. (b) Trimap. (c) r=1. (d) r=10. We recommend viewing the matte in the electronic version of this paper. (a-1) (a-2) (a) (a-1) (a-2) Figure 7. Trimap segmentation. r=17, iteration 1 r=17, iteration 3 r=17, iteration 10 (b) (b-1) (b-2) r=10, iteration 1 r=10, iteration 3 r=10, iteration 10 Figure 8. Propagation and window sizes. The band widthw b 50 in the first row, and w b 30 in the second row,. The window radius r = w b /3. (c) (c-1) (c-2) is nearly uniform. Once the trimap is segmented, we can solve the linear system in each segment. The integral image is only calculated in the bounding box of the unknowns in each segment. We first solve the segments that satisfy condition (c) or (d), as they directly contain both F and B constraints. Then, other segments (that only have F and U, or B and U) are solved in inverse Breadth-First Search order, i.e., the deeper leaves in the KD-trees have a higher priority. As they do not contain F or B constraint, we use the solved matte of the neighboring segment as boundary conditions. The inverse Breadth-First Search order insures that at least one of their neighbors has been solved before. More importantly, we adapt the kernel size to the band width of the unknowns. For each segment, let (w, h) be the width and height of U s bounding box. We approximate the width of the band by w b =n/max(w,h). We set the window radius r = w b /η, where η is a factor. Intuitively, the propagation will influence the other side of the band in η/2 iterations (notice that the kernel radius is 2r). Therefore, the number of iterations needed for convergence is in the order of η, and can be fixed beforehand. In Fig. 8, we show that (d) (d-1) (d-2) Figure 9. Local-global-local Scheme. (a) Segmented trimap of Fig. 1. (b) Local step 1. (c) Global step. (d) Local step 2. Please refer to the text for explanation. We recommend to see the electronic version of these images. the propagation speed is adaptive to the band width. The same number of iterations provide high quality mattes for two bands of different width. A Local-Global-Local Scheme To prevent the result being too locally determined, we propose a local-global-local scheme. Given the trimap segmentation, we first solve the matte for each segment. Here we let η=3. This step propagates information fast. We empirically fix the iteration number to 10 which is sufficient to get a good initial matte for the next step (Fig. 9(b)). Noticeable seams may exist on segment boundaries (Fig. 9(b-1)). Using the above obtained solution as initial guess, we optimize the whole linear system globally. In this step, we set the window radius r to be min(w, h)/50 where w and h
6 are the image s width and height. The iteration number is fixed on 5. This global step removes the seams (Fig. 9(c- 1)). But as the kernel size may be too large for some local regions, there are artifacts due to the failure of the color line assumption (Fig. 9(c-2)). In the final step, we refine this solution locally using the trimap segmentation again. Here we let η = 15 and run 20 iterations. In experiments we found that more iterations improve the result very little. The window size is much smaller than the first local step because we want to maintain the color line model. This suppresses the artifacts due to the large window size used in previous two steps (Fig. 9(d-2)). Time Complexity Denote N as the total area of all the bounding boxes of the unknowns in all segments. The time complexity of the local steps is O(N ), since the time for calculating Lp in each iteration is O(N ) and the iteration number is fixed. For the same reason, the time complexity of the global step is O(M), where M is the size of the bounding box of the whole unknown region. In experiment, we find that the time for the global step is much less than that of the local steps due to fewer iterations. So the total time complexity is O(N ). As N is slightly larger than the number of unknowns N, the running time is almost linear to N. Besides, once the trimap is segmented, the time can be predicted before running the solver. 5. Results We compare our algorithm with the closed-form matting in [12] (i.e., r=1). For closed-form matting, we calculate the matting Laplacian using (5) and then solve the linear system using two different methods: brute force (simply denoted as below), and the coarse-to-fine () scheme described in [12]. For, we fix the convergence tolerance (10 5 ). For, we run three levels and in each level we solve the system using. All algorithms are implemented using C++ on a laptop with an Intel Core Duo 2.0GHz CPU and 4G memory. Table 1 shows the running time. Our algorithm is about 5 to 20 times faster than or for medium and large images. We also study the average running time per unknown pixel (τ = t/n). Our algorithm gives similarτ (about 0.02) for different images. The small variation of ourτ is mainly because the time is linear to N but not N. On the contrary, τ varies severely for the other two algorithms, because it is influenced by the number of unknowns, image content and the shape of the trimap. Their τ is also much larger for mega-pixel images (Fig. 13). To further study the effect of image sizes on the running time, we down-sample a 5.4M-pixel image ( , the first row of Fig. 13) and its trimap to ten different scales and run all algorithms. The curves are shown in Fig time / s Ours image size / mega-pixel Figure 10. Image size vs. running time. (a) (b) (c) Figure 14. Mattes from strokes. (a) Images and strokes. (b) Results of the closed-form matting. (c) Our results. The images and stroke maps are from [12]. The time of our algorithm increases linearly with the image size, while the time of and increases super-linearly. Next, we compare the matte qualities. For medium images, the results are visually similar for fuzzy objects (Fig. 11). However, our method is better at handling holes, like Fig. 12 and the gap between the toys in Fig. 1. Further, in Fig. 13 we show that our large kernel method greatly improves the quality for high resolution images, as a large kernel collects enough samples to describe the color lines. In Fig. 14, we show the results of stroke-based constraints. The resulting mattes are visually very similar. Note that the stroke-based framework has been proposed to reduce the user s labor. But strokes lead to large areas of unknowns, which increase the running time and often make the whole interaction process even longer. Our method provides comparable results in very short time, so facilitates the feedback from the user. Table 2 shows the average rank of mean square error (MSE) on the benchmark data [1]. The MSE rank of our algorithm is comparable with the closed-from matting [12] and Robust Matting [24]. Note that [24, 14, 15] combine the matting Laplacian with data terms, while our method and [12] do not. We believe that our algorithm can also greatly reduce their running time and further advance the state-of-the-art in both efficiency and quality.
7 Fig 1 Fig 11 Fig 12 Fig 13 Fig 14 Fig 15 Fig 16 image size unknowns 180k 48k 49k 143k 57k 1280k 766k 857k 60k 80k 67k 86k 49s (20) 7.7s (10) 9.4s (10) 44s (14) 14s (11) 902s (32) 440s (28) 464s (31) 12.6s (16) 19s (19) 12.4s (11) 36s (51) time 29s (11) 7.1s (9) 6.8s (8) 28s (9) 5.9s (5) 352s (12) 256s (16) 206s (14) 2.4s (3) 4.7s (5) 11.6s (11) 5.7s (8) 2.5s 0.8s 0.9s 3.1s 1.2s 28s 15.7s 14.9s 0.8s 1s 1.1s 0.7s average time per k unknowns (𝜏 ) Table 1. Image sizes and running time. The numbers in the brackets are the ratios to our algorithm s running time. Solving the closed-form matting using brute force Conjugate Gradient () is very slow when the number of unknowns gets larger. The coarse-to-fine scheme () in [12] accelerates the stroke-based examples (Fig. 14 & 16), but still takes hundreds of seconds for high-resolution image (Fig. 13). Our method is about 5 to 20 times faster. Even mega-pixel images with mega unknowns (Fig. 13) only takes 15 to 30 seconds. image trimap ground truth Figure 11. Medium images of fuzzy objects. The images and trimaps are from [1] and [24]. input trimap Figure 12. Medium images of foreground objects with holes. The images and trimaps are from [1]. image trimap ground truth Figure 13. High resolution images. The images and trimaps are from [1]. We recommend to see the electronic version of this paper.
8 Improved color [14] 2 Random walk [7] 6.8 Our method 3.2 Geodesic [2] 7.5 Closed-form [12] 3.3 Bayesian [4] 8.8 Robust [24] 3.7 Easy matting [8] 9 High-res [15] 4.4 Poisson [19] 10.9 Iterative BP [22] 6.6 Table 2. Average rank of mean square error (MSE) on a data set from [1]. Please visit [1] for a comprehensive comparison. input trimap closed-form input closed-form Figure 15. Failure of the color line assumption. strokes closed-form additional strokes Figure 16. Failure case where the constraints are too sparse. Like the closed-form matting, our method may fail when the color line assumption is invalid, e.g., when the foreground/background colors are complicated (Fig. 15). Moreover, our method may give poorer results when the band of unknowns is too wide, because the large kernel increases the probability of breaking the color line assumption (Fig. 16). However, as our method runs quickly, it is convenient for the user to improve the result by refining the trimap. 6. Conclusion and Discussion In this paper, we propose a fast and effective algorithm for image matting. Our novel insight is that solving a large kernel matrix is not necessarily slow. It indeed takes fewer iterations to converge; we only need to focus on reducing the time in each iteration. Besides, a large kernel may also improve quality. This will encourage us to restudy a great category of matrices, like the homogenous Laplacian in Poisson image editing [13], Gaussian affinity Laplacian in colorization [11], or other edge-preserving matrices [5]. Generalizing their definitions to larger kernels will lead to new research topics. References [1] [2] X. Bai and G. Sapiro. A geodesic framework for fast interactive image and video segmentation and matting. ICCV, [3] A. Bousseau, S. Paris, and F. Durand. User-assisted intrinsic images. SIGGRAPH Asia, [4] Y. Chuang, B. Curless, D. Salesin, and R. Szeliski. A bayesian approach to digital matting. CVPR, [5] Z. Farbman, R. Fattal, D. Lischinski, and R. Szeliski. Edgepreserving decompositions for multi-scale tone and detail manipulation. SIGGRAPH, [6] C. Franklin. Summed-area tables for texture mapping. SIG- GRAPH, [7] L. Grady, T. Schiwietz, and S. Aharon. Random walks for interactive alpha-matting. VIIP, [8] Y. Guan, W. Cheny, X. Liang, Z. Ding, and Q. Peng. Easy matting: A stroke based approach for continuous image matting. Eurographics, [9] K. He, J. Sun, and X. Tang. Single image haze removal using dark channel prior. CVPR, [10] E. Hsu, T. Mertens, S. Paris, S. Avidan, and F. Durand. Light mixture estimation for spatially varying white balance. SIG- GRAPH, [11] A. Levin, D. Lischinski, and Y. Weiss. Colorization using optimization. SIGGRAPH, [12] A. Levin, D. Lischinski, and Y. Weiss. A closed-form solution to natural image matting. CVPR, [13] P. Perez, M. Gangnet, and A. Blake. Poisson image editing. SIGGRAPH, [14] C. Rhemann, C. Rother, and M. Gelautz. Improving color modeling for alpha matting. BMVC, [15] C. Rhemann, C. Rother, A. Rav-Acha, M. Gelautz, and T. Sharp. High resolution matting via interactive trimap segmentation. CVPR, [16] C. Rhemann, C. Rother, J. Wang, M. Gelautz, P. Kohli, and P. Rott. A perceptually motivated online benchmark for image matting. CVPR, [17] Y. Saad. Iterative methods for sparse linear systems, page 178. SIAM, [18] D. Singaraju, C. Rother, and C. Rhemann. New appearance models for natural image matting. CVPR, [19] J. Sun, J. Jia, C. Tang, and H. Shum. Poisson matting. SIG- GRAPH, [20] R. Szeliski. Locally adapted hierarchical basis preconditioning. SIGGRAPH, [21] J. Wang, M. Agrawala, and M. Cohen. Soft scissors: An interactive tool for realtime high quality matting. SIGGRAPH, [22] J. Wang and M. Cohen. An iterative optimization approach for unified image segmentation and matting. ICCV, [23] J. Wang and M. Cohen. Image and video matting: A survey. Foundations and Trends in Computer Graphics and Vision, [24] J. Wang and M. Cohen. Optimized color sampling for robust matting. CVPR, [25] Y. Zheng and C. Kambhamettu. Learning based digital matting. ICCV, 2009.
Computational Optical Imaging - Optique Numerique. -- Deconvolution --
Computational Optical Imaging - Optique Numerique -- Deconvolution -- Winter 2014 Ivo Ihrke Deconvolution Ivo Ihrke Outline Deconvolution Theory example 1D deconvolution Fourier method Algebraic method
High Quality Image Deblurring Panchromatic Pixels
High Quality Image Deblurring Panchromatic Pixels ACM Transaction on Graphics vol. 31, No. 5, 2012 Sen Wang, Tingbo Hou, John Border, Hong Qin, and Rodney Miller Presented by Bong-Seok Choi School of Electrical
Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall
Automatic Photo Quality Assessment Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Estimating i the photorealism of images: Distinguishing i i paintings from photographs h Florin
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
DATA ANALYSIS II. Matrix Algorithms
DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where
An Introduction to Machine Learning
An Introduction to Machine Learning L5: Novelty Detection and Regression Alexander J. Smola Statistical Machine Learning Program Canberra, ACT 0200 Australia [email protected] Tata Institute, Pune,
Fast Image Labeling for Creating High-Resolution Panoramic Images on Mobile Devices
2009 11th IEEE International Symposium on Multimedia Fast Image Labeling for Creating High-Resolution Panoramic Images on Mobile Devices Yingen Xiong and Kari Pulli Nokia Research Center Palo Alto, CA
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Mean-Shift Tracking with Random Sampling
1 Mean-Shift Tracking with Random Sampling Alex Po Leung, Shaogang Gong Department of Computer Science Queen Mary, University of London, London, E1 4NS Abstract In this work, boosting the efficiency of
Subspace Analysis and Optimization for AAM Based Face Alignment
Subspace Analysis and Optimization for AAM Based Face Alignment Ming Zhao Chun Chen College of Computer Science Zhejiang University Hangzhou, 310027, P.R.China [email protected] Stan Z. Li Microsoft
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections
Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections Maximilian Hung, Bohyun B. Kim, Xiling Zhang August 17, 2013 Abstract While current systems already provide
High Quality Image Magnification using Cross-Scale Self-Similarity
High Quality Image Magnification using Cross-Scale Self-Similarity André Gooßen 1, Arne Ehlers 1, Thomas Pralow 2, Rolf-Rainer Grigat 1 1 Vision Systems, Hamburg University of Technology, D-21079 Hamburg
Yousef Saad University of Minnesota Computer Science and Engineering. CRM Montreal - April 30, 2008
A tutorial on: Iterative methods for Sparse Matrix Problems Yousef Saad University of Minnesota Computer Science and Engineering CRM Montreal - April 30, 2008 Outline Part 1 Sparse matrices and sparsity
3D Model based Object Class Detection in An Arbitrary View
3D Model based Object Class Detection in An Arbitrary View Pingkun Yan, Saad M. Khan, Mubarak Shah School of Electrical Engineering and Computer Science University of Central Florida http://www.eecs.ucf.edu/
Analysis of Bayesian Dynamic Linear Models
Analysis of Bayesian Dynamic Linear Models Emily M. Casleton December 17, 2010 1 Introduction The main purpose of this project is to explore the Bayesian analysis of Dynamic Linear Models (DLMs). The main
Feature Tracking and Optical Flow
02/09/12 Feature Tracking and Optical Flow Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem Many slides adapted from Lana Lazebnik, Silvio Saverse, who in turn adapted slides from Steve
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
By choosing to view this document, you agree to all provisions of the copyright laws protecting it.
This material is posted here with permission of the IEEE Such permission of the IEEE does not in any way imply IEEE endorsement of any of Helsinki University of Technology's products or services Internal
Probabilistic Models for Big Data. Alex Davies and Roger Frigola University of Cambridge 13th February 2014
Probabilistic Models for Big Data Alex Davies and Roger Frigola University of Cambridge 13th February 2014 The State of Big Data Why probabilistic models for Big Data? 1. If you don t have to worry about
Vision based Vehicle Tracking using a high angle camera
Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu [email protected] [email protected] Abstract A vehicle tracking and grouping algorithm is presented in this work
Robert Collins CSE598G. More on Mean-shift. R.Collins, CSE, PSU CSE598G Spring 2006
More on Mean-shift R.Collins, CSE, PSU Spring 2006 Recall: Kernel Density Estimation Given a set of data samples x i ; i=1...n Convolve with a kernel function H to generate a smooth function f(x) Equivalent
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
Tensor Methods for Machine Learning, Computer Vision, and Computer Graphics
Tensor Methods for Machine Learning, Computer Vision, and Computer Graphics Part I: Factorizations and Statistical Modeling/Inference Amnon Shashua School of Computer Science & Eng. The Hebrew University
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 269 Class Project Report
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 69 Class Project Report Junhua Mao and Lunbo Xu University of California, Los Angeles [email protected] and lunbo
Enhanced LIC Pencil Filter
Enhanced LIC Pencil Filter Shigefumi Yamamoto, Xiaoyang Mao, Kenji Tanii, Atsumi Imamiya University of Yamanashi {[email protected], [email protected], [email protected]}
Bayesian Image Super-Resolution
Bayesian Image Super-Resolution Michael E. Tipping and Christopher M. Bishop Microsoft Research, Cambridge, U.K..................................................................... Published as: Bayesian
On the Interaction and Competition among Internet Service Providers
On the Interaction and Competition among Internet Service Providers Sam C.M. Lee John C.S. Lui + Abstract The current Internet architecture comprises of different privately owned Internet service providers
Detection and Restoration of Vertical Non-linear Scratches in Digitized Film Sequences
Detection and Restoration of Vertical Non-linear Scratches in Digitized Film Sequences Byoung-moon You 1, Kyung-tack Jung 2, Sang-kook Kim 2, and Doo-sung Hwang 3 1 L&Y Vision Technologies, Inc., Daejeon,
Fitting Subject-specific Curves to Grouped Longitudinal Data
Fitting Subject-specific Curves to Grouped Longitudinal Data Djeundje, Viani Heriot-Watt University, Department of Actuarial Mathematics & Statistics Edinburgh, EH14 4AS, UK E-mail: [email protected] Currie,
The primary goal of this thesis was to understand how the spatial dependence of
5 General discussion 5.1 Introduction The primary goal of this thesis was to understand how the spatial dependence of consumer attitudes can be modeled, what additional benefits the recovering of spatial
Topological Data Analysis Applications to Computer Vision
Topological Data Analysis Applications to Computer Vision Vitaliy Kurlin, http://kurlin.org Microsoft Research Cambridge and Durham University, UK Topological Data Analysis quantifies topological structures
Face Model Fitting on Low Resolution Images
Face Model Fitting on Low Resolution Images Xiaoming Liu Peter H. Tu Frederick W. Wheeler Visualization and Computer Vision Lab General Electric Global Research Center Niskayuna, NY, 1239, USA {liux,tu,wheeler}@research.ge.com
Computer Graphics. Geometric Modeling. Page 1. Copyright Gotsman, Elber, Barequet, Karni, Sheffer Computer Science - Technion. An Example.
An Example 2 3 4 Outline Objective: Develop methods and algorithms to mathematically model shape of real world objects Categories: Wire-Frame Representation Object is represented as as a set of points
2.2 Creaseness operator
2.2. Creaseness operator 31 2.2 Creaseness operator Antonio López, a member of our group, has studied for his PhD dissertation the differential operators described in this section [72]. He has compared
Interactive Offline Tracking for Color Objects
Interactive Offline Tracking for Color Objects Yichen Wei Jian Sun Xiaoou Tang Heung-Yeung Shum Microsoft Research Asia, Beijing, China {yichenw,jiansun,xitang,hshum}@microsoft.com Abstract In this paper,
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY
PHOTOGRAMMETRIC TECHNIQUES FOR MEASUREMENTS IN WOODWORKING INDUSTRY V. Knyaz a, *, Yu. Visilter, S. Zheltov a State Research Institute for Aviation System (GosNIIAS), 7, Victorenko str., Moscow, Russia
LABEL PROPAGATION ON GRAPHS. SEMI-SUPERVISED LEARNING. ----Changsheng Liu 10-30-2014
LABEL PROPAGATION ON GRAPHS. SEMI-SUPERVISED LEARNING ----Changsheng Liu 10-30-2014 Agenda Semi Supervised Learning Topics in Semi Supervised Learning Label Propagation Local and global consistency Graph
Efficient online learning of a non-negative sparse autoencoder
and Machine Learning. Bruges (Belgium), 28-30 April 2010, d-side publi., ISBN 2-93030-10-2. Efficient online learning of a non-negative sparse autoencoder Andre Lemme, R. Felix Reinhart and Jochen J. Steil
EECS 556 Image Processing W 09. Interpolation. Interpolation techniques B splines
EECS 556 Image Processing W 09 Interpolation Interpolation techniques B splines What is image processing? Image processing is the application of 2D signal processing methods to images Image representation
Tracking Moving Objects In Video Sequences Yiwei Wang, Robert E. Van Dyck, and John F. Doherty Department of Electrical Engineering The Pennsylvania State University University Park, PA16802 Abstract{Object
(Quasi-)Newton methods
(Quasi-)Newton methods 1 Introduction 1.1 Newton method Newton method is a method to find the zeros of a differentiable non-linear function g, x such that g(x) = 0, where g : R n R n. Given a starting
Spatial Statistics Chapter 3 Basics of areal data and areal data modeling
Spatial Statistics Chapter 3 Basics of areal data and areal data modeling Recall areal data also known as lattice data are data Y (s), s D where D is a discrete index set. This usually corresponds to data
Binary Image Reconstruction
A network flow algorithm for reconstructing binary images from discrete X-rays Kees Joost Batenburg Leiden University and CWI, The Netherlands [email protected] Abstract We present a new algorithm
General Framework for an Iterative Solution of Ax b. Jacobi s Method
2.6 Iterative Solutions of Linear Systems 143 2.6 Iterative Solutions of Linear Systems Consistent linear systems in real life are solved in one of two ways: by direct calculation (using a matrix factorization,
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
MVA ENS Cachan. Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos [email protected]
Machine Learning for Computer Vision 1 MVA ENS Cachan Lecture 2: Logistic regression & intro to MIL Iasonas Kokkinos [email protected] Department of Applied Mathematics Ecole Centrale Paris Galen
AN INTRODUCTION TO NUMERICAL METHODS AND ANALYSIS
AN INTRODUCTION TO NUMERICAL METHODS AND ANALYSIS Revised Edition James Epperson Mathematical Reviews BICENTENNIAL 0, 1 8 0 7 z ewiley wu 2007 r71 BICENTENNIAL WILEY-INTERSCIENCE A John Wiley & Sons, Inc.,
Metrics on SO(3) and Inverse Kinematics
Mathematical Foundations of Computer Graphics and Vision Metrics on SO(3) and Inverse Kinematics Luca Ballan Institute of Visual Computing Optimization on Manifolds Descent approach d is a ascent direction
ANALYSIS, THEORY AND DESIGN OF LOGISTIC REGRESSION CLASSIFIERS USED FOR VERY LARGE SCALE DATA MINING
ANALYSIS, THEORY AND DESIGN OF LOGISTIC REGRESSION CLASSIFIERS USED FOR VERY LARGE SCALE DATA MINING BY OMID ROUHANI-KALLEH THESIS Submitted as partial fulfillment of the requirements for the degree of
The Characteristic Polynomial
Physics 116A Winter 2011 The Characteristic Polynomial 1 Coefficients of the characteristic polynomial Consider the eigenvalue problem for an n n matrix A, A v = λ v, v 0 (1) The solution to this problem
State of Stress at Point
State of Stress at Point Einstein Notation The basic idea of Einstein notation is that a covector and a vector can form a scalar: This is typically written as an explicit sum: According to this convention,
Least Squares Estimation
Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David
Binary Image Scanning Algorithm for Cane Segmentation
Binary Image Scanning Algorithm for Cane Segmentation Ricardo D. C. Marin Department of Computer Science University Of Canterbury Canterbury, Christchurch [email protected] Tom
Fast Multipole Method for particle interactions: an open source parallel library component
Fast Multipole Method for particle interactions: an open source parallel library component F. A. Cruz 1,M.G.Knepley 2,andL.A.Barba 1 1 Department of Mathematics, University of Bristol, University Walk,
COLOR-BASED PRINTED CIRCUIT BOARD SOLDER SEGMENTATION
COLOR-BASED PRINTED CIRCUIT BOARD SOLDER SEGMENTATION Tz-Sheng Peng ( 彭 志 昇 ), Chiou-Shann Fuh ( 傅 楸 善 ) Dept. of Computer Science and Information Engineering, National Taiwan University E-mail: [email protected]
Sachin Patel HOD I.T Department PCST, Indore, India. Parth Bhatt I.T Department, PCST, Indore, India. Ankit Shah CSE Department, KITE, Jaipur, India
Image Enhancement Using Various Interpolation Methods Parth Bhatt I.T Department, PCST, Indore, India Ankit Shah CSE Department, KITE, Jaipur, India Sachin Patel HOD I.T Department PCST, Indore, India
Numerical Methods I Solving Linear Systems: Sparse Matrices, Iterative Methods and Non-Square Systems
Numerical Methods I Solving Linear Systems: Sparse Matrices, Iterative Methods and Non-Square Systems Aleksandar Donev Courant Institute, NYU 1 [email protected] 1 Course G63.2010.001 / G22.2420-001,
Voronoi Treemaps in D3
Voronoi Treemaps in D3 Peter Henry University of Washington [email protected] Paul Vines University of Washington [email protected] ABSTRACT Voronoi treemaps are an alternative to traditional rectangular
How I won the Chess Ratings: Elo vs the rest of the world Competition
How I won the Chess Ratings: Elo vs the rest of the world Competition Yannis Sismanis November 2010 Abstract This article discusses in detail the rating system that won the kaggle competition Chess Ratings:
Introduction to General and Generalized Linear Models
Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby
A Fast Algorithm for Multilevel Thresholding
JOURNAL OF INFORMATION SCIENCE AND ENGINEERING 17, 713-727 (2001) A Fast Algorithm for Multilevel Thresholding PING-SUNG LIAO, TSE-SHENG CHEN * AND PAU-CHOO CHUNG + Department of Electrical Engineering
10.2 ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS. The Jacobi Method
578 CHAPTER 1 NUMERICAL METHODS 1. ITERATIVE METHODS FOR SOLVING LINEAR SYSTEMS As a numerical technique, Gaussian elimination is rather unusual because it is direct. That is, a solution is obtained after
Bildverarbeitung und Mustererkennung Image Processing and Pattern Recognition
Bildverarbeitung und Mustererkennung Image Processing and Pattern Recognition 1. Image Pre-Processing - Pixel Brightness Transformation - Geometric Transformation - Image Denoising 1 1. Image Pre-Processing
An Iterative Image Registration Technique with an Application to Stereo Vision
An Iterative Image Registration Technique with an Application to Stereo Vision Bruce D. Lucas Takeo Kanade Computer Science Department Carnegie-Mellon University Pittsburgh, Pennsylvania 15213 Abstract
Image Super-Resolution Using Deep Convolutional Networks
1 Image Super-Resolution Using Deep Convolutional Networks Chao Dong, Chen Change Loy, Member, IEEE, Kaiming He, Member, IEEE, and Xiaoou Tang, Fellow, IEEE arxiv:1501.00092v3 [cs.cv] 31 Jul 2015 Abstract
Highlight Removal by Illumination-Constrained Inpainting
Highlight Removal by Illumination-Constrained Inpainting Ping Tan Stephen Lin Long Quan Heung-Yeung Shum Microsoft Research, Asia Hong Kong University of Science and Technology Abstract We present a single-image
Class #6: Non-linear classification. ML4Bio 2012 February 17 th, 2012 Quaid Morris
Class #6: Non-linear classification ML4Bio 2012 February 17 th, 2012 Quaid Morris 1 Module #: Title of Module 2 Review Overview Linear separability Non-linear classification Linear Support Vector Machines
GI01/M055 Supervised Learning Proximal Methods
GI01/M055 Supervised Learning Proximal Methods Massimiliano Pontil (based on notes by Luca Baldassarre) (UCL) Proximal Methods 1 / 20 Today s Plan Problem setting Convex analysis concepts Proximal operators
Convolution. 1D Formula: 2D Formula: Example on the web: http://www.jhu.edu/~signals/convolve/
Basic Filters (7) Convolution/correlation/Linear filtering Gaussian filters Smoothing and noise reduction First derivatives of Gaussian Second derivative of Gaussian: Laplacian Oriented Gaussian filters
Component Ordering in Independent Component Analysis Based on Data Power
Component Ordering in Independent Component Analysis Based on Data Power Anne Hendrikse Raymond Veldhuis University of Twente University of Twente Fac. EEMCS, Signals and Systems Group Fac. EEMCS, Signals
Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus
Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus Tihomir Asparouhov and Bengt Muthén Mplus Web Notes: No. 15 Version 8, August 5, 2014 1 Abstract This paper discusses alternatives
Color Segmentation Based Depth Image Filtering
Color Segmentation Based Depth Image Filtering Michael Schmeing and Xiaoyi Jiang Department of Computer Science, University of Münster Einsteinstraße 62, 48149 Münster, Germany, {m.schmeing xjiang}@uni-muenster.de
Digital Imaging and Multimedia. Filters. Ahmed Elgammal Dept. of Computer Science Rutgers University
Digital Imaging and Multimedia Filters Ahmed Elgammal Dept. of Computer Science Rutgers University Outlines What are Filters Linear Filters Convolution operation Properties of Linear Filters Application
P164 Tomographic Velocity Model Building Using Iterative Eigendecomposition
P164 Tomographic Velocity Model Building Using Iterative Eigendecomposition K. Osypov* (WesternGeco), D. Nichols (WesternGeco), M. Woodward (WesternGeco) & C.E. Yarman (WesternGeco) SUMMARY Tomographic
Interactive Level-Set Deformation On the GPU
Interactive Level-Set Deformation On the GPU Institute for Data Analysis and Visualization University of California, Davis Problem Statement Goal Interactive system for deformable surface manipulation
Segmentation of building models from dense 3D point-clouds
Segmentation of building models from dense 3D point-clouds Joachim Bauer, Konrad Karner, Konrad Schindler, Andreas Klaus, Christopher Zach VRVis Research Center for Virtual Reality and Visualization, Institute
DYNAMIC RANGE IMPROVEMENT THROUGH MULTIPLE EXPOSURES. Mark A. Robertson, Sean Borman, and Robert L. Stevenson
c 1999 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or
Recognizing Cats and Dogs with Shape and Appearance based Models. Group Member: Chu Wang, Landu Jiang
Recognizing Cats and Dogs with Shape and Appearance based Models Group Member: Chu Wang, Landu Jiang Abstract Recognizing cats and dogs from images is a challenging competition raised by Kaggle platform
A Noise-Aware Filter for Real-Time Depth Upsampling
A Noise-Aware Filter for Real-Time Depth Upsampling Derek Chan Hylke Buisman Christian Theobalt Sebastian Thrun Stanford University, USA Abstract. A new generation of active 3D range sensors, such as time-of-flight
Environmental Remote Sensing GEOG 2021
Environmental Remote Sensing GEOG 2021 Lecture 4 Image classification 2 Purpose categorising data data abstraction / simplification data interpretation mapping for land cover mapping use land cover class
Real Time Target Tracking with Pan Tilt Zoom Camera
2009 Digital Image Computing: Techniques and Applications Real Time Target Tracking with Pan Tilt Zoom Camera Pankaj Kumar, Anthony Dick School of Computer Science The University of Adelaide Adelaide,
Analysis of Multi-Spacecraft Magnetic Field Data
COSPAR Capacity Building Beijing, 5 May 2004 Joachim Vogt Analysis of Multi-Spacecraft Magnetic Field Data 1 Introduction, single-spacecraft vs. multi-spacecraft 2 Single-spacecraft data, minimum variance
Edge tracking for motion segmentation and depth ordering
Edge tracking for motion segmentation and depth ordering P. Smith, T. Drummond and R. Cipolla Department of Engineering University of Cambridge Cambridge CB2 1PZ,UK {pas1001 twd20 cipolla}@eng.cam.ac.uk
System Identification for Acoustic Comms.:
System Identification for Acoustic Comms.: New Insights and Approaches for Tracking Sparse and Rapidly Fluctuating Channels Weichang Li and James Preisig Woods Hole Oceanographic Institution The demodulation
D-optimal plans in observational studies
D-optimal plans in observational studies Constanze Pumplün Stefan Rüping Katharina Morik Claus Weihs October 11, 2005 Abstract This paper investigates the use of Design of Experiments in observational
Notes on Symmetric Matrices
CPSC 536N: Randomized Algorithms 2011-12 Term 2 Notes on Symmetric Matrices Prof. Nick Harvey University of British Columbia 1 Symmetric Matrices We review some basic results concerning symmetric matrices.
Image Segmentation and Registration
Image Segmentation and Registration Dr. Christine Tanner ([email protected]) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation
December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B. KITCHENS
December 4, 2013 MATH 171 BASIC LINEAR ALGEBRA B KITCHENS The equation 1 Lines in two-dimensional space (1) 2x y = 3 describes a line in two-dimensional space The coefficients of x and y in the equation
Exact Inference for Gaussian Process Regression in case of Big Data with the Cartesian Product Structure
Exact Inference for Gaussian Process Regression in case of Big Data with the Cartesian Product Structure Belyaev Mikhail 1,2,3, Burnaev Evgeny 1,2,3, Kapushev Yermek 1,2 1 Institute for Information Transmission
An Overview Of Software For Convex Optimization. Brian Borchers Department of Mathematics New Mexico Tech Socorro, NM 87801 borchers@nmt.
An Overview Of Software For Convex Optimization Brian Borchers Department of Mathematics New Mexico Tech Socorro, NM 87801 [email protected] In fact, the great watershed in optimization isn t between linearity
University of Lille I PC first year list of exercises n 7. Review
University of Lille I PC first year list of exercises n 7 Review Exercise Solve the following systems in 4 different ways (by substitution, by the Gauss method, by inverting the matrix of coefficients
CS 5614: (Big) Data Management Systems. B. Aditya Prakash Lecture #18: Dimensionality Reduc7on
CS 5614: (Big) Data Management Systems B. Aditya Prakash Lecture #18: Dimensionality Reduc7on Dimensionality Reduc=on Assump=on: Data lies on or near a low d- dimensional subspace Axes of this subspace
TESLA Report 2003-03
TESLA Report 23-3 A multigrid based 3D space-charge routine in the tracking code GPT Gisela Pöplau, Ursula van Rienen, Marieke de Loos and Bas van der Geer Institute of General Electrical Engineering,
