Segmentation-Free Dynamic Scene Deblurring
|
|
|
- Harry Carson
- 9 years ago
- Views:
Transcription
1 Segmentation-Free Dynamic Scene Deblurring Tae Hyun Kim and Kyoung Mu Lee Department of ECE, ASRI, Seoul National University, , Seoul, Korea {lliger9, (a) (c) Figure 1: (a) Blurry image Color coded motion flow (c) Our deblurring result Abstract Most state-of-the-art dynamic scene deblurring methods based on accurate motion segmentation assume that motion blur is small or that the specific type of motion causing the blur is known. In this paper, we study a motion segmentation-free dynamic scene deblurring method, which is unlike other conventional methods. When the motion can be approimated to linear motion that is locally (piel-wise) varying, we can handle various types of blur caused by camera shake, including out-of-plane motion, depth variation, radial distortion, and so on. Thus, we propose a new energy model simultaneously estimating motion flow and the latent image based on robust total variation (TV)-L1 model. This approach is necessary to handle abrupt changes in motion without segmentation. Furthermore, we address the problem of the traditional coarse-to-fine deblurring framework, which gives rise to artifacts when restoring small structures with distinct motion. We thus propose a novel kernel re-initialization method which reduces the error of motion flow propagated from a coarser level. Moreover, a highly effective conve optimization-based solution mitigating the computational difficulties of the TV-L1 model is established. Comparative eperimental results on challenging real blurry images demonstrate the efficiency of the proposed method. 1. Introduction Blind single-image deblurring is a method used to restore a sharp image from an image blurred by camera shake or object motion under low light conditions. This approach has become an active research topic in computer vision because of the recent demand for clear images. However, deblurring is difficult to solve because the problem is highly ill-posed. To solve this problem, many researchers have studied the joint estimation of latent image and blur kernel, thus recasting the deblurring problem as an energy minimization problem based on the general constraint: B = KL + N, (1) where L, B, and N denote the vector forms of the latent image, blurred image, and noise, respectively. The matri K denotes the blur kernel with a row vector corresponding to the blur kernel placed at each piel location. There are various kernel estimation methods depending on the types of motion blur. The 2D translational camera shake is the most comprehensively studied motion [3, 6, 24]. Such camera motion creates a spatially invariant blur kernel, such that the matri operation in (1) can be epressed in a convolution form with the aid of linearity. Therefore, using fast Fourier transform makes fast kernel estimation from 2D translational camera shake possible. Moreover, camera shake, including rotational movement, generalizes the 2D translational camera motion and it enables spatially varying blur by camera rotation [8, 10, 23]. Although these efforts to model real camera motion have yielded promising results
2 (a) (c) (d) Figure 2: (a) Partially blurred image by forward motion. -(c) Segmentation and deblurring results of [13] (d) Our deblurring result. Notably, the taillight and rear fender of the tai are restored to a significantly better degree than those in (c). in the restoration of blurry images through the motion of a camera, applying these approaches to cases where the blur does not come from the global camera motion is difficult. By contrast, without knowing the specific type of motion blur, traditional deterministic filter-based approaches can handle various types of blur [1, 7], but these filters are incapable of handling large motion blur and are sensitive to noise. Additionally, edge statistics can be also used to estimate locally varying blur kernels [12], but, the method is also restricted to small motion blur. Therefore, some researchers have studied methods that can be used to deal with large blur without knowing the specific type of motion blur. Harmeling et al. [9] proposed a method based on the assumption that local blur kernel is epressed as a weighted sum of the neighboring blur kernels and Ji et al. [11] interpolated initially given kernels robustly. Although these approaches can handle smoothly varying blur kernel, these methods cannot handle abrupt changes in the blur kernel, which commonly occur in a dynamic scene containing multiple moving objects. To address this problem, some researches have focused on deblurring dynamic scenes and established approaches that commonly require accurate motion segmentation. In [16], Levin segmented blur motions by comparing likelihoods with a set of given 1D bo filters. However, limited number of 1D bo filters were used. Thus, the poor segmentation could cause undesirable artifacts. CouzinieDevy et al. etended the work of Levin in [4] by casting the problem as a multi-label segmentation problem and estimating locally varying blur. However, the method could not handle large blur because of the eponential increases in the the number of candidate labels. This condition restricts the blur kernels to small 2D Gaussian or linear. The recent work of Kim et al. [13] proposed a method to estimate blur kernels, latent image, and motion segments jointly. The eistence of camera shake, including rotation and 2D translational motion of objects, was assumed, but the method also fails in both segmentation and deblurring when the captured image is blurred by unepected blur effects, such as forward motion, depth variation, or radial distortion. For eample, the tai is blurred by its forward motion, and such an unepected motion causes a failure of both segmentation and deblurring as shown in Figs. 2 and 2(c). In sum, state-of-the-art dynamic scene deblurring methods require accurate motion segmentation for specific types of motion blur. To mitigate this restriction, we propose a method to deblur dynamic scenes without segmentation and without restricting the types of motion blur, when the locally varying blur kernels can be approimated to 2D motion vectors. Although this restriction ecludes non-linear motion, numerous types of motions can be linearized in practical situations [4, 5, 16]. We observed that this assumption holds for many real blurry images and that this approach works even when small rotational camera shake eists, as shown in our eperiments. In the previous work of Dai et al. [5], the authors estimated motion flow via the alpha channel [17] of the blurry image. However, they used a constraint that is different from ours and applied implicit segmentation based on RANSAC. In addition, the result depends on the accuracy of the given alpha channel. By contrast, we propose a new framework that enables abrupt changes in motion without segmentation based on the robust TV-L1 model. As illustrated in Fig.1, the proposed approach estimates the motion flow and latent image. Our method does not require any comple model to handle specific types of motion blur [10, 23, 27] and does not depend on the accurate motion segmentation. In addition, the proposed method is embedded into the traditional coarse-to-fine framework to handle large blur. Our finding is that small structures with distinct motion blur give rise to serious artifacts in the coarse-tofine framework. Thus, we also propose a novel method to re-initialize the motion flow and reduce the error propagated in the coarse to fine framework. In this paper, we introduce a new deblurring framework that jointly estimates the spatially varying motion flow and the latent image. We also provide a highly practical solver optimizing TV-L1 model based on conve optimization. As shown in Fig. 2(d), we achieve better results with the aid of accurately estimated motion flow even when the stateof-the-art methods fail. We demonstrate the effectiveness and practicality of our new deblurring framework with test results on challenging real images on which conventional techniques fail.
3 2. Dynamic Scene Deblurring Model To solve the ill-posed blind deblurring problem, various energy models that are composed of the regularization and data terms have been proposed to estimate the sharp image and the blur kernel in the following form E = E data (L, K, B) + E reg (L, K), (2) where the data term E data measures the data fidelity, and the regularization term E reg enforces the smoothness constraint to the latent image as well as to the blur kernel. In [13], the authors argued that conventional energy model in (2) is invalid for dynamic scene deblurring because it principally requires motion segmentation. However, in this paper, we propose a new energy model based on (2) for dynamic scene deblurring. The proposed model does not require eplicit motion segmentation. Further details on the model are described in the following sections L1 based Robust Blur Model ( 2, 2 ) (, ) y ( 2, 2 ) Figure 3: Blur kernel corresponding to motion flow u In our study, we assumed that the locally varying blur kernel can be linearized in terms of a 2D motion vector, u = (u, v) T, which can be epressed as h(, y) = { 1 u δ(v uy), 0, otherwise. u if 2, y v 2, (3) where h(, y) denotes the blur kernel corresponding to motion flow u as illustrated in Fig. 3, and δ is the Dirac delta function. Thus, we can cast the kernel estimation problem as a motion estimation problem. The data term of our new energy model is given by E data (L, u, B) = λ (k(, u)) T L B(), (4) where R 2 is the inde of the discrete locations in the image domain, and (k(, u)) T denotes the discretized vector form of h with a sum of elements equal to one. This vector also corresponds to a row vector of the kernel matri, K, where indicates the inde of the row. The operator {, y } denotes the partial derivatives in the horizontal and vertical directions [3, 13]. The parameter λ controls the importance of the data term. (a) (c) Figure 4: Comparison of motion flow with and without edgemap. (a) Blurry image Color coded motion flow without edgemap (c) Color coded motion flow with edge-map As mentioned in [13], we use only gradient maps of the latent and blurry images in the data model to reduce ringing artifacts. In addition, the proposed model must enable abrupt changes in the blur kernel near motion boundary because the goal is deblurring a dynamic scene without segmentation. Therefore, we propose an L1-based model in the data term that allows discontinuities in the flow field. Although the robust L1 model was proposed before to estimate the latent image in a non-blind deblurring process [24], this model has not been used in the estimation of blur kernel, to the best of our knowledge, because of its computational difficulties. Despite its difficulties, we adopt the L1 model for estimating not only the latent image but also the blur kernel TV based Robust Regularization The smoothness of variables must be enforced to obtain a reliable solution when the problem is ill-posed. In our energy model, two primal unknown variables are the latent image L and the motion flow u, and each has different kinds of regularization, as given by E reg (L, u) = E reg (L) + E reg (u). (5) More details on the regularization of each variable are described in the following sections Regularization on Latent Image In (1), noise is added to our blurry image. Thus, we should suppress the noise in the latent image while preserving edges. To address this problem, many researchers studied the prior of latent image [14, 15, 18]. The sparse gradient prior is known to describe the edge statistics of natural images well. In the same vein, we adopt the TV model to regularize the latent image, as in [13, 24]. The formulation is given by, E reg (L) = L. (6) Edge-aware Regularization on Motion Flow To have similar blur kernels among neighboring piels, we regularize the motion flow. Similar to the proposed data term in the preceding section, the key to regularizing the motion flow is how well the motion boundary is preserved.
4 To this end, we propose an edge-map coupled TV model to regularize conditionally, which yields E reg (u) = µ g() u, (7) where the parameter µ controls the strength of regularization, and the edge-map, g(), measuring the similarity between neighboring piels is defined as g() = ep( ( ma( B, L 0 ) σ I ) 2 ). (8) The parameter σ I denotes the decay eponent and L 0 is an intermediate latent image propagated from a coarser level in the coarse-to-fine framework. In particular, the blurred image itself is also used to obtain the edge-map, because an unblurred region may eist in a partially blurred image, and it works when L 0 is smoothed out. The outstanding performance of edge-aware regularization is demonstrated in Fig. 4. As epected, the proposed regularizer preserves motion discontinuities much better than the model without edge-map. 3. Optimization Framework The proposed single-image blind deblurring model introduced in the previous sections includes the data and regularization terms, and the objective function is given by min L +µ g() u +λ (k(, u)) T L B(). L,u (9) Our final objective in (9) is highly comple, that is, to obtain a global optimum. Thus we divide the original problem into two easier subproblems for L and u. We then apply alternating optimization procedure, which was introduced in [3, 21, 26]. The overall algorithm of the proposed method is given in Algorithm 1. Moreover, our approach is embedded into the traditional coarse-to-fine framework [6, 19] to handle large motion blur. However, we found that it gives rise to severe artifacts when applied to a dynamic scene, which has small structures with distinct motion blurs. Thus, we propose a novel re-initialization method to solve this problem. Further detail is given in the following section Kernel Re initialization The conventional coarse-to-fine approach is widely used in various fields and also has been shown to yield successful results in the deblurring of static scene. This method can be used in dynamic scene deblurring, but it has limitations that have not been observed in the restoration of a static scene. For eample, a small structure, such as a thin line, that has distinctive motion blur, cannot be seen at the coarser level. Therefore, the motion flow of such structure is estimated from its neighbors, which ehibit different motion, Algorithm 1 Overview of the proposed method Input: A blurry image B Output: Latent image L and motion flow u 1: Build an image pyramid with 10 levels and a scale factor of 0.8 2: Kernel re-initialization. (Sec. 3.1) 3: for t = 1 to 3 do 4: Continuous optimization of motion flow u with fied L. (Sec. 3.2) 5: Continuous optimization of latent image L with fied u. (Sec. 3.3) 6: end for 7: Propagate variables to the net pyramid level if eists. 8: Repeat steps 2-7 from coarse to fine pyramid level. with the aid of regularization. However, this process generates reconstruction error toward the finer level when the small structure appears suddenly in the blurry image. Notably, a similar problem has been reported in other vision applications [22, 25]. Local and sparse reconstruction error may not raise severe artifacts in the estimation of global camera motion, which is more serious in a local approach, such as the method proposed in this study. To mitigate this problem, we propose a novel kernel reinitialization method for both motion flow and the latent image. The key idea is to detect the erroneous region and deblur it using a deterministic filter [20], after which accurate motion flow is estimated. Hence, we cast the problem of detecting the erroneous region as a labeling problem, that is, min e e()(c (k(, u)) T L B() ) + 1 e(), (10) ν where the vector variable e is piel-wise binary indicator variable, e() {0, 1}, and the constant c is a positive threshold value. The first term introduced in (10) is the likelihood term that causes e() to be one when the reconstruction error, (k(, u)) T L B(), is higher than c. The second term is a prior giving the sparsity on the variable e, as we assume high reconstruction error is sparsely distributed, and the parameter ν controls the importance of the prior. Through the continuous relaation of e, we can obtain the approimated solution of (10) quickly by adapting the first-order primal dual algorithm [2], that is, r() n+1 r() = n +σe() n ma(1,r() n +σe() n ) e() n+1 (e (e() n τr() n+1 )) 2 = arg min + e 2τ ν e()(c (k(, u)) T L B() ) e() n+1 = ma(0, min(1, e() n+1 )), (11) where n 0 indicates the iteration number, vector r denotes the dual variable of e and the positive update steps σ and τ control the convergence rate [2].
5 (a) where u 0 is an initial motion flow of u, and ρ(, u 0 ) denotes the first-order derivative. Through approimation, the proposed energy model becomes conve near u 0, which results in u = arg min u µ g() u +λ(ρ(, u 0 )+ ρ(, u 0 ) T (u u 0 )), (14) To solve this, we adapted the conve optimization algorithm in [2], and the primal dual update process is given by (c) (d) (e) Figure 5: (a) Estimated latent image in a coarse level Yellow color denotes the detected erroneous region. (c) Cropped result in the coarse level (d) Cropped result in the finest level without the use of re-initialization (e) Cropped result in the finest level with the use of re-initialization After detection of the erroneous region, we re-initialize the propagated motion of this region to be zero, which denotes the impulse blur kernel. By applying sharp image restoration in section 3.3, we restore the small structure with distinct blur motion. However, the newly restored small structure remains blurry because it is estimated from the impulse blur kernel. Thus, we apply a deterministic filter [20] to deblur the structure and to facilitate the fast convergence of motion estimation. For this process, we use the prediction step introduced in [3]. The overall process is shown in Algorithm 2. The necessity of our re-initialization step is illustrated in Fig. 5. Notably an unseen thin line in the coarse level in Fig. 5(c) is successfully restored using our proposed method, as compared in Figs. 5(d) to (e). Algorithm 2 Kernel re-initialization algorithm Input: A blurry image B, intermediate latent image L and motion flow u propagated from coarser pyramid level. Output: Re-initialized motion flow u and edge-enhanced L 1: for n = 1 to 50 do 2: Compute e using (11) 3: end for 4: u() 0, if e()> 0.5 5: Continuous optimization of L with re-initialized u (Sec. 3.3) 6: Enhance edge by applying prediction step in [3] 3.2. Motion Flow Estimation For the latent image L being fied, the proposed energy model in (9) is simplified, but the data term ρ(, u) is nonconve in the argument u, which makes the optimization intractable ρ(, u) = (k(, u)) T L B(). (12) To make the optimization tractable, we linearize the data function via the Taylor approimation to obtain ρ(, u) ρ(, u 0 ) + ρ(, u 0 ) T (u u 0 ), (13) { p n+1 p = n +σ(ga)u n ma(1,p n +σ(ga)u n ) u n+1 = (u n τ(ga) T p n+1 ) τ( λ µ ) ρ(, u 0), (15) where p denotes the dual variable of u on the vector space. The continuous linear operator A calculates the difference between neighboring piels, and the diagonal matri G is the weighting matri denoted as G = diag(g()). This update process is easy to implement and converges quickly (n=100) Sharp Image Estimation For the motion flow u and corresponding kernel matri K being fied, the proposed energy model becomes a wellknown non-blind deblurring model. However, as we use the L1 model in the proposed data term, which requires high computations. To address this problem, we adopt the quadratic relaation method [22] to facilitate the solution. Thus, we introduce auiliary variables f, which yields min L,f L + λ f () 1 ((k(, u)) T L B() + f ()) 2. 2θ (16) If the fied parameter θ is set to a very small value, then the minimization of (16) is close to that of its original TV- L1 model. Using decomposition, both the function of L and functions of f become a thousand times easier to solve, and each variable is solved by [2] in the alternating optimization process. The update process for f while L being fied is epressed as q () n+1 q = () n +σf () n ma(1,q () n +σf () n ) f () n+1 (f (f () n τq = arg min () n+1 )) 2 + f 2τ 1 ((k(, u)) T L B() + f ()) 2, 2θλ (17) where the vector q is the dual variable of f. Similarly, the update process for L while f being fied
6 is given by s m+1 s = m +σ L AL m ma(1,s m +σ L AL m ) L m+1 (L (L m τ L A T s m+1 )) 2 = arg min + L 2τ L 1 (K L B + f ) 2, 2θλ (18) where m 0 indicates iteration number, the parameters σ L and τ L denote the update steps and the vector s is the dual variable of L. In addition, we adopt conjugate gradient (CG) method to update primal variable, L, in (18). The overall procedure of our sharp image estimation is shown in Algorithm 3. Algorithm 3 Sharp Image Estimation Input: A blurry image B and motion flow u Output: Sharp image L 1: for n = 1 to 3 do 2: Update f using (17) 3: for m = 1 to 30 do 4: Update L using (18) 5: end for 6: end for 3.4. Implementation In implementation, we use fied parameters for most eperiments ecept for parameter λ. Because each blurry image has a different amount of noise and a different blur kernel size, the parameter λ adjusting the smoothness of motion flow and latent image can be changed. We set the parameters as µ = 1, σ I = , c = 0.05, ν = 0.02, 1 σ =, τ = 1, σ L = 10, τ L = , and 2 (2) 2 (2) θ = 0.1 λ. The value of λ ranges from 10 to 50. In the coarse-to-fine framework, we build an image pyramids with 10 levels and a scale factor of 0.8, and use bicubic interpolation for both motion flow and the latent image in propagation. Initially, all dual variables are set as zeros. In addition, the blurry image itself is used for initial latent image L, and we use small random values for initial motion flow u. However, as our blur kernel is symmetric (i.e. k(, u) = k(, u)), we constrain the motion flow, (u, v), to be on a set, B = {(u, v) R 2 u 0} {(u, v) R 2 u = 0, v < 0}. Our current Matlab implementation requires approimately 25 minutes to deblur a VGA image. The running time can be greatly reduced using optimized C or CUDA on a modern GPGPU. 4. Eperimental Results In this section, we show the outstanding performance and efficiency of the proposed work. In Fig. 6, deblurring results and estimated motion flow are illustrated for real challenging dynamic scenes. In qualitative analysis, the edges in deblurred images are restored keenly without segmentation and without restricting the type of motion blur. The color codes of estimated motion flows in tetured regions are as accurate as we epected ecept homogeneous regions that do not raise severe artifacts. Notably, the motion flow in the the bottom row is corresponding to the depth map of the latent image, and this could be another application of our work that is estimating depth map from a blurry single-image. In Fig. 7, deblurring results for camera shakes which includes rotational camera movements are illustrated. Because the proposed method is based on the approimation of locally linear blur kernel, the results may not accurate than the results from methods estimating the global camera motion in less-tetured regions. However, the restored strong edges from both synthetic image in the top row and the real image in the bottom row demonstrate that our approimation is valid for small rotational camera shake to some degree. In addition, we compared our results with the state-ofthe art deblurring methods in various challenging situations. First, in Fig. 8(a), the blurry image is degraded by severe radial distortion and it causes serious problems in conventional methods but ours works well. Second, in Fig. 8, because the cause of blur is specific forward motion which can not be handled in conventional methods, ours outperforms and restores the edges of characters and arrow more sharply. Third, in Fig. 8(c), [26] fails in deblurring as depth discontinuity eist in the blurry image. Although it is possible to deblur with segmentation based method, but [13] also fails, because the background is too narrow to be segmented. Thus, both conventional methods provide unsatisfactory results. Finally, in Fig. 8(d), as the synthetic image, which has depth discontinuity, is blurred by rotational camera shake, and thus, both the conventional methods fail but our method shows successful result. 5. Discussion and Limitations In this study, we have presented an efficient dynamic scene deblurring method that does not require accurate motion segmentation with the aid of robust TV-L1 based model. In addition, the proposed method can handle various types of blur kernels when the kernels can be approimated to linear motions. We also provided an efficient and practical solution optimizing TV-L1 model, and demonstrated that the proposed method outperforms the state-of-the-art methods in various challenging situations. On the other hand, the remaining problem is that, in case of knowing the specific types of motion blurs, conventional deblurring methods estimating the global camera motion work well and the results in less-tetured region are better than ours. Therefore our future work will combine such a global and our local approach in a unified framework.
7 (a) (c) (d) Figure 6: (a) Real blurry dynamic scenes Deblurring results (c) Comparisons of cropped results (d) Estimated motion flows Acknowledgments This research was supported in part by the MKE(The Ministry of Knowledge Economy), Korea and Microsoft Research, under IT/SW Creative research program supervised by the NIPA(National IT Industry Promotion Agency) (NIPA-2013-H ). References [1] P. Bhat, C. L. Zitnick, M. Cohen, and B. Curless. Gradientshop: A gradient-domain optimization framework for image and video filtering. ACM Transactions on Graphics (TOG), 29(10), [2] A. Chambolle and T. Pock. A first-order primal-dual algorithm for conve problems with applications to imaging. Journal of Mathematical Imaging and Vision, 40(1): , May , 5 [3] S. Cho and S. Lee. Fast motion deblurring. In SIGGRAPH, , 3, 4, 5 [4] F. Couzinie-Devy, J. Sun, K. Alahari, and J. Ponce. Learning to estimate and remove non-uniform image blur. In CVPR, [5] S. Dai and Y. Wu. Motion from blur. In CVPR, [6] R. Fergus, B. Singh, A. Hertzmann, S. T. Roweis, and W. Freeman. Removing camera shake from a single photograph. In SIGGRAPH, , 4 [7] G. Gilboa, N. Sochen, and Y. Y. Zeevi. Image enhancement and denoising by comple diffusion processes. IEEE Trans. Pattern Analysis Machine Intelligence, 26(8): , [8] A. Gupta, N. Joshi, L. Zitnick, M. Cohen, and B. Curless. Single image deblurring using motion density functions. In ECCV,
8 (a) (c) (d) Figure 8: From left to right: Blurry images, results of Xu et al [26], results of Kim et al [13], and results of the proposed method (a) (c) Figure 7: (a) Blurry images from rotational camera shakes Deblurring results (c) Comparisons of cropped results [9] S. Harmeling, H. Michael, and B. Schoelkopf. Space-variant singleimage blind deconvolution for removing camera shake. In NIPS, [10] M. Hirsch, C. J. Schuler, S. Harmeling, and B. Scholkopf. Fast removal of non-uniform camera shake. In ICCV, , 2 [11] H. Ji and K. Wang. A two-stage approach to blind spatially-varying motion deblurring. In CVPR, [12] N. Joshi, R. Szeliski, and D. Kriegman. Psf estimation using sharp edge prediction. In CVPR, [13] T. H. Kim, B. Ahn, and K. M. Lee. Dynamic scene deblurring. In ICCV, , 3, 6, 8 [14] D. Krishnan and R. Fergus. Fast image deconvolution using hyperlaplacian priors. In NIPS, [15] D. Krishnan, T. Tay, and R. Fergus. Blind deconvolution using a normalized sparsity measure. In CVPR, [16] A. Levin. Blind motion deblurring using image statistics. In NIPS, [17] A. Levin, D. Lischinski, and Y. Weiss. A closed form solution to natural image matting. IEEE Trans. Pattern Analysis Machine Intelligence, 30(2): , [18] A. Levin and Y. Weiss. User assisted separation of reflections from a single image using a sparsity prior. IEEE Trans. Pattern Analysis Machine Intelligence, 29(9): , [19] A. Levin, Y. Weiss, F. Durand, and W. T. Freeman. Understanding and evaluating blind deconvolution algorithms. In CVPR, [20] S. Osher and L. I. Rudin. Feature-oriented image enhancement using shock filters. SIAM Journal on Numerical Analysis, 27(4): , , 5 [21] Q. Shan, J. Jia, and A. Agarwala. High-quality motion deblurring from a single image. In SIGGRAPH, [22] F. Steinbruecker, T. Pock, and D. Cremers. Large displacement optical flow computation without warping. In ICCV, , 5 [23] O. Whyte, J. Sivic, A. Zisserman, and J. Ponce. Non-uniform deblurring for shaken images. International Journal of Computer Vision, 98(2): , , 2 [24] L. Xu and J. Jia. Two-phase kernel estimation for robust motion deblurring. In ECCV, , 3 [25] L. Xu, J. Jia, and Y. Matsushita. Motion detail preserving optical flow estimation. IEEE Trans. Pattern Analysis Machine Intelligence, 34(9): , [26] L. Xu, S. Zheng, and J. Jia. Unnatural l0 sparse representation for natural image deblurring. In CVPR, , 6, 8 [27] S. Zheng, L. Xu,, and J. Jia. Forward motion deblurring. In ICCV,
RSA algorithm for blurred on blinded deconvolution technique
RSA algorithm for blurred on technique *Kirti Bhadauria 1, *U. Dutta 2, **Priyank gupta 3 *Computer Science and Engineering Department, Maharana Pratap College of technology Putli Ghar Road, Near Collectorate,
High Quality Image Deblurring Panchromatic Pixels
High Quality Image Deblurring Panchromatic Pixels ACM Transaction on Graphics vol. 31, No. 5, 2012 Sen Wang, Tingbo Hou, John Border, Hong Qin, and Rodney Miller Presented by Bong-Seok Choi School of Electrical
Numerical Methods For Image Restoration
Numerical Methods For Image Restoration CIRAM Alessandro Lanza University of Bologna, Italy Faculty of Engineering CIRAM Outline 1. Image Restoration as an inverse problem 2. Image degradation models:
Moving Least Squares Approximation
Chapter 7 Moving Least Squares Approimation An alternative to radial basis function interpolation and approimation is the so-called moving least squares method. As we will see below, in this method the
Optical Flow. Shenlong Wang CSC2541 Course Presentation Feb 2, 2016
Optical Flow Shenlong Wang CSC2541 Course Presentation Feb 2, 2016 Outline Introduction Variation Models Feature Matching Methods End-to-end Learning based Methods Discussion Optical Flow Goal: Pixel motion
Bildverarbeitung und Mustererkennung Image Processing and Pattern Recognition
Bildverarbeitung und Mustererkennung Image Processing and Pattern Recognition 1. Image Pre-Processing - Pixel Brightness Transformation - Geometric Transformation - Image Denoising 1 1. Image Pre-Processing
Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections
Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections Maximilian Hung, Bohyun B. Kim, Xiling Zhang August 17, 2013 Abstract While current systems already provide
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
Understanding and evaluating blind deconvolution algorithms
Understanding and evaluating blind deconvolution algorithms Anat Levin,, Yair Weiss,3, Fredo Durand, William T. Freeman,4 MIT CSAIL, Weizmann Institute of Science, 3 Hebrew University, 4 Adobe Abstract
Computational Optical Imaging - Optique Numerique. -- Deconvolution --
Computational Optical Imaging - Optique Numerique -- Deconvolution -- Winter 2014 Ivo Ihrke Deconvolution Ivo Ihrke Outline Deconvolution Theory example 1D deconvolution Fourier method Algebraic method
Fast Matting Using Large Kernel Matting Laplacian Matrices
Fast Matting Using Large Kernel Matting Laplacian Matrices Kaiming He 1 1 Department of Information Engineering The Chinese University of Hong Kong Jian Sun 2 2 Microsoft Research Asia Xiaoou Tang 1,3
jorge s. marques image processing
image processing images images: what are they? what is shown in this image? What is this? what is an image images describe the evolution of physical variables (intensity, color, reflectance, condutivity)
High Quality Image Magnification using Cross-Scale Self-Similarity
High Quality Image Magnification using Cross-Scale Self-Similarity André Gooßen 1, Arne Ehlers 1, Thomas Pralow 2, Rolf-Rainer Grigat 1 1 Vision Systems, Hamburg University of Technology, D-21079 Hamburg
Canny Edge Detection
Canny Edge Detection 09gr820 March 23, 2009 1 Introduction The purpose of edge detection in general is to significantly reduce the amount of data in an image, while preserving the structural properties
Context-Constrained Hallucination for Image Super-Resolution
Context-Constrained Hallucination for Image Super-Resolution Jian Sun Xi an Jiaotong University Xi an, P. R. China [email protected] Jiejie Zhu Marshall F. Tappen EECS, University of Central Florida
A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow
, pp.233-237 http://dx.doi.org/10.14257/astl.2014.51.53 A Study on SURF Algorithm and Real-Time Tracking Objects Using Optical Flow Giwoo Kim 1, Hye-Youn Lim 1 and Dae-Seong Kang 1, 1 Department of electronices
Mean-Shift Tracking with Random Sampling
1 Mean-Shift Tracking with Random Sampling Alex Po Leung, Shaogang Gong Department of Computer Science Queen Mary, University of London, London, E1 4NS Abstract In this work, boosting the efficiency of
EECS 556 Image Processing W 09. Interpolation. Interpolation techniques B splines
EECS 556 Image Processing W 09 Interpolation Interpolation techniques B splines What is image processing? Image processing is the application of 2D signal processing methods to images Image representation
Applications to Data Smoothing and Image Processing I
Applications to Data Smoothing and Image Processing I MA 348 Kurt Bryan Signals and Images Let t denote time and consider a signal a(t) on some time interval, say t. We ll assume that the signal a(t) is
DYNAMIC RANGE IMPROVEMENT THROUGH MULTIPLE EXPOSURES. Mark A. Robertson, Sean Borman, and Robert L. Stevenson
c 1999 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or
Convolution. 1D Formula: 2D Formula: Example on the web: http://www.jhu.edu/~signals/convolve/
Basic Filters (7) Convolution/correlation/Linear filtering Gaussian filters Smoothing and noise reduction First derivatives of Gaussian Second derivative of Gaussian: Laplacian Oriented Gaussian filters
Diagonal preconditioning for first order primal-dual algorithms in convex optimization
Diagonal preconditioning for first order primal-dual algorithms in conve optimization Thomas Pock Institute for Computer Graphics and Vision Graz University of Technology [email protected] Antonin Chambolle
ROBUST VEHICLE TRACKING IN VIDEO IMAGES BEING TAKEN FROM A HELICOPTER
ROBUST VEHICLE TRACKING IN VIDEO IMAGES BEING TAKEN FROM A HELICOPTER Fatemeh Karimi Nejadasl, Ben G.H. Gorte, and Serge P. Hoogendoorn Institute of Earth Observation and Space System, Delft University
Feature Tracking and Optical Flow
02/09/12 Feature Tracking and Optical Flow Computer Vision CS 543 / ECE 549 University of Illinois Derek Hoiem Many slides adapted from Lana Lazebnik, Silvio Saverse, who in turn adapted slides from Steve
Highlight Removal by Illumination-Constrained Inpainting
Highlight Removal by Illumination-Constrained Inpainting Ping Tan Stephen Lin Long Quan Heung-Yeung Shum Microsoft Research, Asia Hong Kong University of Science and Technology Abstract We present a single-image
Bayesian Image Super-Resolution
Bayesian Image Super-Resolution Michael E. Tipping and Christopher M. Bishop Microsoft Research, Cambridge, U.K..................................................................... Published as: Bayesian
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 269 Class Project Report
Automatic 3D Reconstruction via Object Detection and 3D Transformable Model Matching CS 69 Class Project Report Junhua Mao and Lunbo Xu University of California, Los Angeles [email protected] and lunbo
Single Depth Image Super Resolution and Denoising Using Coupled Dictionary Learning with Local Constraints and Shock Filtering
Single Depth Image Super Resolution and Denoising Using Coupled Dictionary Learning with Local Constraints and Shock Filtering Jun Xie 1, Cheng-Chuan Chou 2, Rogerio Feris 3, Ming-Ting Sun 1 1 University
A Duality Based Algorithm for TV-L 1 -Optical-Flow Image Registration
A Duality Based Algorithm for TV-L 1 -Optical-Flow Image Registration Thomas Pock 1, Martin Urschler 1, Christopher Zach 2, Reinhard Beichel 3, and Horst Bischof 1 1 Institute for Computer Graphics & Vision,
Admin stuff. 4 Image Pyramids. Spatial Domain. Projects. Fourier domain 2/26/2008. Fourier as a change of basis
Admin stuff 4 Image Pyramids Change of office hours on Wed 4 th April Mon 3 st March 9.3.3pm (right after class) Change of time/date t of last class Currently Mon 5 th May What about Thursday 8 th May?
Convolutional Feature Maps
Convolutional Feature Maps Elements of efficient (and accurate) CNN-based object detection Kaiming He Microsoft Research Asia (MSRA) ICCV 2015 Tutorial on Tools for Efficient Object Detection Overview
Support Vector Machine. Tutorial. (and Statistical Learning Theory)
Support Vector Machine (and Statistical Learning Theory) Tutorial Jason Weston NEC Labs America 4 Independence Way, Princeton, USA. [email protected] 1 Support Vector Machines: history SVMs introduced
CMOS Image Sensor Noise Reduction Method for Image Signal Processor in Digital Cameras and Camera Phones
CMOS Image Sensor Noise Reduction Method for Image Signal Processor in Digital Cameras and Camera Phones Youngjin Yoo, SeongDeok Lee, Wonhee Choe and Chang-Yong Kim Display and Image Processing Laboratory,
Fast Image Labeling for Creating High-Resolution Panoramic Images on Mobile Devices
2009 11th IEEE International Symposium on Multimedia Fast Image Labeling for Creating High-Resolution Panoramic Images on Mobile Devices Yingen Xiong and Kari Pulli Nokia Research Center Palo Alto, CA
Affine Transformations
A P P E N D I X C Affine Transformations CONTENTS C The need for geometric transformations 335 C2 Affine transformations 336 C3 Matri representation of the linear transformations 338 C4 Homogeneous coordinates
Epipolar Geometry. Readings: See Sections 10.1 and 15.6 of Forsyth and Ponce. Right Image. Left Image. e(p ) Epipolar Lines. e(q ) q R.
Epipolar Geometry We consider two perspective images of a scene as taken from a stereo pair of cameras (or equivalently, assume the scene is rigid and imaged with a single camera from two different locations).
High Performance GPU-based Preprocessing for Time-of-Flight Imaging in Medical Applications
High Performance GPU-based Preprocessing for Time-of-Flight Imaging in Medical Applications Jakob Wasza 1, Sebastian Bauer 1, Joachim Hornegger 1,2 1 Pattern Recognition Lab, Friedrich-Alexander University
MAP-Inference for Highly-Connected Graphs with DC-Programming
MAP-Inference for Highly-Connected Graphs with DC-Programming Jörg Kappes and Christoph Schnörr Image and Pattern Analysis Group, Heidelberg Collaboratory for Image Processing, University of Heidelberg,
Color Segmentation Based Depth Image Filtering
Color Segmentation Based Depth Image Filtering Michael Schmeing and Xiaoyi Jiang Department of Computer Science, University of Münster Einsteinstraße 62, 48149 Münster, Germany, {m.schmeing xjiang}@uni-muenster.de
Synthetic Aperture Radar (SAR)
asic Concept of SAR: Consider fied array Synthetic Aperture Radar (SAR) L Angular resolution θ λ /L v(t i ) Σ Claim Equivalent to Moving Antenna: θ Assumes phase coherence in oscillator during each image
Colour Image Segmentation Technique for Screen Printing
60 R.U. Hewage and D.U.J. Sonnadara Department of Physics, University of Colombo, Sri Lanka ABSTRACT Screen-printing is an industry with a large number of applications ranging from printing mobile phone
A Correlation-Based Approach to Robust Point Set Registration
A Correlation-Based Approach to Robust Point Set Registration Yanghai Tsin 1 and Takeo Kanade 2 1 Siemens Corporate Research, 755 College Road East, Princeton, NJ 854, USA 2 Carnegie Mellon University,
Probabilistic Models for Big Data. Alex Davies and Roger Frigola University of Cambridge 13th February 2014
Probabilistic Models for Big Data Alex Davies and Roger Frigola University of Cambridge 13th February 2014 The State of Big Data Why probabilistic models for Big Data? 1. If you don t have to worry about
Classifying Manipulation Primitives from Visual Data
Classifying Manipulation Primitives from Visual Data Sandy Huang and Dylan Hadfield-Menell Abstract One approach to learning from demonstrations in robotics is to make use of a classifier to predict if
A Reliability Point and Kalman Filter-based Vehicle Tracking Technique
A Reliability Point and Kalman Filter-based Vehicle Tracing Technique Soo Siang Teoh and Thomas Bräunl Abstract This paper introduces a technique for tracing the movement of vehicles in consecutive video
WAVELET FRAME BASED IMAGE RESTORATION WITH MISSING/DAMAGED PIXELS
WAVELET FRAME BASED IMAGE RESTORATION WITH MISSING/DAMAGED PIXELS HUI JI, ZUOWEI SHEN, AND YUHONG XU Abstract. This paper addresses the problem of how to recover degraded images with partial image pixels
How To Analyze Ball Blur On A Ball Image
Single Image 3D Reconstruction of Ball Motion and Spin From Motion Blur An Experiment in Motion from Blur Giacomo Boracchi, Vincenzo Caglioti, Alessandro Giusti Objective From a single image, reconstruct:
Simultaneous Gamma Correction and Registration in the Frequency Domain
Simultaneous Gamma Correction and Registration in the Frequency Domain Alexander Wong [email protected] William Bishop [email protected] Department of Electrical and Computer Engineering University
Efficient online learning of a non-negative sparse autoencoder
and Machine Learning. Bruges (Belgium), 28-30 April 2010, d-side publi., ISBN 2-93030-10-2. Efficient online learning of a non-negative sparse autoencoder Andre Lemme, R. Felix Reinhart and Jochen J. Steil
A Noise-Aware Filter for Real-Time Depth Upsampling
A Noise-Aware Filter for Real-Time Depth Upsampling Derek Chan Hylke Buisman Christian Theobalt Sebastian Thrun Stanford University, USA Abstract. A new generation of active 3D range sensors, such as time-of-flight
The Image Deblurring Problem
page 1 Chapter 1 The Image Deblurring Problem You cannot depend on your eyes when your imagination is out of focus. Mark Twain When we use a camera, we want the recorded image to be a faithful representation
Low-resolution Character Recognition by Video-based Super-resolution
2009 10th International Conference on Document Analysis and Recognition Low-resolution Character Recognition by Video-based Super-resolution Ataru Ohkura 1, Daisuke Deguchi 1, Tomokazu Takahashi 2, Ichiro
Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability
Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller
Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression
Logistic Regression Department of Statistics The Pennsylvania State University Email: [email protected] Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max
How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm
IJSTE - International Journal of Science Technology & Engineering Volume 1 Issue 10 April 2015 ISSN (online): 2349-784X Image Estimation Algorithm for Out of Focus and Blur Images to Retrieve the Barcode
Visualization by Linear Projections as Information Retrieval
Visualization by Linear Projections as Information Retrieval Jaakko Peltonen Helsinki University of Technology, Department of Information and Computer Science, P. O. Box 5400, FI-0015 TKK, Finland [email protected]
Single Image Super-Resolution using Gaussian Process Regression
Single Image Super-Resolution using Gaussian Process Regression He He and Wan-Chi Siu Department of Electronic and Information Engineering The Hong Kong Polytechnic University {07821020d, [email protected]}
Image Normalization for Illumination Compensation in Facial Images
Image Normalization for Illumination Compensation in Facial Images by Martin D. Levine, Maulin R. Gandhi, Jisnu Bhattacharyya Department of Electrical & Computer Engineering & Center for Intelligent Machines
Vision based Vehicle Tracking using a high angle camera
Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu [email protected] [email protected] Abstract A vehicle tracking and grouping algorithm is presented in this work
Face Model Fitting on Low Resolution Images
Face Model Fitting on Low Resolution Images Xiaoming Liu Peter H. Tu Frederick W. Wheeler Visualization and Computer Vision Lab General Electric Global Research Center Niskayuna, NY, 1239, USA {liux,tu,wheeler}@research.ge.com
Lecture 14. Point Spread Function (PSF)
Lecture 14 Point Spread Function (PSF), Modulation Transfer Function (MTF), Signal-to-noise Ratio (SNR), Contrast-to-noise Ratio (CNR), and Receiver Operating Curves (ROC) Point Spread Function (PSF) Recollect
Improvements in quality and quantification of 3D PET images
Improvements in quality and quantification of 3D PET images The resolution of PET depends on some effects impossible to be readily corrected The implementation of 3D spatially-variant PSF (radially asymmetric)
Tracking Moving Objects In Video Sequences Yiwei Wang, Robert E. Van Dyck, and John F. Doherty Department of Electrical Engineering The Pennsylvania State University University Park, PA16802 Abstract{Object
SUPER-RESOLUTION (SR) has been an active research
498 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 15, NO. 3, APRIL 2013 A Self-Learning Approach to Single Image Super-Resolution Min-Chun Yang and Yu-Chiang Frank Wang, Member, IEEE Abstract Learning-based approaches
Detection and Restoration of Vertical Non-linear Scratches in Digitized Film Sequences
Detection and Restoration of Vertical Non-linear Scratches in Digitized Film Sequences Byoung-moon You 1, Kyung-tack Jung 2, Sang-kook Kim 2, and Doo-sung Hwang 3 1 L&Y Vision Technologies, Inc., Daejeon,
Segmentation of building models from dense 3D point-clouds
Segmentation of building models from dense 3D point-clouds Joachim Bauer, Konrad Karner, Konrad Schindler, Andreas Klaus, Christopher Zach VRVis Research Center for Virtual Reality and Visualization, Institute
Graph Based Processing of Big Images
Graph Based Processing of Big Images HUI HAN CHIN DSO NATIONAL LABORATORIES, SINGAPORE 31 OCT 2014 Introduction Member of Technical Staff at Signal Processing Laboratory, Sensors Division, DSO National
Face Recognition in Low-resolution Images by Using Local Zernike Moments
Proceedings of the International Conference on Machine Vision and Machine Learning Prague, Czech Republic, August14-15, 014 Paper No. 15 Face Recognition in Low-resolution Images by Using Local Zernie
Accurate Blur Models vs. Image Priors in Single Image Super-Resolution
Accurate Blur Models vs. Image Priors in Single Image Super-Resolution Netalee Efrat, Daniel Glasner, Alexander Apartsin, Boaz Nadler, Anat Levin Dept. of Computer Science and Applied Math The Weizmann
The Visual Internet of Things System Based on Depth Camera
The Visual Internet of Things System Based on Depth Camera Xucong Zhang 1, Xiaoyun Wang and Yingmin Jia Abstract The Visual Internet of Things is an important part of information technology. It is proposed
Latest Results on High-Resolution Reconstruction from Video Sequences
Latest Results on High-Resolution Reconstruction from Video Sequences S. Lertrattanapanich and N. K. Bose The Spatial and Temporal Signal Processing Center Department of Electrical Engineering The Pennsylvania
NEIGHBORHOOD REGRESSION FOR EDGE-PRESERVING IMAGE SUPER-RESOLUTION. Yanghao Li, Jiaying Liu, Wenhan Yang, Zongming Guo
NEIGHBORHOOD REGRESSION FOR EDGE-PRESERVING IMAGE SUPER-RESOLUTION Yanghao Li, Jiaying Liu, Wenhan Yang, Zongming Guo Institute of Computer Science and Technology, Peking University, Beijing, P.R.China,
Final Year Project Progress Report. Frequency-Domain Adaptive Filtering. Myles Friel. Supervisor: Dr.Edward Jones
Final Year Project Progress Report Frequency-Domain Adaptive Filtering Myles Friel 01510401 Supervisor: Dr.Edward Jones Abstract The Final Year Project is an important part of the final year of the Electronic
Sharpening through spatial filtering
Sharpening through spatial filtering Stefano Ferrari Università degli Studi di Milano [email protected] Elaborazione delle immagini (Image processing I) academic year 2011 2012 Sharpening The term
A Study on the Comparison of Electricity Forecasting Models: Korea and China
Communications for Statistical Applications and Methods 2015, Vol. 22, No. 6, 675 683 DOI: http://dx.doi.org/10.5351/csam.2015.22.6.675 Print ISSN 2287-7843 / Online ISSN 2383-4757 A Study on the Comparison
Factorizing Scene Albedo and Depth from a Single Foggy Image
Factorizing Scene Albedo and Depth from a Single Foggy Image Louis Kratz Ko Nishino Drexel University Philadelphia, PA {lak24, kon}@drexel.edu Abstract Atmospheric conditions induced by suspended particles,
Couple Dictionary Training for Image Super-resolution
IEEE TRANSACTIONS ON IMAGE PROCESSING 1 Couple Dictionary Training for Image Super-resolution Jianchao Yang, Student Member, IEEE, Zhaowen Wang, Student Member, IEEE, Zhe Lin, Member, IEEE, Scott Cohen,
Image Segmentation and Registration
Image Segmentation and Registration Dr. Christine Tanner ([email protected]) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation
On Convergence Rate of Concave-Convex Procedure
On Convergence Rate of Concave-Conve Procedure Ian En-Hsu Yen [email protected] Po-Wei Wang [email protected] Nanyun Peng Johns Hopkins University Baltimore, MD 21218 [email protected] Shou-De
Distributed Structured Prediction for Big Data
Distributed Structured Prediction for Big Data A. G. Schwing ETH Zurich [email protected] T. Hazan TTI Chicago M. Pollefeys ETH Zurich R. Urtasun TTI Chicago Abstract The biggest limitations of learning
A Novel Hole filling method based on Projection onto Convex Set in DIBR
3rd International Conference on Multimedia Technology ICMT 2013) A Novel Hole filling method based on Projection onto Convex Set in DIBR Weiquan Wang1 and Yingyun Yang2 and Qian Liang3 Abstract. Depth
On SDP- and CP-relaxations and on connections between SDP-, CP and SIP
On SDP- and CP-relaations and on connections between SDP-, CP and SIP Georg Still and Faizan Ahmed University of Twente p 1/12 1. IP and SDP-, CP-relaations Integer program: IP) : min T Q s.t. a T j =
Real-time Visual Tracker by Stream Processing
Real-time Visual Tracker by Stream Processing Simultaneous and Fast 3D Tracking of Multiple Faces in Video Sequences by Using a Particle Filter Oscar Mateo Lozano & Kuzahiro Otsuka presented by Piotr Rudol
Part-Based Recognition
Part-Based Recognition Benedict Brown CS597D, Fall 2003 Princeton University CS 597D, Part-Based Recognition p. 1/32 Introduction Many objects are made up of parts It s presumably easier to identify simple
Digital Imaging and Multimedia. Filters. Ahmed Elgammal Dept. of Computer Science Rutgers University
Digital Imaging and Multimedia Filters Ahmed Elgammal Dept. of Computer Science Rutgers University Outlines What are Filters Linear Filters Convolution operation Properties of Linear Filters Application
Parabolic Equations. Chapter 5. Contents. 5.1.2 Well-Posed Initial-Boundary Value Problem. 5.1.3 Time Irreversibility of the Heat Equation
7 5.1 Definitions Properties Chapter 5 Parabolic Equations Note that we require the solution u(, t bounded in R n for all t. In particular we assume that the boundedness of the smooth function u at infinity
3D Scanner using Line Laser. 1. Introduction. 2. Theory
. Introduction 3D Scanner using Line Laser Di Lu Electrical, Computer, and Systems Engineering Rensselaer Polytechnic Institute The goal of 3D reconstruction is to recover the 3D properties of a geometric
Two-Frame Motion Estimation Based on Polynomial Expansion
Two-Frame Motion Estimation Based on Polynomial Expansion Gunnar Farnebäck Computer Vision Laboratory, Linköping University, SE-581 83 Linköping, Sweden [email protected] http://www.isy.liu.se/cvl/ Abstract.
Subspace Analysis and Optimization for AAM Based Face Alignment
Subspace Analysis and Optimization for AAM Based Face Alignment Ming Zhao Chun Chen College of Computer Science Zhejiang University Hangzhou, 310027, P.R.China [email protected] Stan Z. Li Microsoft
A System for Capturing High Resolution Images
A System for Capturing High Resolution Images G.Voyatzis, G.Angelopoulos, A.Bors and I.Pitas Department of Informatics University of Thessaloniki BOX 451, 54006 Thessaloniki GREECE e-mail: [email protected]
Probabilistic Latent Semantic Analysis (plsa)
Probabilistic Latent Semantic Analysis (plsa) SS 2008 Bayesian Networks Multimedia Computing, Universität Augsburg [email protected] www.multimedia-computing.{de,org} References
Support Vector Machines with Clustering for Training with Very Large Datasets
Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France [email protected] Massimiliano
Lecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
Super-resolution Reconstruction Algorithm Based on Patch Similarity and Back-projection Modification
1862 JOURNAL OF SOFTWARE, VOL 9, NO 7, JULY 214 Super-resolution Reconstruction Algorithm Based on Patch Similarity and Back-projection Modification Wei-long Chen Digital Media College, Sichuan Normal
Edge tracking for motion segmentation and depth ordering
Edge tracking for motion segmentation and depth ordering P. Smith, T. Drummond and R. Cipolla Department of Engineering University of Cambridge Cambridge CB2 1PZ,UK {pas1001 twd20 cipolla}@eng.cam.ac.uk
Image Super-Resolution as Sparse Representation of Raw Image Patches
Image Super-Resolution as Sparse Representation of Raw Image Patches Jianchao Yang, John Wright, Yi Ma, Thomas Huang University of Illinois at Urbana-Champagin Beckman Institute and Coordinated Science
