Understanding and evaluating blind deconvolution algorithms
|
|
|
- Joleen Bradford
- 9 years ago
- Views:
Transcription
1 Understanding and evaluating blind deconvolution algorithms Anat Levin,, Yair Weiss,3, Fredo Durand, William T. Freeman,4 MIT CSAIL, Weizmann Institute of Science, 3 Hebrew University, 4 Adobe Abstract Blind deconvolution is the recovery of a sharp version of a image when the blur kernel is unknown. Recent algorithms have afforded dramatic progress, yet many aspects of the problem remain challenging and hard to understand. The goal of this paper is to analyze and evaluate recent blind deconvolution algorithms both theoretically and experimentally. We explain the previously reported failure of the naive MAP approach by demonstrating that it mostly favors no-blur explanations. On the other hand we show that since the kernel size is often smaller than the image size a MAP estimation of the kernel alone can be well constrained and accurately recover the true blur. The plethora of recent deconvolution techniques makes an experimental evaluation on ground-truth data important. We have collected blur data with ground truth and compared recent algorithms under equal settings. Additionally, our data demonstrates that the shift-invariant blur assumption made by most algorithms is often violated.. Introduction Blind deconvolution is the problem of recovering a sharp version of an input blurry image when the blur kernel is unknown []. Mathematically, we wish to decompose a image y as y = k x () where x is a visually plausible sharp image, and k is a non negative blur kernel, whose support is small compared to the image size. This problem is severely ill-posed and there is an infinite set of pairs (x, k) explaining any observed y. For example, One undesirable solution that perfectly satisfies eq. is the no-blur explanation: k is the delta (identity) kernel and x = y. The ill-posed nature of the problem implies that additional assumptions on x or k must be introduced. Blind deconvolution is the subject of numerous papers in the signal and image processing literature, to name a few consider [,,, 5, 7] and the survey in []. Despite the exhaustive research, results on real world images are rarely produced. Recent algorithms have proposed to address the ill-posedness of blind deconvolution by characterizing x using natural image statistics [, 4, 4,, 7, 3, ]. While this principle has lead to tremendous progress, the results are still far from perfect. Blind deconvolution algorithms exhibit some common building principles, and vary in others. The goal of this paper is to analyze the problem and shed new light on recent algorithms. What are the key challenges and what are the important components that make blind deconvolution possible? Additionally, which aspects of the problem should attract further research efforts? One of the puzzling aspects of blind deconvolution is the failure of the MAP approach. Recent papers emphasize the usage of a sparse derivative prior to favor sharp images. However, a direct application of this principle has not yielded the expected results and all algorithms have required additional components, such as marginalization across all possible images [, 4, 4], spatially-varying terms [7, 9], or solvers that vary their optimization energy over time [9]. In this paper we analyze the source of the MAP failure. We show that counter-intuitively, the most favorable solution under a sparse prior is usually a blurry image and not a sharp one. Thus, the global optimum of the MAP approach is the no-blur explanation. We discuss solutions to the problem and analyze the answers provided by existing algorithms. We show that one key property making blind deconvolution possible is the strong asymmetry between the dimensionalities of x and k. While the number of unknowns in x increases with image size, the dimensionality of k remains small. Therefore, while a simultaneous MAP estimation of both x and k fails, a MAP estimation of k alone (marginalizing over x), is well constrained and recovers an accurate kernel. We suggest that while the sparse prior is helpful, the key component making blind deconvolution possible is not the choice of prior, but the thoughtful choice of estimator. Furthermore, we show that with a proper estimation rule, blind deconvolution can be performed even with a weak Gaussian prior. Finally, we collect motion- data with ground truth. This data allows us to quantitatively compare recent blind deconvolution algorithms. Our evaluation suggest that the variational Bayes approach of [4] outperforms all existing alternatives. This data also shows that the shift invariance convolution model involved in most existing algorithms is often violated and that realistic camera shake includes in-plane rotations.. MAP x,k estimation and its limitations In this paper y denotes an observed blurry image, which is a convolution of an unknown sharp image x with an unknown blur kernel k, plus noise n (this paper assumes i.i.d. Gaussian noise): y = k x + n. ()
2 Using capital letters for the Fourier transform of a signal: Y ω = K ωx ω + N ω. (3) The goal of blind deconvolution is to infer both k and x given a single input y. Additionally, k is non negative, and its support is often small compared to the image size. The simplest approach is a maximum-a-posteriori (MAP x,k ) estimation, seeking a pair (ˆx, ˆk) maximizing: p(x, k y) p(y x, k)p(x)p(k). (4) For simplicity of the exposition, we assume a uniform prior on k. The likelihood term p(y x, k) is the data fitting term log p(y x, k) = λ k x y. The prior p(x) favors natural images, usually based on the observation that their gradient distribution is sparse. A common measure is log p(x) = X g x,i(x) α + g y,i(x) α + C (5) i where g x,i (x) and g y,i (x) denote the horizontal and vertical derivatives at pixel i (we use the simple [ ] filter) and C is a constant normalization term. Exponent values α < lead to sparse priors and natural images usually correspond to α in the range of [.5,.] []. Other choices include a Laplacian prior α =, and a Gaussian prior α =. While natural image gradients are very non-gaussian, we examine this model because it enables an analytical treatment. The MAP x,k approach seeks (ˆx, ˆk) minimizing (ˆx, ˆk) = arg min x,k λ k x y + X i Eq. () reveals an immediate limitation: g x,i(x) α + g y,i(x) α. Claim Let x be an arbitrarily large image sampled from the prior p(x), and y = k x. The pair (x, k) optimizing the MAP x,k score satisfies x and k. Proof: For every pair (x, k) we use a scalar s to define a new pair x = s x, k = /s k with equal data fitting k x y = k x y. While the data fitting term is constant, the prior term improves as s. This observation is not surprising. The most likely image under the prior in Eq. (5) is a flat image with no gradients. One attempt to fix the problem is to assume the mean intensity of the and sharp images should be equal, and constrain the sum of k: i k i =. This eliminates the zero solution, but usually the no-blur solution is still favored. To understand this, consider the D signals x in Fig. that were convolved with a (truncated) Gaussian kernel k of standard deviation 4 pixels. We compare two interpretations: ) the true kernel: y = k x. ) the delta kernel (no blur) y = k y. We evaluate the logp(x, k y) score (Eq. ()), while varying the α parameter in the prior. For step edges (Fig. (a)) MAP x,k usually succeeds. The edge is sharper than its version and while the Gaussian prior favors the blurry explanation, appropriate sparse priors (α < ) favor the correct sharp explanation. We keep estimation variables in subscript to distinguish between a MAP estimation of both x and k, to a MAP estimation of k alone. () (a) (b) Σ i g i (x) α Σ i g i (x) α Σ i g i (x) α 4 α 4 α α (c) Figure. The MAP x,k score evaluated on toy D signals. Left: sharp and signals. Right: sum of gradients log p(x) = P i gi(x) α as a function of α. 5 5 windows 5 5 windows windows 3% % % Figure. MAP x,k failure on real image windows. Windows in which the sharp explanation is favored are marked in red. The percent of windows in which the sharp version is favored decreases with window size. In contrast, Fig. (b) presents a narrow peak. Blurring reduces the peak height, and as a result, the Laplacian prior α = favors the blurry x (k is delta) because the absolute sum of gradients is lower. Examining Fig. (b-right) suggests that the explanation is winning for smaller α values as well. The sharp explanation is favored only for low alpha values, approaching a binary penalty. However, the sparse models describing natural images are not binary, they are usually in the range α [.5,.] []. The last signal considered in Fig. (c) is a row cropped from a natural image, illustrating that natural images contain a lot of medium contrast texture and noise, corresponding to the narrow peak structure. This dominates the statistics more than step edges. As a result, blurring a natural image reduces the overall contrast and, as in Fig. (b), even sparse priors favor the blurry x explanation.
3 p(x) x 4 x l= l=3 l=9 E[ log p(x)] l (blur width) α=. α=.5 α= 9 (a) p l for α =.5 (b) E p l[ log p (x l )] Figure 3. (a) Comparison of gradient histograms for and un images sampled from p (x). Blur reduces the average gradient magnitude. (b) Expected negative likelihood reduces (probability increases) with blur. To confirm the above observation, we the image in Fig. with a Gaussian kernel of standard deviation 3 pixels. We compared the sum of the gradients in the and sharp images using α =.5. For 5 5 windows the image is favored over 97% of the windows, and this phenomenon increases with window size. For windows, the version is favored at all windows. Another observation is that if the sharp explanation does win, it happens next to significant edges. To understand this, note that blur has two opposite effects on the image likelihood: ) it makes the signal derivatives less sparse, and that reduces the likelihood. ) It reduces the derivatives variance and that increases its likelihood. For very specific images, like ideal step edges, the first effect dominants and blur reduces the likelihood. However, for most natural images the second effect is stronger and blur increases the likelihood. To illustrate this, let x be a sequence sampled i.i.d. from p (x i ) e γ x i α, x l a sequence obtained by convolving x with a width l box filter (normalizing the kernel sum to ), and p l its probability distribution. The expected negative log likelihood (effecting the MAP x,k ) of x l under the sharp distribution p is: E p l[ log p (x l )] = p l (x)log p (x)dx. Fig. 3(a) plots p l for α =.5, and Fig. 3(b) the expected likelihood as a function of l. The variance is reduced by convolution, and hence the negative log-likelihood reduces as well. Revisiting the literature on the subject, Fergus et al. [4] report that their initial attempts to approach blind deconvolution with MAP x,k failed, resulting in either the explanation or a binary two-tone image, depending on parameter tunings. Algorithms like [7, ] explicitly detect edges in the image (either manually or automatically), and seek a kernel which transfers these edges into binary ones. This is motivated by the example in Fig., suggesting that MAP x,k could do the right thing around step edges. Another algorithm which makes usage of this property is [9]. It optimizes a semi-map x,k score, but explicitly detects smooth image regions and reweights their contribution. Thus, the MAP x,k score is dominated by edges. We discuss this algorithm in detail in [3]. Earlier blind deconvolution papers which exploit a MAP x,k approach avoid the delta solution using other assumptions which are less applicable for real world images. For example, [] assumes x contains an object on a flat background with a known compact support. All these examples highlight the fact that the prior alone does not favor the desired result. The source of the problem is that for all α values, the most likely event of the prior in Eq. (5) is the fully flat image. This phenomenon is robust to the exact choice of prior, and replacing the model in Eq. (5) with higher order derivatives or with more sophisticated natural image priors [, 3] does not change the result. We also note that the problem is present even if the derivatives signal is sampled exactly from p(x) and the prior is perfectly correct in the generative sense. In the next section we suggest that, to overcome the MAP x,k limitation, one should reconsider the choice of estimator. We revisit a second group of blind deconvolution algorithms derived from this idea. 3. MAP k estimation The limitations of MAP estimation in the case of few measurements have been pointed out many times in estimation theory and statistical signal processing [9, ]. Indeed, in the MAP x,k problem we can never collect enough measurements because the number of unknowns grows with the image size. In contrast, estimation theory tells us [9] that, given enough measurements, MAP estimators do approach the true solution. Therefore, the key to success is to exploit a special property of blind deconvolution: the strong asymmetry between the dimensionalities of the two unknowns. While the dimensionality of x increases with the image size, the support of the kernel is fixed and small relative to the image size. The image y does provide a large number of measurements for estimating k. As we prove below, for an increasing image size, a MAP k estimation of k alone (marginalizing over x) can recover the true kernel with an increasing accuracy. This result stands in contrast to Claim which stated that a MAP x,k estimator continues to fail even as the number of measurements goes to infinity. This leads to an alternative blind deconvolution strategy: use a MAP k estimator to recover the kernel and, given the kernel, solve for x using a non blind deconvolution algorithm. Before providing a formal proof, we attempt to gain an intuition about the difference between MAP k and MAP x,k scores. A MAP k estimator selects ˆk = arg max k p(k y), where p(k y) = p(y k)p(k)/p(y), and p(y k) is obtained by marginalizing over x, and evaluating the full volume of possible x interpretations: Z p(y k) = p(x, y k)dx. (7) To see the role of marginalization, consider the scalar blind deconvolution problem illustrated in []. Suppose a scalar y is observed, and should be decomposed as y = k x + n. Assume a zero mean Gaussian prior on the noise and signal,
4 x max p(x,k) p(k) k k k (a) p(x, k y) (b) p(k y ) (c) p(k y,..y N ) Figure 4. A toy blind deconvolution problem with one scalar y = kx+n (replotted from []). (a) The joint distribution p(x,k y). A maximum is obtained for x, k. (b) The marginalized score p(k y) produce an optimum closer to the true k. (c) The uncertainty of p(k y) reduces given multiple observations y j = kx j + n j. x N(, σ ), n N(, η ). Then p(k) N= N=5 N=5 N= P(x,k y) e η kx y x σ. () Fig. 4(a) illustrate the D distribution P(x, k y). Unsurprisingly, it is maximized by x, k. On the other hand, p(y k) is the integral over all x explanations: Z P(y k) e η kx y x σ dx. (9) This integral is not maximized by k. In fact, if we consider the first term only e η kx y dx, it clearly favors k values because they allow a larger volume of possible x values. To see that, note that for every k and every ǫ > the size of the set of x values satisfying kx y < ǫ is ǫ/k, maximized as k. Combining the two terms in (9) leads to an example in the middle of the range, and we show in Sec. 3.. that x σ, which make sense because x now behaves like a typical sample from the prior. This is the principle of genericity described in Bayesian terms by []. Fig. 4(b) plots P(y k), which is essentially summing the columns of Fig. 4(a). Now consider blur in real images: for the delta kernel there is only a single solution x = y satisfying k x = y. However, while the delta spectrum is high everywhere, the true kernel is usually a low pass, and has low spectrum values. Referring to the notation of Eq. (3), if K ω =, an infinite subspace of possible explanations is available as X ω can be arbitrary (and with noise, any low K ω values increase the uncertainty, even if they are not exactly ). Hence, the true kernel gets an advantage in the p(y k) score. We prove that for sufficiently large images, p(k y) is guaranteed to favor the true kernel. Claim Let x be an arbitrarily large image, sampled from the prior p(x), and y = k x + n. Then p(k y) is maximized by the true kernel k. Moreover, if argmax k p(y k) is unique, p(k y) approaches a delta function. Note that Claim does not guarantee that the MAP k is unique. For example, if the kernel support is not constrained enough, multiple spatial shifts of the kernel provide equally good solutions. The problem can be easily avoided by a weak prior on k (e.g. favoring centered kernels). Proof: We divide the image into small disjoint windows {y,..., y n } and treat them as i.i.d. samples y j p(y k ). We then select k ML = argmax k j p(yj k). Applying the standard consistency theorem for maximum likelihood estimators [9] we know that given enough samples, the ML approaches the true parameters. That is, when n p(k ML ({y,..., y n }) = k ). () Due to the local form of the prior p(x) (Eq. (5)), taking sufficiently far away disjoint windows will ensure that p(y k) j p(yj k). Thus, p(y k) is maximized by k ML. Also, if we select a m times larger image y, p(y k) = p(y k) m. Thus, if p(y k) < max k p(y k) then p(y k). Finally, if p(k ) >, then k MAP, k ML are equal on large images since arg max k p(y k) = arg max k p(y k)p(k), and thus, k MAP k. Similarly, if max k p(y k) is unique, p(k y) approaches a delta function. Fig. 4(c) plots p(y k) for a scalar blind deconvolution task with N observations y j = kx j +n j, illustrating that as N increases, the uncertainty around the solution decreases (compare with Fig. 4(b)). In [3] we also justify the MAP k approach from the loss function perspective. 3.. Examples of MAP k estimation Claim reduces to a robust blind deconvolution strategy: use MAP k estimator to recover k MAP = arg max k p(k y), and then use k MAP to solve for x using some non blind deconvolution algorithm. To illustrate the MAP k approach, we start with the simple case of a Gaussian prior on p(x), as it permits a derivation in closed form. 3.. The Gaussian prior The prior on X in Eq. (5) is a convolution and thus diagonal in the frequency domain. If G x, G y denote the Fourier transform of the derivatives g x, g y, then: X N(, diag(σ ω)) σ ω = β( G x,ω + G y,ω ). () Note that since a derivative filter is zero at low frequencies and high at higher frequencies, this is similar to the classical /f power spectrum law for images. Denoting noise variance by η, we can express p(x, Y ; K) = p(y X; K)p(X) as: p(x,y ; K) e η KωXω Yω σ ω X ω. () (see [3] for details). Conditioned on k, the mean and mode of a Gaussian are equal: X MAP ω = K ω + η σ ω «K T ω Y ω. (3) Eq. (3) is the classic Wiener filter [5]. One can also integrate X and express p(y K) analytically. This is also a diagonal zero mean Gaussian with Y N(, diag(φ ω)), φ ω = σ ω K ω + η. (4)
5 Eq. (4) is maximized when φ ω = Y ω, and for blind deconvolution, this implies: «ˆK ω = max, Yω η. (5) The image estimated using ˆK satisfies X ω σω. Therefore MAP k does not result in a trivial X = solution as MAP x,k would, but in a solution whose variance matches the prior variance σ, that is, a solution which looks like a typical sample from the prior p(x). Another way to interpret the MAP k, is to note that σ ω X Kω log + η log p(y K) = log p(x MAP, Y ; K) σ ω ω () Referring to Eq. (), the second term is just the log determinant of the covariance of p(x Y ; K). This second term is optimized when K ω =, i.e. by kernels with more blur. That is, log p(y K) is equal to the MAP x,k score of the mode plus a term favoring kernels with blur. The discussion above suggests that the Gaussian MAP k provides a reasonable solution to blind deconvolution. In the experiment section we evaluate this algorithm and show that, while weaker than the sparse prior, it can provide acceptable solutions. This stands in contrast to the complete failure of a MAP x,k approach, even with the seemingly better sparse prior. This demonstrates that a careful choice of estimator is actually more critical than the choice of prior. Note that Eq. (5) is accurate if every frequency is estimated independently. In practice, the solution can be further constrained, because the limited spatial support of k implies that the frequency coefficients {K ω } are linearly dependent. Another important issue is that Eq. (5) provides information on the kernel power spectrum alone but leaves uncertainty about the phase. Many variants of Gaussian blind deconvolution algorithms are available in the image processing literature (e.g. [, 5]) but in most cases only symmetric kernels are considered since their phase is known to be zero. However, realistic camera shake kernels are usually not symmetric. In [3] we describe a Gaussian blind deconvoltion algorithm which attempts to recover non symmetric kernels as well. 3.. Approximation strategies with a sparse prior The challenge with the MAP k approach is that for a general sparse prior, p(k y) (Eq. (7)) cannot be computed in closed form. Several previous blind deconvolution algorithms can be viewed as approximation strategies for MAP k, although the authors might not have motivated them in this way. A simple approximation is proposed by Levin [4], for the D blur case. It assumes that the observed derivatives of y are independent (this is usually weaker than assuming independent derivatives of x): log p(y k) = i log p(g x,i(y) k). Since p(g x,i (y) k) is a D distributions, it can be expressed as a D table, or a histogram h k. «+C The independence assumption implies that instead of summing over image pixels, one can express p(y k) by summing over histogram bins: log p(y k) = X i log p(g x,i(y) k) = X j h j log(h k j) (7) where h denotes the gradients histogram in the observed image and j is a bin index. In a second step, note that maximizing Eq. (7) is equivalent to minimizing the histogram distance between the observed and expected histograms h,h k. This is because the Kullback Leibler divergence is equal to the negative log likelihood, plus a constant that does not depend on k (the negative entropy): D KL(h, h k ) = X j h j log(h j) X j h j log(h k j ). () Since the KL divergence is non-negative, the likelihood is maximized when the histograms h, h k are equal. This very simple approach is already able to avoid the delta solution but as we demonstrate in Sec. 4. it is not accurately identifying the exact filter width. A stronger approximation is the variational Bayes meanfield approach taken by Fergus et al. [4]. The idea is to build an approximating distribution with a simpler parametric form: p(x,k y) q(x, k) = Y i q(g i,x(x))q(g i,y(x)) Y j q(k j). (9) Since q is expressed in the gradient domain this does not recover x directly. Thus, they also pick the MAP k kernel from q and then solve for x using non blind deconvolution. A third way to approximate the MAP k is the Laplace approximation [], which is a generalization of Eq. (): log p(y k) log p(x MAP, y;k) log A + C () A = log p(x,y; k) x x=x MAP. () i x j The Laplace approximation states that p(y k) can be expressed by the probability of the mode x MAP plus the log determinant of the variance around the mode. As discussed above, higher variance is usually achieved when k contains more zero frequencies, i.e. more blur. Therefore, the Laplace approximation suggests that p(y k) is the MAP x,k score plus a term pulling toward kernels with more blur. Unfortunately, in the non Gaussian case the covariance matrix isn t diagonal and exact inversion is less trivial. Some earlier blind deconvolution approaches [, 7] can be viewed as simplified forms of a blur favoring term. For example, they bias towered blurry kernels by adding a term penalizing the high frequencies of k or with an explicit prior on the kernel. Another approach was exploit by Bronstein et al. [3]. They note that in the absence of noise and with invertible kernels p(k y) can be exactly evaluated for sparse priors as well. This reduces to optimizing the sparsity of the image plus the log determinant of the kernel spectrum.
6 Exact Zero sheet MAP x,k MAP x,k +edge reweight Gaussian prior independent approx variational approx Bronstein et. al. Figure 5. log p(y k) scores using various approximation strategies on D image signals. Successful algorithms locate the minimum score at the true, denoted by the dashed line. likelihood N= N=5 N= Figure. The uncertainty in kernel estimation decreses with more samples. For as little at N = columns it is already tightly peaked at the true answer. 4. Evaluating blind deconvolution algorithms In this section we qualitatively compare blind deconvolution strategies on the same data. We start with a synthetic D example and in the second part turn to real D motion. 4.. D evaluation As a first test, we use a set of signals of size cropped from a natural image. These small D signals allow us to evaluate the marginalization integral in Eq. (7) exactly even for a sparse prior. The signals were convolved with a 5-tap box filter (cyclic convolution was used) and an i.i.d. Gaussian noise with standard deviation. was added. We explicitly search over the explanations of all box filters of size l =,.., 7 taps (all filters normalized to ). The explicit search allows comparison of the score of different blind deconvolution strategies without folding in optimization errors. (In practice optimization errors do have a large effect on the successes of blind deconvolution algorithms.) The exact log p(y k) score is minimized by the true box width l = 5. We tested the zero sheet separation (e.g. []), an earlier image processing approach with no probabilistic formulation. This algorithm measures the Fourier magnitude of y at the zero frequencies of each box filter k. If the image was indeed convolved with that filter, low Fourier content is expected. However, this approach considers the zero frequencies alone ignoring all other information, and is known to be noise sensitive. It is also limited to kernel families from a simple parametric form and with a clear zeros structure. Supporting the example in Sec., a pure MAP x,k approach (p(y k) p(x MAP, y k)) favors no-blur (l = ). Reweighting the derivative penalty around edges can improve the situation, but the delta solution still provides a noticeable local optimum. The correct minimum is favored with a variational Bayes approximation [4] and with the semi Laplace approximation of [3]. The independence approximation [4] is able to overcome the delta solution, but does not localize the solution very accurately (minimum at l = 4 instead of l = 5.) Finally, the correct solution is identified even with the poor image prior provided by a Gaussian model, demonstrating (a) Figure 7. Ground truth data acquisition. (a) Calibration image. (b) Smear of points at 4 corners, demonstrating that the spatially uniform blur model is violated. that the choice of estimator (MAP x,k v.s. MAP k ), is more critical than the actual prior (Gaussian v.s. sparse). Since claim guaranties success only for large images, we attempt to evaluate how large an image should be in practice. Fig. plots the uncertainty in p(k y) for multiple random samples of N columns. The probability is tightly peaked at the right answer for as little as N = columns. The search space in Fig. is limited to the single parameter family of box filters. In real motion deblurring one searches over a larger family of kernels and a larger uncertainty is expected. 4.. D evaluation To compare blind deconvolution algorithms we have collected data with ground truth. We capture a sharp version a planar scene (Fig. 7(a)) by mounting the camera on a tripod, as well as a few shots. Using the sharp reference we solve for a non-negative kernel k minimizing k x y. The scene in Fig. 7(a) includes high frequency noise patterns which helps stabilizing the constraints on k. The central area of the frame includes four real images used as input to the various blind deconvolution algorithms. We first observed that assuming a uniform blur over the image is not realistic even for planar scenes. For example Fig. 7(b) shows traces of points at 4 corners of an image captured by a hand-held camera, with a clear variation between the corners. This suggests that an in-plane rotation (rotation around the z-axis) is a significant component of human hand shake. Yet, since a uniform assumption is made by most algorithms, we need to evaluate them on data (b)
7 Figure. Ground truth data: 4 images and blur kernels, resulting in 3 test images Percentage 4 Fergus Shan Shan, sps deconv MAP xk Gaussian prior above 4 Error ratios Figure 9. Evaluation results: Cumulative histogram of the deconvolution error ratio across test examples. which obeys their assumption. To capture images with spatially invariant blur we placed the camera on a tripod, locking the Z-axis rotation handle of the tripod but loosening the X and Y handles. We calibrated the blur of such images and cropped windows from each, leading to 3 test images displayed in Fig. and available online levina/papers/levinetalcvpr9data.zip We used an 5mm lens and a.3 seconds exposure. The kernels support varied from to 5 pixels. We can measure the SSD error between a deconvolved output and the ground truth. However, wider kernels result in larger deconvolution error even with the true kernel. To normalize this effect, we measure the ratio between deconvolution error with the estimated kernel and deconvolution with the truth kernel. In Fig. 9 we plot the cumulative histogram of error ratios (e.g. bin r = 3 counts the percentage of test examples achieving error ratio below 3). Empirically, we noticed that error ratios above are already visually implausible. One test image is presented in Fig., all others included in [3]. We have evaluated the algorithms of Fergus et al. [4] and Shan et al. [9] (each using the authors implementation), as well as MAP k estimation using a Gaussian prior [3], and a simplified MAP x,k approach constraining k i = (we used coordinate descent, iterating between holding x constant and solving for k, and then holding k constant and solving for x ). The algorithms of [4, 7, 3] were not tested because the first was designed for D motion only, and the others focus on smaller blur kernels. We made our best attempt to adjust the parameters of Shan et al. [9], but run all test images with equal parameters. Fergus et al. [4] used Richardson-Lucy non blind deconvolution in their code. Since this algorithm is a source for ringing artifacts, we improved the results using the kernel estimated by the authors code with the (non blind) sparse deconvolution of []. Similarly, we used sparse deconvolution with the kernel estimated by Shan et al. The bars in Fig. 9 and the visual results in [3] suggest that Fergus et al. s algorithm [4] significantly outperforms all other alternatives. Many of the artifacts in the results of [4] can be attributed to the Richardson-Lucy artifacts, or to non uniform blur in their test images. Our comparison also suggests that applying sparse deconvolution using the kernels outputted by Shan et al. [9] improves their results. As expected, the naive MAP x,k approach outputs small kernels approaching the delta solution. 5. Discussion This paper analyzes the major building blocks of recent blind deconvolution algorithms. We illustrate the limitation of the simple MAP x,k approach, favoring the no-blur (delta kernel) explanation. One class of solutions involves explicit edge detection. A more principled strategy exploits the dimensionality asymmetry, and estimates MAP k while marginalizing over x. While the computational aspects involved with this marginalization are more challenging, existing approximations are powerful. We have collected motion blur data with ground truth and quantitatively compared existing algorithms. Our comparison suggests that the variational Bayes approximation [4] significantly outperforms all existing alternatives. The conclusions from our analysis are useful for directing future blind deconvolution research. In particular, we
8 Input Fergus et. al. error ratio=.7 Naive MAP x,k error ratio=5. Ground truth Shan et. al. error ratio=5. Gaussian prior error ratio=. Figure. Visual deconvolution results by various deconvolution algorithms. See [3] for more examples. note that modern natural image priors [, 3] do not overcome the MAP x,k limitation (and in our tests did not change the observation in Sec. ). While it is possible that blind deconvolution can benefit from future research on natural image statistics, this paper suggests that better estimators for existing priors may have more impact on future blind deconvolution algorithms. Additionally, we observed that the popular spatially uniform blur assumption is usually unrealistic. Thus, it seems that blur models which can relax this assumption [] have a high potential to improve blind deconvolution results. Acknowledgments: We thank the Israel Science Foundation, the Royal Dutch/Shell Group, NGA NEGI-5-4-4, MURI Grant N , NSF CAREER award Fredo Durand acknowledges a Microsoft Research New Faculty Fellowship and a Sloan Fellowship. References [] G. R. Ayers and J. C. Dainty. Interative blind deconvolution method and its applications. Opt. Lett., 9. [] D. Brainard and W. Freeman. Bayesian color constancy. JOSA, 997. [3] M. M. Bronstein, A. M. Bronstein, M. Zibulevsky, and Y. Y. Zeevi. Blind deconvolution of images using optimal sparse representations. Image Processing, IEEE Transactions on, 4():7 73, 5. [4] R. Fergus, B. Singh, A. Hertzmann, S.T. Roweis, and W.T. Freeman. Removing camera shake from a single photograph. SIGGRAPH,. [5] R. C. Gonzalez and R. E. Woods. Digital Image Processing. Prentice Hall, January. [] Jiaya Jia. Single image motion deblurring using transparency. In CVPR, 7. [7] N. Joshi, R. Szeliski, and D. Kriegman. Psf estimation using sharp edge prediction. In CVPR,. [] A. K. Katsaggelos and K. T. Lay. Maximum likelihood blur identification and image restoration using the em algorithm. IEEE Trans. Signal Processing, 99. [9] S. M. Kay. Fundamentals of Statistical Signal Processing: Estimation Theory. Prentice Hall, 997. [] D. Kundur and D. Hatzinakos. Blind image deconvolution. IEEE Signal Processing Magazine, 99. [] R. G. Lane and R. H. T. Bates. Automatic multidimensional deconvolution. J. Opt. Soc. Am. A, 4():, 97. [] A. Levin, R. Fergus, F. Durand, and W. Freeman. Image and depth from a conventional camera with a coded aperture. SIGGRAPH, 7. [3] A. Levin, Y. Weiss, F. Durand, and W.T. Freeman. Understanding and evaluating blind deconvolution algorithms. Technical report, MIT-CSAIL-TR-9-4, 9. [4] Anat Levin. Blind motion deblurring using image statistics. In Advances in Neural Information Processing Systems (NIPS),. [5] A. C. Likas and N. P. Galatsanos. A variational approach for bayesian blind image deconvolution. IEEE Trans. on Signal Processing, 4. [] J. W. Miskin and D. J. C. MacKay. Ensemble learning for blind image separation and deconvolution. In Advances in Independent Component Analysis. Springer,. [7] R. Molina, A. K. Katsaggelos, J. Abad, and J. Mateos. A bayesian approach to blind deconvolution based on dirichlet distributions. In ICASSP, 997. [] S. Roth and M.J. Black. Fields of experts: A framework for learning image priors. In CVPR, 5. [9] Q. Shan, J. Jia, and A. Agarwala. High-quality motion deblurring from a single image. SIGGRAPH,. [] Q. Shan, W. Xiong, and J. Jia. Rotational motion deblurring of a rigid object from a single image. In ICCV, 7. [] E P Simoncelli. Bayesian denoising of visual images in the wavelet domain. In Bayesian Inference in Wavelet Based Models. Springer-Verlag, New York, 999. [] E. Thiébaut and J.-M. Conan. Strict a priori constraints for maximum-likelihood blind deconvolution. J. Opt. Soc. Am. A, (3):45 49, 995. [3] Y. Weiss and W. T. Freeman. What makes a good model of natural images? In CVPR, 7.
RSA algorithm for blurred on blinded deconvolution technique
RSA algorithm for blurred on technique *Kirti Bhadauria 1, *U. Dutta 2, **Priyank gupta 3 *Computer Science and Engineering Department, Maharana Pratap College of technology Putli Ghar Road, Near Collectorate,
Numerical Methods For Image Restoration
Numerical Methods For Image Restoration CIRAM Alessandro Lanza University of Bologna, Italy Faculty of Engineering CIRAM Outline 1. Image Restoration as an inverse problem 2. Image degradation models:
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
High Quality Image Deblurring Panchromatic Pixels
High Quality Image Deblurring Panchromatic Pixels ACM Transaction on Graphics vol. 31, No. 5, 2012 Sen Wang, Tingbo Hou, John Border, Hong Qin, and Rodney Miller Presented by Bong-Seok Choi School of Electrical
Basics of Statistical Machine Learning
CS761 Spring 2013 Advanced Machine Learning Basics of Statistical Machine Learning Lecturer: Xiaojin Zhu [email protected] Modern machine learning is rooted in statistics. You will find many familiar
Bayesian Image Super-Resolution
Bayesian Image Super-Resolution Michael E. Tipping and Christopher M. Bishop Microsoft Research, Cambridge, U.K..................................................................... Published as: Bayesian
Computational Optical Imaging - Optique Numerique. -- Deconvolution --
Computational Optical Imaging - Optique Numerique -- Deconvolution -- Winter 2014 Ivo Ihrke Deconvolution Ivo Ihrke Outline Deconvolution Theory example 1D deconvolution Fourier method Algebraic method
CHAPTER 2 Estimating Probabilities
CHAPTER 2 Estimating Probabilities Machine Learning Copyright c 2016. Tom M. Mitchell. All rights reserved. *DRAFT OF January 24, 2016* *PLEASE DO NOT DISTRIBUTE WITHOUT AUTHOR S PERMISSION* This is a
Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections
Blind Deconvolution of Barcodes via Dictionary Analysis and Wiener Filter of Barcode Subsections Maximilian Hung, Bohyun B. Kim, Xiling Zhang August 17, 2013 Abstract While current systems already provide
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
Practical Guide to the Simplex Method of Linear Programming
Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear
Introduction to General and Generalized Linear Models
Introduction to General and Generalized Linear Models General Linear Models - part I Henrik Madsen Poul Thyregod Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby
99.37, 99.38, 99.38, 99.39, 99.39, 99.39, 99.39, 99.40, 99.41, 99.42 cm
Error Analysis and the Gaussian Distribution In experimental science theory lives or dies based on the results of experimental evidence and thus the analysis of this evidence is a critical part of the
A Coefficient of Variation for Skewed and Heavy-Tailed Insurance Losses. Michael R. Powers[ 1 ] Temple University and Tsinghua University
A Coefficient of Variation for Skewed and Heavy-Tailed Insurance Losses Michael R. Powers[ ] Temple University and Tsinghua University Thomas Y. Powers Yale University [June 2009] Abstract We propose a
Admin stuff. 4 Image Pyramids. Spatial Domain. Projects. Fourier domain 2/26/2008. Fourier as a change of basis
Admin stuff 4 Image Pyramids Change of office hours on Wed 4 th April Mon 3 st March 9.3.3pm (right after class) Change of time/date t of last class Currently Mon 5 th May What about Thursday 8 th May?
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written
The Image Deblurring Problem
page 1 Chapter 1 The Image Deblurring Problem You cannot depend on your eyes when your imagination is out of focus. Mark Twain When we use a camera, we want the recorded image to be a faithful representation
Bayesian probability theory
Bayesian probability theory Bruno A. Olshausen arch 1, 2004 Abstract Bayesian probability theory provides a mathematical framework for peforming inference, or reasoning, using probability. The foundations
Epipolar Geometry. Readings: See Sections 10.1 and 15.6 of Forsyth and Ponce. Right Image. Left Image. e(p ) Epipolar Lines. e(q ) q R.
Epipolar Geometry We consider two perspective images of a scene as taken from a stereo pair of cameras (or equivalently, assume the scene is rigid and imaged with a single camera from two different locations).
DYNAMIC RANGE IMPROVEMENT THROUGH MULTIPLE EXPOSURES. Mark A. Robertson, Sean Borman, and Robert L. Stevenson
c 1999 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or
Lecture 14. Point Spread Function (PSF)
Lecture 14 Point Spread Function (PSF), Modulation Transfer Function (MTF), Signal-to-noise Ratio (SNR), Contrast-to-noise Ratio (CNR), and Receiver Operating Curves (ROC) Point Spread Function (PSF) Recollect
Lecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall
Automatic Photo Quality Assessment Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Estimating i the photorealism of images: Distinguishing i i paintings from photographs h Florin
An Introduction to Machine Learning
An Introduction to Machine Learning L5: Novelty Detection and Regression Alexander J. Smola Statistical Machine Learning Program Canberra, ACT 0200 Australia [email protected] Tata Institute, Pune,
Vector and Matrix Norms
Chapter 1 Vector and Matrix Norms 11 Vector Spaces Let F be a field (such as the real numbers, R, or complex numbers, C) with elements called scalars A Vector Space, V, over the field F is a non-empty
Statistical Machine Translation: IBM Models 1 and 2
Statistical Machine Translation: IBM Models 1 and 2 Michael Collins 1 Introduction The next few lectures of the course will be focused on machine translation, and in particular on statistical machine translation
High Quality Image Magnification using Cross-Scale Self-Similarity
High Quality Image Magnification using Cross-Scale Self-Similarity André Gooßen 1, Arne Ehlers 1, Thomas Pralow 2, Rolf-Rainer Grigat 1 1 Vision Systems, Hamburg University of Technology, D-21079 Hamburg
Language Modeling. Chapter 1. 1.1 Introduction
Chapter 1 Language Modeling (Course notes for NLP by Michael Collins, Columbia University) 1.1 Introduction In this chapter we will consider the the problem of constructing a language model from a set
Measuring Line Edge Roughness: Fluctuations in Uncertainty
Tutor6.doc: Version 5/6/08 T h e L i t h o g r a p h y E x p e r t (August 008) Measuring Line Edge Roughness: Fluctuations in Uncertainty Line edge roughness () is the deviation of a feature edge (as
Low-resolution Character Recognition by Video-based Super-resolution
2009 10th International Conference on Document Analysis and Recognition Low-resolution Character Recognition by Video-based Super-resolution Ataru Ohkura 1, Daisuke Deguchi 1, Tomokazu Takahashi 2, Ichiro
Investment Statistics: Definitions & Formulas
Investment Statistics: Definitions & Formulas The following are brief descriptions and formulas for the various statistics and calculations available within the ease Analytics system. Unless stated otherwise,
Adaptive Online Gradient Descent
Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650
PS 271B: Quantitative Methods II. Lecture Notes
PS 271B: Quantitative Methods II Lecture Notes Langche Zeng [email protected] The Empirical Research Process; Fundamental Methodological Issues 2 Theory; Data; Models/model selection; Estimation; Inference.
Monotonicity Hints. Abstract
Monotonicity Hints Joseph Sill Computation and Neural Systems program California Institute of Technology email: [email protected] Yaser S. Abu-Mostafa EE and CS Deptartments California Institute of Technology
The CUSUM algorithm a small review. Pierre Granjon
The CUSUM algorithm a small review Pierre Granjon June, 1 Contents 1 The CUSUM algorithm 1.1 Algorithm............................... 1.1.1 The problem......................... 1.1. The different steps......................
Inner Product Spaces
Math 571 Inner Product Spaces 1. Preliminaries An inner product space is a vector space V along with a function, called an inner product which associates each pair of vectors u, v with a scalar u, v, and
Tutorial on Markov Chain Monte Carlo
Tutorial on Markov Chain Monte Carlo Kenneth M. Hanson Los Alamos National Laboratory Presented at the 29 th International Workshop on Bayesian Inference and Maximum Entropy Methods in Science and Technology,
19 LINEAR QUADRATIC REGULATOR
19 LINEAR QUADRATIC REGULATOR 19.1 Introduction The simple form of loopshaping in scalar systems does not extend directly to multivariable (MIMO) plants, which are characterized by transfer matrices instead
Applications to Data Smoothing and Image Processing I
Applications to Data Smoothing and Image Processing I MA 348 Kurt Bryan Signals and Images Let t denote time and consider a signal a(t) on some time interval, say t. We ll assume that the signal a(t) is
EECS 556 Image Processing W 09. Interpolation. Interpolation techniques B splines
EECS 556 Image Processing W 09 Interpolation Interpolation techniques B splines What is image processing? Image processing is the application of 2D signal processing methods to images Image representation
Date: April 12, 2001. Contents
2 Lagrange Multipliers Date: April 12, 2001 Contents 2.1. Introduction to Lagrange Multipliers......... p. 2 2.2. Enhanced Fritz John Optimality Conditions...... p. 12 2.3. Informative Lagrange Multipliers...........
1 Solving LPs: The Simplex Algorithm of George Dantzig
Solving LPs: The Simplex Algorithm of George Dantzig. Simplex Pivoting: Dictionary Format We illustrate a general solution procedure, called the simplex algorithm, by implementing it on a very simple example.
1 Prior Probability and Posterior Probability
Math 541: Statistical Theory II Bayesian Approach to Parameter Estimation Lecturer: Songfeng Zheng 1 Prior Probability and Posterior Probability Consider now a problem of statistical inference in which
By choosing to view this document, you agree to all provisions of the copyright laws protecting it.
This material is posted here with permission of the IEEE Such permission of the IEEE does not in any way imply IEEE endorsement of any of Helsinki University of Technology's products or services Internal
Robert Collins CSE598G. More on Mean-shift. R.Collins, CSE, PSU CSE598G Spring 2006
More on Mean-shift R.Collins, CSE, PSU Spring 2006 Recall: Kernel Density Estimation Given a set of data samples x i ; i=1...n Convolve with a kernel function H to generate a smooth function f(x) Equivalent
Linear Programming. March 14, 2014
Linear Programming March 1, 01 Parts of this introduction to linear programming were adapted from Chapter 9 of Introduction to Algorithms, Second Edition, by Cormen, Leiserson, Rivest and Stein [1]. 1
Visualization by Linear Projections as Information Retrieval
Visualization by Linear Projections as Information Retrieval Jaakko Peltonen Helsinki University of Technology, Department of Information and Computer Science, P. O. Box 5400, FI-0015 TKK, Finland [email protected]
MBA 611 STATISTICS AND QUANTITATIVE METHODS
MBA 611 STATISTICS AND QUANTITATIVE METHODS Part I. Review of Basic Statistics (Chapters 1-11) A. Introduction (Chapter 1) Uncertainty: Decisions are often based on incomplete information from uncertain
Analysis of Bayesian Dynamic Linear Models
Analysis of Bayesian Dynamic Linear Models Emily M. Casleton December 17, 2010 1 Introduction The main purpose of this project is to explore the Bayesian analysis of Dynamic Linear Models (DLMs). The main
Jitter Measurements in Serial Data Signals
Jitter Measurements in Serial Data Signals Michael Schnecker, Product Manager LeCroy Corporation Introduction The increasing speed of serial data transmission systems places greater importance on measuring
How To Analyze Ball Blur On A Ball Image
Single Image 3D Reconstruction of Ball Motion and Spin From Motion Blur An Experiment in Motion from Blur Giacomo Boracchi, Vincenzo Caglioti, Alessandro Giusti Objective From a single image, reconstruct:
Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus
Auxiliary Variables in Mixture Modeling: 3-Step Approaches Using Mplus Tihomir Asparouhov and Bengt Muthén Mplus Web Notes: No. 15 Version 8, August 5, 2014 1 Abstract This paper discusses alternatives
Accurate Blur Models vs. Image Priors in Single Image Super-Resolution
Accurate Blur Models vs. Image Priors in Single Image Super-Resolution Netalee Efrat, Daniel Glasner, Alexander Apartsin, Boaz Nadler, Anat Levin Dept. of Computer Science and Applied Math The Weizmann
Convolution. 1D Formula: 2D Formula: Example on the web: http://www.jhu.edu/~signals/convolve/
Basic Filters (7) Convolution/correlation/Linear filtering Gaussian filters Smoothing and noise reduction First derivatives of Gaussian Second derivative of Gaussian: Laplacian Oriented Gaussian filters
TCOM 370 NOTES 99-4 BANDWIDTH, FREQUENCY RESPONSE, AND CAPACITY OF COMMUNICATION LINKS
TCOM 370 NOTES 99-4 BANDWIDTH, FREQUENCY RESPONSE, AND CAPACITY OF COMMUNICATION LINKS 1. Bandwidth: The bandwidth of a communication link, or in general any system, was loosely defined as the width of
Logistic Regression. Jia Li. Department of Statistics The Pennsylvania State University. Logistic Regression
Logistic Regression Department of Statistics The Pennsylvania State University Email: [email protected] Logistic Regression Preserve linear classification boundaries. By the Bayes rule: Ĝ(x) = arg max
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
Correlation and Convolution Class Notes for CMSC 426, Fall 2005 David Jacobs
Correlation and Convolution Class otes for CMSC 46, Fall 5 David Jacobs Introduction Correlation and Convolution are basic operations that we will perform to extract information from images. They are in
Canny Edge Detection
Canny Edge Detection 09gr820 March 23, 2009 1 Introduction The purpose of edge detection in general is to significantly reduce the amount of data in an image, while preserving the structural properties
Predict the Popularity of YouTube Videos Using Early View Data
000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050
Linear Classification. Volker Tresp Summer 2015
Linear Classification Volker Tresp Summer 2015 1 Classification Classification is the central task of pattern recognition Sensors supply information about an object: to which class do the object belong
MATH10212 Linear Algebra. Systems of Linear Equations. Definition. An n-dimensional vector is a row or a column of n numbers (or letters): a 1.
MATH10212 Linear Algebra Textbook: D. Poole, Linear Algebra: A Modern Introduction. Thompson, 2006. ISBN 0-534-40596-7. Systems of Linear Equations Definition. An n-dimensional vector is a row or a column
CCNY. BME I5100: Biomedical Signal Processing. Linear Discrimination. Lucas C. Parra Biomedical Engineering Department City College of New York
BME I5100: Biomedical Signal Processing Linear Discrimination Lucas C. Parra Biomedical Engineering Department CCNY 1 Schedule Week 1: Introduction Linear, stationary, normal - the stuff biology is not
JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions
Edith Cowan University Research Online ECU Publications Pre. JPEG compression of monochrome D-barcode images using DCT coefficient distributions Keng Teong Tan Hong Kong Baptist University Douglas Chai
4. Continuous Random Variables, the Pareto and Normal Distributions
4. Continuous Random Variables, the Pareto and Normal Distributions A continuous random variable X can take any value in a given range (e.g. height, weight, age). The distribution of a continuous random
Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.
Statistical Learning: Chapter 4 Classification 4.1 Introduction Supervised learning with a categorical (Qualitative) response Notation: - Feature vector X, - qualitative response Y, taking values in C
People have thought about, and defined, probability in different ways. important to note the consequences of the definition:
PROBABILITY AND LIKELIHOOD, A BRIEF INTRODUCTION IN SUPPORT OF A COURSE ON MOLECULAR EVOLUTION (BIOL 3046) Probability The subject of PROBABILITY is a branch of mathematics dedicated to building models
Clarify Some Issues on the Sparse Bayesian Learning for Sparse Signal Recovery
Clarify Some Issues on the Sparse Bayesian Learning for Sparse Signal Recovery Zhilin Zhang and Bhaskar D. Rao Technical Report University of California at San Diego September, Abstract Sparse Bayesian
Least-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
1 The Brownian bridge construction
The Brownian bridge construction The Brownian bridge construction is a way to build a Brownian motion path by successively adding finer scale detail. This construction leads to a relatively easy proof
Globally Optimal Crowdsourcing Quality Management
Globally Optimal Crowdsourcing Quality Management Akash Das Sarma Stanford University [email protected] Aditya G. Parameswaran University of Illinois (UIUC) [email protected] Jennifer Widom Stanford
STA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! [email protected]! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
The continuous and discrete Fourier transforms
FYSA21 Mathematical Tools in Science The continuous and discrete Fourier transforms Lennart Lindegren Lund Observatory (Department of Astronomy, Lund University) 1 The continuous Fourier transform 1.1
Elasticity Theory Basics
G22.3033-002: Topics in Computer Graphics: Lecture #7 Geometric Modeling New York University Elasticity Theory Basics Lecture #7: 20 October 2003 Lecturer: Denis Zorin Scribe: Adrian Secord, Yotam Gingold
Probabilistic Models for Big Data. Alex Davies and Roger Frigola University of Cambridge 13th February 2014
Probabilistic Models for Big Data Alex Davies and Roger Frigola University of Cambridge 13th February 2014 The State of Big Data Why probabilistic models for Big Data? 1. If you don t have to worry about
CALCULATIONS & STATISTICS
CALCULATIONS & STATISTICS CALCULATION OF SCORES Conversion of 1-5 scale to 0-100 scores When you look at your report, you will notice that the scores are reported on a 0-100 scale, even though respondents
Component Ordering in Independent Component Analysis Based on Data Power
Component Ordering in Independent Component Analysis Based on Data Power Anne Hendrikse Raymond Veldhuis University of Twente University of Twente Fac. EEMCS, Signals and Systems Group Fac. EEMCS, Signals
Introduction to Support Vector Machines. Colin Campbell, Bristol University
Introduction to Support Vector Machines Colin Campbell, Bristol University 1 Outline of talk. Part 1. An Introduction to SVMs 1.1. SVMs for binary classification. 1.2. Soft margins and multi-class classification.
Rodenstock Photo Optics
Rogonar Rogonar-S Rodagon Apo-Rodagon N Rodagon-WA Apo-Rodagon-D Accessories: Modular-Focus Lenses for Enlarging, CCD Photos and Video To reproduce analog photographs as pictures on paper requires two
MATH 4330/5330, Fourier Analysis Section 11, The Discrete Fourier Transform
MATH 433/533, Fourier Analysis Section 11, The Discrete Fourier Transform Now, instead of considering functions defined on a continuous domain, like the interval [, 1) or the whole real line R, we wish
Part II Redundant Dictionaries and Pursuit Algorithms
Aisenstadt Chair Course CRM September 2009 Part II Redundant Dictionaries and Pursuit Algorithms Stéphane Mallat Centre de Mathématiques Appliquées Ecole Polytechnique Sparsity in Redundant Dictionaries
Building an Advanced Invariant Real-Time Human Tracking System
UDC 004.41 Building an Advanced Invariant Real-Time Human Tracking System Fayez Idris 1, Mazen Abu_Zaher 2, Rashad J. Rasras 3, and Ibrahiem M. M. El Emary 4 1 School of Informatics and Computing, German-Jordanian
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
6.4 Normal Distribution
Contents 6.4 Normal Distribution....................... 381 6.4.1 Characteristics of the Normal Distribution....... 381 6.4.2 The Standardized Normal Distribution......... 385 6.4.3 Meaning of Areas under
A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images
A Complete Gradient Clustering Algorithm for Features Analysis of X-ray Images Małgorzata Charytanowicz, Jerzy Niewczas, Piotr A. Kowalski, Piotr Kulczycki, Szymon Łukasik, and Sławomir Żak Abstract Methods
Semi-Supervised Support Vector Machines and Application to Spam Filtering
Semi-Supervised Support Vector Machines and Application to Spam Filtering Alexander Zien Empirical Inference Department, Bernhard Schölkopf Max Planck Institute for Biological Cybernetics ECML 2006 Discovery
The equivalence of logistic regression and maximum entropy models
The equivalence of logistic regression and maximum entropy models John Mount September 23, 20 Abstract As our colleague so aptly demonstrated ( http://www.win-vector.com/blog/20/09/the-simplerderivation-of-logistic-regression/
Enhancement of scanned documents in Besov spaces using wavelet domain representations
Enhancement of scanned documents in Besov spaces using wavelet domain representations Kathrin Berkner 1 Ricoh Innovations, Inc., 2882 Sand Hill Road, Suite 115, Menlo Park, CA 94025 ABSTRACT After scanning,
Some Essential Statistics The Lure of Statistics
Some Essential Statistics The Lure of Statistics Data Mining Techniques, by M.J.A. Berry and G.S Linoff, 2004 Statistics vs. Data Mining..lie, damn lie, and statistics mining data to support preconceived
Econometrics Simple Linear Regression
Econometrics Simple Linear Regression Burcu Eke UC3M Linear equations with one variable Recall what a linear equation is: y = b 0 + b 1 x is a linear equation with one variable, or equivalently, a straight
The Optimality of Naive Bayes
The Optimality of Naive Bayes Harry Zhang Faculty of Computer Science University of New Brunswick Fredericton, New Brunswick, Canada email: hzhang@unbca E3B 5A3 Abstract Naive Bayes is one of the most
Bargaining Solutions in a Social Network
Bargaining Solutions in a Social Network Tanmoy Chakraborty and Michael Kearns Department of Computer and Information Science University of Pennsylvania Abstract. We study the concept of bargaining solutions,
Descriptive statistics Statistical inference statistical inference, statistical induction and inferential statistics
Descriptive statistics is the discipline of quantitatively describing the main features of a collection of data. Descriptive statistics are distinguished from inferential statistics (or inductive statistics),
Probabilistic Latent Semantic Analysis (plsa)
Probabilistic Latent Semantic Analysis (plsa) SS 2008 Bayesian Networks Multimedia Computing, Universität Augsburg [email protected] www.multimedia-computing.{de,org} References
The Heat Equation. Lectures INF2320 p. 1/88
The Heat Equation Lectures INF232 p. 1/88 Lectures INF232 p. 2/88 The Heat Equation We study the heat equation: u t = u xx for x (,1), t >, (1) u(,t) = u(1,t) = for t >, (2) u(x,) = f(x) for x (,1), (3)
Inference of Probability Distributions for Trust and Security applications
Inference of Probability Distributions for Trust and Security applications Vladimiro Sassone Based on joint work with Mogens Nielsen & Catuscia Palamidessi Outline 2 Outline Motivations 2 Outline Motivations
The Effects of Start Prices on the Performance of the Certainty Equivalent Pricing Policy
BMI Paper The Effects of Start Prices on the Performance of the Certainty Equivalent Pricing Policy Faculty of Sciences VU University Amsterdam De Boelelaan 1081 1081 HV Amsterdam Netherlands Author: R.D.R.
ECON 459 Game Theory. Lecture Notes Auctions. Luca Anderlini Spring 2015
ECON 459 Game Theory Lecture Notes Auctions Luca Anderlini Spring 2015 These notes have been used before. If you can still spot any errors or have any suggestions for improvement, please let me know. 1
