Single-Image Super-resolution Using Sparse Regression and Natural Image Prior
|
|
|
- Prosper Dennis
- 10 years ago
- Views:
Transcription
1 1 Single-Image Super-resolution Using Sparse Regression and Natural Image Prior Kwang In Kim and Younghee Kwon Abstract This paper proposes a framework for single-image super-resolution. The underlying idea is to learn a map from input low-resolution images to target high-resolution images based on example pairs of input and output images. Kernel ridge regression (KRR) is adopted for this purpose. To reduce the time complexity of training and testing for KRR, a sparse solution is found by combining the ideas of kernel matching pursuit and gradient descent. As a regularized solution, KRR leads to a better generalization than simply storing the examples as it has been done in existing example-based algorithms and results in much less noisy images. However, this may introduce blurring and ringing artifacts around major edges as sharp changes are penalized severely. A prior model of a generic image class which takes into account the discontinuity property of images is adopted to resolve this problem. Comparison with existing algorithms shows the effectiveness of the proposed method. Index Terms Computer vision, machine learning, image enhancement, display algorithms 1 INTRODUCTION Single-image super-resolution refers to the task of constructing a high-resolution enlargement of a single low-resolution image. This problem is inherently ill-posed as there are generally multiple high-resolution images that can be reduced to the same low-resolution image. Accordingly, for this problem, one has to rely on strong prior information. This information is available either in the explicit form of a distribution or energy functional defined on the image class [1], [], [3], [4], and/or in the implicit form of example images which leads to examplebased super-resolution [5], [6], [7], [8], [9], [10], [11], [1], [13]. Previous example-based super-resolution algorithms can roughly be characterized as nearest neighbor (NN)-based estimation: during the training phase, pairs of low-resolution and corresponding high-resolution image patches (sub-windows of images) are collected. Then, in the super-resolution phase, each patch of the given low-resolution image is compared to the stored low-resolution patches, and the high-resolution patch corresponding to the nearest low-resolution patch and satisfying a certain spatial neighborhood compatibility is selected as the output. For instance, Freeman et al. [6] posed the image super-resolution as the problem of estimating high-frequency details by interpolating the input low-resolution image into the desired scale (which results in a blurred image). Then, the super-resolution is performed by the NN-based estimation of high-frequency patches based on the corresponding patches of input low-frequency image and resolving the compatibility of output patches using a Markov network. K. I. Kim is with Max-Planck-Institut für biologische Kybernetik, Spemannstr. 38, D-7076 Tübingen, Germany. [email protected] Y. Kwon is with KAIST, Gusong-dong, Yuseong-gu, Daejeon, Korea. [email protected] Although this method (and also other NN-based methods) has already shown impressive performance, there is still room for improvement if one views the image super-resolution as a regression problem, i.e., finding a map from the space of lowresolution image patches to the space of target high-resolution patches. It is well known in the machine learning community that NN-based estimation suffers from overfitting when the target function is highly complex or the data is high-dimensional [14], which is the case for image super-resolution. Accordingly, it is reasonable to expect that NN-based methods can be improved by adopting learning algorithms with regularization capability to avoid overfitting. Indeed, attempts have already been made to regularize the estimator. Chang et al. [1] regularized the NN estimator by representing the input and target image patches with linear combinations (calculated from locally linear embedding) of stored training patches (k-nns) while Datsenko and Elad [13] proposed a maximum a posteriori (MAP) framework where the prior penalizes the deviation of the solution from a weighted average of k-nns. The weights are then chosen in a manner similar to robust regression such that the contributions of the outliers are weakened. A rather straightforward approach would be to regularize the regressor directly. Based on the framework of Freeman et al. [6], [7], Kim et al. [15] have posed the problem of estimating the high-frequency details as a regression problem which is then resolved by support vector regression (SVR). Meanwhile, Ni and Nguyen [16] utilized SVR in the frequency domain and posed the super-resolution as a kernel learning problem. While SVR produced a significant improvement over existing example-based methods, it has drawbacks in building a practical system: 1. as a regularization framework, SVR tends to smooth sharp edges and produce oscillations along the major edges (ringing artifacts). These might lead to low reconstruction error on average, but is visually implausible;. SVR results in a dense solution, i.e., the regression function is expanded in the whole set of training data points and accordingly is computationally demanding both in training and in testing: optimizing the hyper-parameters based on crossvalidation indicated that the optimum value of ɛ for the ɛ- insensitive loss function of SVR is close to zero [15]. The current work extends the framework of Kim et al. [15]. 1 Kernel ridge regression (KRR) is utilized in place of SVR. Since the L -loss adopted by KRR is differentiable, we construct the sparse basis set based on the combination of the kernel matching pursuit (KMP) [18] and gradient descent, and thereby reduce the time complexity of training and testing for regression. As the regularizer of KRR is the same as that of SVR, the problem of ringing artifacts still remains. This is resolved by exploiting a prior over image structure which takes into account the discontinuity of pixel values across edges. LEARNING IMAGE SUPER-RESOLUTION Adopting the framework of Freeman et al. [6], [7], for the super-resolution of a given image, we firstly interpolate the input into the desired scale using cubic spline interpolation (henceforth referred to as interpolation ). Then, the highfrequency details which are missing in the interpolation (X) are estimated based on its band frequency components (LX) 1. A short version of this paper appeared in the proceedings of DAGM008 [17].
2 Fig. 1. Overview of super-resolution shown with examples: (a) input image is interpolated into the desired scale, (b) a set of candidate images is generated as the result of regression, (c) candidates are combined based on estimated confidences; The combined result is sharper and less noisy than individual candidates, which however shows ringing artifacts, and (d) post-processing removes ringing artifacts and further enhances edges. extracted by applying the Laplacian to X. The estimate (Y ) can then be added to X to produce the super-resolved image. A local patch-based regression (cf. Sect..1) is adopted for the estimation: LX is scanned with a patch (of size M, or M M) to produce a patch-valued regression result (of size N) for each pixel. As the output patches overlap with their neighbors, this results in a set of candidates for each pixel location which constitutes a 3-D image Z. Each candidate is obtained based on different local observation of input image and accordingly contains different partial information of the underlying high-resolution image. A single high-resolution image is then obtained as a convex combination for each pixel of the set of candidate pixels based on their estimated confidence. To enhance the visual quality around the major edges, the results are post-processed based on the prior of natural images proposed by Tappen et al. [1] (cf. Sect..). Figure 1 summarizes the super-resolution process..1 Regression and Combination The training patch pairs for the regression are randomly sampled from a set of low-resolution and corresponding desired high-resolution images (cf. Sect. 3). To avoid that the learning is distracted by uninformative patterns, the patches whose norms are close to zero are excluded from the training set. Furthermore, to increase the efficiency of the training set, the data are contrast-normalized [7]: during the construction of the training set, both the input patch and corresponding desired patches are normalized by dividing them by the L 1 -norm of the input patch. For an unseen image patch, the input is again normalized before the regression and the corresponding output is inverse normalized. For a given set of training data points {(x 1, y 1),..., (x l, y l )} R M R N, we minimize the following regularized cost functional for the regressor f = {f 1,..., f N }: E(f) = 1 ( (f i (x j) yj) i + λ i f i H ), i (1) i=1,...,n j=1,...,l where y j = [yj 1,..., yj N ] and H i is a reproducing kernel Hilbert space (RKHS). Due to the reproducing property (i.e. f, k(x, ) H = f(x)), the minimizer of above functional is expanded in kernel functions: f i ( ) = a i jk i (x j, ), for i = 1,..., N, and f i H i = j=1,...,l m,n=1,...,l a i ma i nk i (x m, x n), for i = 1,..., N, where k i is the reproducing kernel [19] for H i, e.g., to be a Gaussian kernel ) k i x y (x, y) = exp (. σk i Equation (1) is the sum of individual convex cost functionals for each scalar-valued regressor f i and the minimum can obtained analytically. However, this requires the construction and inversion of N kernel matrices ([K i (m,n)] l,l = k i (x m, x n), for i = 1,..., N) in training and N l kernel evaluations in testing, which becomes prohibitive even for a relatively small number of training data points (e.g., l 10, 000) (cf. [17] for details). In this paper, this problem is approached by trading the complexity off with the optimality of the solution by 1. tying the regularization parameter and the kernel parameter for the regressors (i.e. λ = λ i and σ k = σk i for i = 1,..., N) and. finding the minimizer of (1) only within the span of a sparse basis set {k(b 1, ),..., k(b lb, )} (l b l): f i ( ) = a i jk(b j, ), for i = 1,..., N. j=1,...,l b In this case, by sharing the evaluations of kernel functions, the time complexity of patch-valued regression reduces down to the case of scalar-valued regression, and eventually, the time complexity of testing becomes O(M l b ). Since the solution is obtained by A = (K bx K bx + λk bb ) 1 K bx Y, where [K bx(m,n) ] lb,l = k(b m, x n), [K bb(m,n) ] lb,l b = k(b m, b n), Y = [y1,..., yl ], and [A (j,i) ] lb,n = a i j, for a given fixed set of basis points B = {b 1,..., b lb }, the time complexity of training is O(lb 3 + l l b M). In general, the total training time depends on the method of finding B. Since the cost functional (1) is a differentiable function of the basis points B, it can afford gradient-based optimization as already demonstrated in the context of sparse Gaussian process regression [0]. Assuming that the evaluation of the derivative of k with respect to a basis vector takes O(M)-time, which is the case for a Gaussian kernel ( k(x, b) = b σ k k(x, b)(x b)), the evaluation of derivatives of (1) with respect to B and corresponding coefficient matrix A takes O(M l l b ):. With a slight abuse of the Matlab notation, A (m:n,:) stands for the submatrix of A obtained by extracting the rows of A from m to n. Likewise, A (:,m) is defined as the m-th column of A.
3 3 ( ) A E(f) = K bx K bxa Y + λk bb A E(f) = K bx(j,:) (K bxa Y)A (j,:) b j b j +λ K bb(j,:) b j AA (j,:), for j = 1,..., l b. () However, due to the non-convexity of (1) with respect to B, naïve gradient descent is susceptible to local minima and accordingly a good heuristic is required to initialize the solution. The KMP is adopted for this purpose. In KMP (with prefitting) [18], the basis points are selected from the training data points in an incremental way: for given n 1 basis points, the n-th basis point is chosen such that the cost functional (1) is minimized when A is optimized accordingly. 3 The basic idea is to assume that at the n-th step of KMP, the chosen basis point b n plus the accumulation of basis points obtained until the (n 1)-th step (B n 1) constitute a good initial search point. Then, at each step of KMP, B n can be subsequently optimized by gradient descent. Naïve implementation of this idea is still very expensive. To reduce further the complexity, the following simplifications are adopted: 1. In the KMP step, instead of evaluating the whole training set for choosing b n, only l c (l c l) points are considered;. Gradient descent of B n and corresponding A (1:n,:) are performed only at the every r-th KMP step. Instead, for each KMP step, only b n and A (n,:) are optimized. In this case, the gradient of (1) with respect to b n can be evaluated at O(M l)-cost. 4 Furthermore, similarly to [1], for a given b n the optimal A (n,:) can be analytically calculated at the same cost (cf. [17]). At the n-th step, the l c-candidate basis points for KMP are selected based on a rather cheap criterion. One approach might be to choose data points which show the largest distances between the corresponding function outputs obtained at the (n 1)-th step and the desired training outputs (i.e., to use the training error). However, this might tend to choose outliers as they will show relatively large training errors for regularized regression. To avoid this problem, the neighborhood context of each data point is exploited: we define a cost functional which measures the distance between the current function output and the output of a localized KRR C(x j) = K bx(1:n,:)a bx(1:n,:) g j(x j), for j = 1,..., l, where g j = [ g j 1,..., g j N ] is the localized KRR centered at the given input x j, which is obtained by collecting nearest neighbors (NNs) of x j and training the full KRR based on only these NNs. The candidate points are then chosen as the training data points corresponding to the l c-largest values of C. As a regularization method, the use of localized KRRs can effectively single out the outliers. Furthermore, in the preliminary experiments with 10,000 data points (where it was 3. In the original form of KMP, the regularization was implicitly performed by controlling the number of basis points l b (i.e., λ = 0). However, in the current problem, for a given upper bound of l b, we constantly observed a better generalization performance when we assign l b with that upper bound and control λ instead. 4. It should be noted that [[K bx ] n 1,l [A] n 1,N ] l,n (cf. ()) is stored at the (n 1)-th step. Accordingly, at the n-th step, augmenting a single row of K bx and A, respectively is sufficient for calculating the gradient. Cost Random selection K means Gradient descent KMP+gradient descent KMP # basis points Fig.. Performance of different sparse solution methods evaluated in terms of the cost functional (1) for the case of magnification factor 3; A fixed set of hyper-parameters were used for all cases such that the comparison can be made directly in (1). The performance of randomized algorithms (random selection, k- means, gradient descent) are calculated as averages of results from 0 different experiments with random initializations. The lengths of error bars correspond to twice the standard deviations. possible to train the full KRR), 5 it turned out that the outputs of localized KRR on training data points are very close to the full KRR outputs: the average squared distance between the outputs of full KRR and localized KRR was less than 1% of the mean squared training error of full KRR. Accordingly, they could be regarded as a rough estimation of full KRR solution which one might have obtained by training on all l data points. However, it should be noted that the localized KRRs cannot be directly applied for regression as they might interpolate poorly on non-training data points. To gain an insight into the performance of our basis construction method, a set of experiments has been performed with different sparse solution methods, including random selection (of basis points from the training set), KMP, k-means algorithm (clustering of training data points), naïve gradient descent (with basis initialized by k-means), and the proposed combination of KMP and gradient descent. 6 Fig. summarizes the results. The KMP showed an improved performance over the k-means algorithm and random selection which build the basis set without reflecting the cost functional to be optimized. Both of the two gradient descent methods outperformed KMP which chooses the basis points from the training set. The improved performance of gradient descent in combination with KMP could be attributed to the better initialization of the solution for the subsequent gradient descent step. As the result of the patch-based regression step, N candidates are generated for each pixel location. This setting is motivated by the observation that 1. by sharing the hyperparameters and basis points, the computational complexity of patch-valued learning reduces to the scalar-valued learning;. the candidates contain information of different input image locations which are actually diverse enough such that the 5. For preliminary experiments mentioned in this paper, we used only 10,000 training data points for training the regression part to facilitate fast evaluation. 6. For this and all the other experiments in this paper, we set the size of interval r and the number of candidate basis points l c to 10 and 100, respectively.
4 4 combination can boost the performance: for the magnification factor case, constructing an image by choosing the best and the worst (in terms of the distance to the ground truth) candidates from each spatial location of Z resulted in an average peak signal-to-noise ratio (PSNR) difference of 7.84dB (cf. the accompanying technical report [] for details). Certainly, the ground truth is not available at actual super-resolution stage and accordingly a way of constructing a single pixel out of N candidates is required. In this paper, the final estimation of the pixel value for an image location (x, y) is obtained as the convex combination of candidates given in the form of a softmax: 7 Y (x, y) = w i(x, y)z(x, y, i), where w i(x, y) = exp ( i=1,...,n ) di(x, y) / ( exp σ C j=1,...,n ) dj(x, y) σ C and {d 1(x, y),..., d N (x, y)} is the estimation of distances between the unknown desired output and each candidate. This estimate is calculated using a set of linear regressors: d i(x, y) = P Z(x, y) W i, for i = 1,..., N, where P Z(x, y) is a vector constructed by concatenating all columns of a spatial patch (of size R R N) of Z centered at (x, y) and the parameters {W i} are optimized based on the patch-based regression results (Z) for a subset of training images (cf. Sect. 3). There are a few hyper-parameters to be tuned: for the regression part, the input and output patch sizes (M and N, respectively), KRR parameters (σ k and λ), and the number of basis points (l b ) and for the combination part, the input patch size (R) and the weight parameter (σ C). We fix l b, N, and R at 300, 5(5 5), and 49(7 7), respectively. These values are determined by trading the quality of super-resolution off with the computational complexity. We observed constant increase of the performance as l b increases and becomes larger than 300. Similar tendency was also observed with increasing N(< M) and R while the run-time complexity increases linearly with all these parameters. The remaining hyper-parameters are chosen based on error rates of super-resolution results for a set of validation images. However, directly optimizing these many parameters is computationally very demanding, especially due to the large time complexity of choosing basis points. With 00,000 training data points, training a sparse KRR for a given fixed parameters took around a day on a 3GHz machine (for the magnification factor case). To retain the complexity of the whole process at a moderate level, we firstly calculate a rough estimation of parameters based on a fixed set of basis points which is obtained from the k-means algorithm. Then, the full validation is performed only at the vicinity of the rough estimation. For the distance measure of k-means clustering, we use the following combination of Euclidean distances from both the input and output spaces, which leaded to an improved performance (in terms of the KRR cost (1)) over the case of using only the input space distance: d([x i, y i], [x j, y j]) = x i x j + (σ X /σ Y) y i y j, 7. Discussion on alternative combination methods can be found in []. TABLE 1 Parameters for experiments Mag. factor 3 4 M σ k σ C λ σ N σ R T M T M where σ X and σ Y are variances of distances between pairs of training data points in the input space and output space, respectively. It should be noted that the optimization of hyper-parameters for the regression and combination parts should not be separated: choosing the hyper-parameters of regression part based on cross-validation of regression data (pairs of input and output patches) leaded to much more conservative estimation (i.e., σ k and λ are larger) than the case of optimizing jointly the regression and combination parts. This can be explained by (further) regularization effect of the combination part which can be regarded as an instance of ensemble estimator. It has been well known that in general, ensembles of individual estimators can lead to lower variances (expectation of variance of the output for a given set of training data points) and accordingly are smoother than individual estimators (Ch. 7 of [3] and references therein). This makes the optimization criteria a non-differentiable function of hyper-parameters and prevents us from using a rather sophisticate parameter optimization methods, e.g., gradient ascent of the marginal likelihood [0]. In the experiments, we focused on the desired magnification factors at {, 3, 4} along each dimension. Application to other magnification factors should be straightforward. Table 1 summarizes the optimized parameters.. Post-processing Based on Image Prior As demonstrated in Fig. 1, the result of the proposed regression-based method is significantly better than the interpolation. However, detailed visual inspection along the major edges (edges showing rapid and strong change of pixel values) reveals ringing artifacts. In general, regularization methods (depending on the specific class of regularizer) including KRR and SVR tend to fit the data with a smooth function. Accordingly, at the sharp changes of the function (edges in the case of images), either edges are smoothed or oscillation occurs to compensate the resulting loss of smoothness. This might happen for all the levels of images demonstrating the discontinuity. However, the magnitude of oscillation is in proportion to the magnitude of changes and accordingly only visible at the vicinity of major edges. While this problem can indirectly be resolved by imposing less aggressive regularization at the edges, a more direct approach is to rely on the prior knowledge of discontinuity of images. In this work, we use a modification of the natural image prior (NIP) framework proposed by Tappen et al. [1] to the pixels at the vicinity of edges:
5 5 (a) (b) Fig. 3. Factor graph representation for the optimization of (3): (a) NIP term (message propagation from node j to node i) and (b) deviation penalty term of node j; the message from the observation variable node yj to the factor node [j] is a constant. P ({x} {y}) = α xj xi exp σn (j,i NS (j)) " # Y xj yj exp, σr j 1 C Y (3) where {y} represents the observed variables corresponding to the pixel values of Y, {x} represents the latent variable, NS (j) stands for the 8-connected neighbors of the pixel location j, and C is a normalization constant. With the objective of achieving the maximum probability (equivalently, the minimum energy as the inverse of (3)) for a given image, the second product term has the role of preventing the final solution flowing far away from the input regression-based super-resolution result Y, while the first product term (NIP term) tends to smooth the image based on the costs x j x i. The role of α(< 1) is to re-weight the costs such that the largest difference is stressed relatively less than the others and accordingly large changes of pixel values are relatively less penalized. Furthermore, the cost term x j x i α becomes piece-wise concave with boundary points (i.e., boundaries between concave intervals) at NS (j) such that if the second term is removed, the minimum energy for a pixel j is achieved by assigning it with the value of a neighbor, rather than a certain weighted average of neighborhood values which might have been the case when α > 1. Accordingly, this distribution prefers a strong edge rather than a set of small edges and can be used to resolve the problem of smoothing around major edges. The optimization of (3) is performed by a max-sum type belief propagation (BP) similarly to [1]. To facilitate the optimization, we reuse the candidate set generated from the regression step so that the best candidates are chosen by the BP. Accordingly, all possible outputs for each pixel location are constrained to be the N candidates generated during the regression step. In the original NIP framework, the second term is replaced by the reconstruction constraint which measures the distance between the input low-resolution image and an image reconstructed from the high-resolution configuration according to the down-sampling model (blurring and sub-sampling) [1], [4]. The reconstruction constraint corresponds to a generative model, and with the suitable prior (e.g., NIP), provides a MAP framework. However, without the existence of multiple images, which might have guided better the reconstruction, relying on the reconstruction constraint in the proposed method could result in noisy images as the down-sampling process has the effect of removing noises and can make it harder Fig. 4. Gallery of test images (disjoint from training images): we refer to the images in the text by its position in raster order. to penalize the noisy configuration. (cf. [5]).8 Furthermore, we have found that it is not straightforward to control the contribution of NIP part to prevent this effect as it often leaded to a piece-wise constant image. Accordingly, in this work, we simply penalize the deviation from the regression output (Y ) which is far less noisy. The main disadvantage of the proposed scheme in comparison to the original NIP is that the intuitive probabilistic interpretation of super-resolution process [1] is no longer possible. However, on the other hand, since the resulting message structure is significantly simpler than the original version, the optimization can be made much faster: 1 xj yj ν[j] j (xj ) = σr α 1 xj xi µ[i,j] i (xi ) = max µj [i,j] (xj ) xj σn X µj [i,j] (xj ) = ν[j] j (xj ) + µ[j,k] j (xj ). k NS (j)\i These (logarithms of) messages can be derived from (3) based on the factor graph representation of Fig. 3. The message ν[j] j represents the reconstruction constraint at the variable node j while the other two messages correspond to the propagation of a belief from j to i based on the NIP cost. The outgoing message µj [i,j] from j to the factor node [i, j] is composed of the sum of ν[j] j and all the messages from the neighboring factor nodes of j except for the node [i, j]. The message µ[i,j] i is calculated as the maximum of the sum of µj [i,j] and (the logarithm of) the NIP cost over all the latent values xj. The major edges are found by thresholding each pixel based on the L norm of the Laplacian and the range of pixel values in the local patches, i.e., classifying a pixel into major edge class if the norm of Laplacian and the maximum difference of pixel values within a local patch are larger than thresholds TM 1 and TM, respectively (cf. Table 1; see [] for details of parameter optimization). While the improvements in terms of PSNR are not significant (e.g., for the case of magnification factor, on average dB from the combined regression result) the improved visual quality at major edges demonstrate the effectiveness of using the prior of natural images (Figs. 1 and 5). 3 E XPERIMENTS For training and quantitative evaluation, a set of pairs of high-resolution and corresponding low-resolution images were 8. In original work of Tappen et al. [1], the set of possible configurations is much more constrained than that of our method: candidates are -size image patches rather than individual pixels. Accordingly, in their method this problem is not as serious as the case of naı vely using the reconstruction constraint in the proposed method (cf. Appendix in the supplementary material for more discussion).
6 6 TABLE Performance of different example-based super-resolution algorithms: mean improvement (standard deviation) of PSNR values from the input interpolation Mag. factor 3 4 NN 0.11(0.4) N/A -0.85(0.56) LLE -0.18(0.31) -0.17(0.45) -0.5(0.3) NIP -0.50(0.51) N/A N/A SVR 1.31(0.41) 0.8(0.30) 0.79(0.44) Proposed method 1.91(0.58) 1.34(0.47) 1.15(0.56) obtained by blurring and subsampling 9 a set of high-resolution images (the test images are shown in Fig. 4). For comparison, several different example-based image super-resolution methods were implemented, which include Freeman et al. s fast NN-based method [7], Chang et al. s LLE-based method [1], Tappen et al. s NIP [1], 10 and our previous SVR-based method [15] (trained based on only 10,000 data points). Experiments with Tappen et al. s NIP were performed only at the magnification factor as it was not straightforward to implement it for the other magnification factors. For the same reason, Freeman et al. s NN method was applied only to the case of magnification factors and 4. For comparison with nonexample-based methods which are not implemented by us, we performed super-resolution on several images downloaded from the websites of the authors of [3], [4], [6]. 11 To obtain super-resolution results at image boundary, which are not directly available as M > N for the proposed methods and similarity for other example-based methods, the input images were extended by symmetrically replicating pixel values across the image boundary. For the experiments with color images, we applied the model trained on intensity images to each RGB channel and combined them. Figures 5 and 6 show examples of super-resolution. All the example-based super-resolution methods outperformed the spline interpolation in terms of visual plausibility. The NNbased method and the original NIP produced sharper images at the expense of introducing noise which, even with the improved visual quality, led to lower PSNR values than the interpolations (Table ). The results of LLE are less noisy. However, it tended to smooth out texture details as observed in the third image of Fig.5(c) and accordingly produced low PSNR values. The SVR produced less noisy images, but it generated smoothed edges and perceptually distracting ring artifacts which have almost disappeared in the results of the proposed method (e.g., the first and the fourth images of Fig. 5(d)). Disregarding the post-processing stage, we measured on average 0.60dB improvement of PSNRs for the proposed method from the SVR (magnification factor case). This could be attributed to the sparsity of the solution which enabled training on a large data set and the effectiveness of the 9. We use spline resampling which is naturally unbiased to any specific direction in the generation of low-resolution images []. 10. The original NIP algorithm was developed for super-resolving the pixel subsampled image. Accordingly, for the experiments with NIP, the low resolution images were generated by pixel subsampling. The visual qualities of the super-resolution results are not significantly different from the results obtained from spline resampling. However, the quantitative results should not be directly compared with other methods. The parameters used for experiments in the current work simply follow those described in [1]. 11. The original images and the results of [3], [4], and [6] are courtesy of Shengyang Dai, Raanan Fattal, and William T. Freeman, respectively. candidate combination scheme. Moreover, in comparison to SVR, the proposed method requires much less processing time: super-resolving a size image into requires around 7 seconds for the proposed method and 18 minutes for the SVR-based method on a 3GHz machine. For quantitative comparison, PSNRs of different algorithms are summarized in Table. An interesting property of NN-based method is that it introduced certain texture details which were absent in the input low-resolution images and even in the ground truth images. Sometimes, these pseudo textures provided more realistic images than others (e.g., the fifth image of Fig. 5(b)). On the other hand, the proposed method did not generate such new texture details but instead provided a coherent enhancement of existing texture and edge patterns (cf. Fig. 5(g)). As noted in [4], a preference between the two techniques may depend on the specific image and subjective concerns. In comparison with non-example-based methods of Dai et al. [3] and Fattal [4], the proposed method resulted in a better preservation of texture details and more natural transitions of pixel values across strong edges as shown in the stripe pattern of Fig. 6(c). Furthermore, the results of the proposed method look less jagged as observed in petals in the first row of Fig. 6(f). 4 DISCUSSION Except for the preprocessing part (interpolation and the calculation of Laplacian), the proposed method is application agnostic, i.e., the learning part is independent of specific problem at hand. In principle, this generic learning part can be applied to any problem when suitable examples of input and target output images are available. Accordingly, future work will include exploring the potential of learning-based approaches, including the proposed method, for various image enhancement and understanding applications. In the appendix provided in the accompanying supplementary material of the current paper, we show an application of the proposed method to artifact removal of JPEG encoded images. ACKNOWLEDGMENTS The contents of this paper have been greatly benefited from discussions with Gökhan Bakır, Christian Walder, Matthew Blaschko, Christoph Lampert, Daewon Lee, and Matthias Hein and also from the comments of anonymous reviewers. The idea of using localized KRR was originated by Christian Walder. The authors would like to thank Dong Ho Kim for helping with experiments. K. I. Kim was supported by the Korea Research Foundation Grant funded by the Korean Government (MOEHRD) (KRF D00348). REFERENCES [1] M. F. Tappen, B. C. Russel, and W. T. Freeman, Exploiting the sparse derivative prior for super-resolution and image demosaicing, in Proc. IEEE Workshop on Statistical and Computational Theories of Vision, 003. [] D. Tschumperlé and R. Deriche, Vector-valued image regularization with pdes: a common framework for different applications, IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 7, no. 4, pp , 005. [3] S. Dai, M. Han, W. Xu, Y. Wu, and Y. Gong, Soft edge smothness prior for alpha channel super resolution, in Proc. IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 007, pp. 1 8.
7 7 (a) (b) (c) (d) N/A (e) (f) (g) (h) Fig. 5. Super-resolution examples of example-based algorithms: (a) interpolations, (b)-(f) super-resolution results of NN [7], LLE [1], NIP [1], SVR [15], and proposed method, respectively, (g) original high-resolution images, and (h) differences of the images in (f) and (a), respectively, which correspond to the details estimated by the proposed method. Magnification factors are and 4 for the first four columns and the last two columns, respectively. Experiments with NIP [1] were performed only at the magnification factor case (see texts for details). Please refer to the electronic version of the current paper for better visualization. [4] R. Fattal, Image upsampling via imposed edge statistics, ACM Trans. Graphics (Proc. SIGGRAPH 007), vol. 6, no. 3, pp. 95:1 95:8, 007. [5] S. Baker and T. Kanade, Limits on super-resolution and how to break them, IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 4, no. 9, pp , 00. [6] W. T. Freeman, E. C. Pasztor, and O. T. Carmichael, Learning low-level vision, International Journal of Computer Vision, vol. 40, no. 1, pp. 5 47, 000. [7] W. T. Freeman, T. R. Jones, and E. C. Pasztor, Examplebased super-resolution, IEEE Computer Graphics and Applications, vol., no., pp , 00. [8] A. Hertzmann, C. E. Jacobs, N. Oliver, B. Curless, and D. H. Salesin, Image analogies, in Computer Graphics (Proc. Siggraph 001). NY: ACM Press, 001, pp [9] K. I. Kim, M. O. Franz, and B. Schölkopf, Iterative kernel principal component analysis for image modeling, IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 7, no. 9, pp , 005. [10] L. C. Pickup, S. J. Roberts, and A. Zissermann, A sampled texture prior for image super-resolution, in Advances in Neural Information Processing Systems. Cambridge, MA: MIT Press, 004. [11] C. V. Jiji and S. Chaudhuri, Single-frame image super-resolution through contourlet learning, Journal of Multidimensional System and Signal Processing, vol. 006, pp. 1 11, 006. [1] H. Chang, D.-Y. Yeung, and Y. Xiong, Super-resolution through neighbor embedding, in Proc. IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 004, pp [13] D. Datsenko and M. Elad, Example-based single image superresolution: a global MAP approach with outlier rejection, Journal of Multidimensional System and Signal Processing, vol. 18, no. 3, pp , 007. [14] T. Hastie, R. Tibshirani, and J. H. Friedman, The Elements of Statistical Learning. New York: Springer-Verlag, 001. [15] K. I. Kim, D. H. Kim, and J.-H. Kim, Example-based learning for image super-resolution, in Proc. the third Tsinghua-KAIST Joint
8 8 (a) (b) (c) (d) (e) (f) (g) (h) (i) (j) (k) (l) Fig. 6. Comparison between super-resolution results of several different algorithms: (a), (d), and (g) interpolations (magnification factors 3, 4, and 4, respectively), (b) and (h) Dai et al. [3], (e) Fattal [4], (i) Freeman et al. [6], (j) Freeman et al. [7], (k) Chang et al. [1], and (c), (f), and (l) proposed method. Workshop on Pattern Recognition, 004, pp [16] K. Ni and T. Q. Nguyen, Image superresolution using support vector regression, IEEE Trans. Image Processing, vol. 16, no. 6, pp , 007. [17] K. I. Kim and Y. Kwon, Example-based learning for single image super-resolution, in Proc. DAGM, 008, pp [18] P. Vincent and Y. Bengio, Kernel matching pursuit, Machine Learning, vol. 48, pp , 00. [19] B. Scho lkopf and A. Smola, Learning with Kernels. Cambridge, MA: MIT Press, 00. [0] E. Snelson and Z. Ghahramani, Sparse gaussian processes using pseudo-inputs, in Advances in Neural Information Processing Systems. Cambridge, MA: MIT Press, 006. [1] S. S. Keerthi and W. Chu, A matching pursuit approach to sparse gaussian process regression, in Advances in Neural Information Processing Systems. Cambridge, MA: MIT Press, 005. [] K. I. Kim and Y. Kwon, Example-based learning for single-image super-resolution and JPEG artifact removal, Max-Planck-Insitut fu r biologische Kybernetik, Tu bingen, Tech. Rep. 173, August 008. [Online]. Available: attachments/techreport-173 [0].pdf [3] S. Haykin, Neural Networks: A Comprehensive Foundation, nd ed. New Jersey: Prentice Hall, [4] Z. Lin and H.-Y. Shum, Fundamental limits of reconstructionbased superresolution algorithms under local translation, IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 6, no. 1, pp , 004.
How To Solve Krr
Example-based Learning for Single-Image Super-resolution Kwang In Kim 1 and Younghee Kwon 2 1 Max-Planck-Institute für biologische Kybernetik, Spemannstr. 38, D-72076 Tübingen, Germany 2 Korea Advanced
A Learning Based Method for Super-Resolution of Low Resolution Images
A Learning Based Method for Super-Resolution of Low Resolution Images Emre Ugur June 1, 2004 [email protected] Abstract The main objective of this project is the study of a learning based method
Super-resolution Reconstruction Algorithm Based on Patch Similarity and Back-projection Modification
1862 JOURNAL OF SOFTWARE, VOL 9, NO 7, JULY 214 Super-resolution Reconstruction Algorithm Based on Patch Similarity and Back-projection Modification Wei-long Chen Digital Media College, Sichuan Normal
SUPER-RESOLUTION (SR) has been an active research
498 IEEE TRANSACTIONS ON MULTIMEDIA, VOL. 15, NO. 3, APRIL 2013 A Self-Learning Approach to Single Image Super-Resolution Min-Chun Yang and Yu-Chiang Frank Wang, Member, IEEE Abstract Learning-based approaches
NEIGHBORHOOD REGRESSION FOR EDGE-PRESERVING IMAGE SUPER-RESOLUTION. Yanghao Li, Jiaying Liu, Wenhan Yang, Zongming Guo
NEIGHBORHOOD REGRESSION FOR EDGE-PRESERVING IMAGE SUPER-RESOLUTION Yanghao Li, Jiaying Liu, Wenhan Yang, Zongming Guo Institute of Computer Science and Technology, Peking University, Beijing, P.R.China,
Single Image Super-Resolution using Gaussian Process Regression
Single Image Super-Resolution using Gaussian Process Regression He He and Wan-Chi Siu Department of Electronic and Information Engineering The Hong Kong Polytechnic University {07821020d, [email protected]}
High Quality Image Magnification using Cross-Scale Self-Similarity
High Quality Image Magnification using Cross-Scale Self-Similarity André Gooßen 1, Arne Ehlers 1, Thomas Pralow 2, Rolf-Rainer Grigat 1 1 Vision Systems, Hamburg University of Technology, D-21079 Hamburg
Image Super-Resolution as Sparse Representation of Raw Image Patches
Image Super-Resolution as Sparse Representation of Raw Image Patches Jianchao Yang, John Wright, Yi Ma, Thomas Huang University of Illinois at Urbana-Champagin Beckman Institute and Coordinated Science
A Novel Method to Improve Resolution of Satellite Images Using DWT and Interpolation
A Novel Method to Improve Resolution of Satellite Images Using DWT and Interpolation S.VENKATA RAMANA ¹, S. NARAYANA REDDY ² M.Tech student, Department of ECE, SVU college of Engineering, Tirupati, 517502,
Bayesian Image Super-Resolution
Bayesian Image Super-Resolution Michael E. Tipping and Christopher M. Bishop Microsoft Research, Cambridge, U.K..................................................................... Published as: Bayesian
High Quality Image Deblurring Panchromatic Pixels
High Quality Image Deblurring Panchromatic Pixels ACM Transaction on Graphics vol. 31, No. 5, 2012 Sen Wang, Tingbo Hou, John Border, Hong Qin, and Rodney Miller Presented by Bong-Seok Choi School of Electrical
Image Super-Resolution via Sparse Representation
1 Image Super-Resolution via Sparse Representation Jianchao Yang, Student Member, IEEE, John Wright, Student Member, IEEE Thomas Huang, Life Fellow, IEEE and Yi Ma, Senior Member, IEEE Abstract This paper
Context-Constrained Hallucination for Image Super-Resolution
Context-Constrained Hallucination for Image Super-Resolution Jian Sun Xi an Jiaotong University Xi an, P. R. China [email protected] Jiejie Zhu Marshall F. Tappen EECS, University of Central Florida
Face Model Fitting on Low Resolution Images
Face Model Fitting on Low Resolution Images Xiaoming Liu Peter H. Tu Frederick W. Wheeler Visualization and Computer Vision Lab General Electric Global Research Center Niskayuna, NY, 1239, USA {liux,tu,wheeler}@research.ge.com
Supporting Online Material for
www.sciencemag.org/cgi/content/full/313/5786/504/dc1 Supporting Online Material for Reducing the Dimensionality of Data with Neural Networks G. E. Hinton* and R. R. Salakhutdinov *To whom correspondence
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
Class #6: Non-linear classification. ML4Bio 2012 February 17 th, 2012 Quaid Morris
Class #6: Non-linear classification ML4Bio 2012 February 17 th, 2012 Quaid Morris 1 Module #: Title of Module 2 Review Overview Linear separability Non-linear classification Linear Support Vector Machines
Support Vector Machines with Clustering for Training with Very Large Datasets
Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France [email protected] Massimiliano
Super-Resolution Through Neighbor Embedding
Super-Resolution Through Neighbor Embedding Hong Chang, Dit-Yan Yeung, Yimin Xiong Department of Computer Science Hong Kong University of Science and Technology Clear Water Bay, Kowloon, Hong Kong {hongch,
LABEL PROPAGATION ON GRAPHS. SEMI-SUPERVISED LEARNING. ----Changsheng Liu 10-30-2014
LABEL PROPAGATION ON GRAPHS. SEMI-SUPERVISED LEARNING ----Changsheng Liu 10-30-2014 Agenda Semi Supervised Learning Topics in Semi Supervised Learning Label Propagation Local and global consistency Graph
Accurate and robust image superresolution by neural processing of local image representations
Accurate and robust image superresolution by neural processing of local image representations Carlos Miravet 1,2 and Francisco B. Rodríguez 1 1 Grupo de Neurocomputación Biológica (GNB), Escuela Politécnica
Introduction to Support Vector Machines. Colin Campbell, Bristol University
Introduction to Support Vector Machines Colin Campbell, Bristol University 1 Outline of talk. Part 1. An Introduction to SVMs 1.1. SVMs for binary classification. 1.2. Soft margins and multi-class classification.
Super-Resolution from a Single Image
Super-Resolution from a Single Image Daniel Glasner Shai Bagon Michal Irani Dept. of Computer Science and Applied Mathematics The Weizmann Institute of Science Rehovot 76100, Israel Abstract Methods for
Applications to Data Smoothing and Image Processing I
Applications to Data Smoothing and Image Processing I MA 348 Kurt Bryan Signals and Images Let t denote time and consider a signal a(t) on some time interval, say t. We ll assume that the signal a(t) is
Multidimensional Scaling for Matching. Low-resolution Face Images
Multidimensional Scaling for Matching 1 Low-resolution Face Images Soma Biswas, Member, IEEE, Kevin W. Bowyer, Fellow, IEEE, and Patrick J. Flynn, Senior Member, IEEE Abstract Face recognition performance
Low-resolution Character Recognition by Video-based Super-resolution
2009 10th International Conference on Document Analysis and Recognition Low-resolution Character Recognition by Video-based Super-resolution Ataru Ohkura 1, Daisuke Deguchi 1, Tomokazu Takahashi 2, Ichiro
Sachin Patel HOD I.T Department PCST, Indore, India. Parth Bhatt I.T Department, PCST, Indore, India. Ankit Shah CSE Department, KITE, Jaipur, India
Image Enhancement Using Various Interpolation Methods Parth Bhatt I.T Department, PCST, Indore, India Ankit Shah CSE Department, KITE, Jaipur, India Sachin Patel HOD I.T Department PCST, Indore, India
The Role of Size Normalization on the Recognition Rate of Handwritten Numerals
The Role of Size Normalization on the Recognition Rate of Handwritten Numerals Chun Lei He, Ping Zhang, Jianxiong Dong, Ching Y. Suen, Tien D. Bui Centre for Pattern Recognition and Machine Intelligence,
An Introduction to Machine Learning
An Introduction to Machine Learning L5: Novelty Detection and Regression Alexander J. Smola Statistical Machine Learning Program Canberra, ACT 0200 Australia [email protected] Tata Institute, Pune,
A NEW SUPER RESOLUTION TECHNIQUE FOR RANGE DATA. Valeria Garro, Pietro Zanuttigh, Guido M. Cortelazzo. University of Padova, Italy
A NEW SUPER RESOLUTION TECHNIQUE FOR RANGE DATA Valeria Garro, Pietro Zanuttigh, Guido M. Cortelazzo University of Padova, Italy ABSTRACT Current Time-of-Flight matrix sensors allow for the acquisition
Redundant Wavelet Transform Based Image Super Resolution
Redundant Wavelet Transform Based Image Super Resolution Arti Sharma, Prof. Preety D Swami Department of Electronics &Telecommunication Samrat Ashok Technological Institute Vidisha Department of Electronics
PERFORMANCE ANALYSIS OF HIGH RESOLUTION IMAGES USING INTERPOLATION TECHNIQUES IN MULTIMEDIA COMMUNICATION SYSTEM
PERFORMANCE ANALYSIS OF HIGH RESOLUTION IMAGES USING INTERPOLATION TECHNIQUES IN MULTIMEDIA COMMUNICATION SYSTEM Apurva Sinha 1, Mukesh kumar 2, A.K. Jaiswal 3, Rohini Saxena 4 Department of Electronics
Couple Dictionary Training for Image Super-resolution
IEEE TRANSACTIONS ON IMAGE PROCESSING 1 Couple Dictionary Training for Image Super-resolution Jianchao Yang, Student Member, IEEE, Zhaowen Wang, Student Member, IEEE, Zhe Lin, Member, IEEE, Scott Cohen,
Visualization by Linear Projections as Information Retrieval
Visualization by Linear Projections as Information Retrieval Jaakko Peltonen Helsinki University of Technology, Department of Information and Computer Science, P. O. Box 5400, FI-0015 TKK, Finland [email protected]
Single Depth Image Super Resolution and Denoising Using Coupled Dictionary Learning with Local Constraints and Shock Filtering
Single Depth Image Super Resolution and Denoising Using Coupled Dictionary Learning with Local Constraints and Shock Filtering Jun Xie 1, Cheng-Chuan Chou 2, Rogerio Feris 3, Ming-Ting Sun 1 1 University
Image Interpolation by Pixel Level Data-Dependent Triangulation
Volume xx (200y), Number z, pp. 1 7 Image Interpolation by Pixel Level Data-Dependent Triangulation Dan Su, Philip Willis Department of Computer Science, University of Bath, Bath, BA2 7AY, U.K. mapds,
Image Segmentation and Registration
Image Segmentation and Registration Dr. Christine Tanner ([email protected]) Computer Vision Laboratory, ETH Zürich Dr. Verena Kaynig, Machine Learning Laboratory, ETH Zürich Outline Segmentation
Least Squares Estimation
Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN-13: 978-0-470-86080-9 ISBN-10: 0-470-86080-4 Editors Brian S Everitt & David
Machine Learning and Pattern Recognition Logistic Regression
Machine Learning and Pattern Recognition Logistic Regression Course Lecturer:Amos J Storkey Institute for Adaptive and Neural Computation School of Informatics University of Edinburgh Crichton Street,
Statistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
Adaptive Online Gradient Descent
Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches
Modelling, Extraction and Description of Intrinsic Cues of High Resolution Satellite Images: Independent Component Analysis based approaches PhD Thesis by Payam Birjandi Director: Prof. Mihai Datcu Problematic
JPEG compression of monochrome 2D-barcode images using DCT coefficient distributions
Edith Cowan University Research Online ECU Publications Pre. JPEG compression of monochrome D-barcode images using DCT coefficient distributions Keng Teong Tan Hong Kong Baptist University Douglas Chai
Performance Verification of Super-Resolution Image Reconstruction
Performance Verification of Super-Resolution Image Reconstruction Masaki Sugie Department of Information Science, Kogakuin University Tokyo, Japan Email: [email protected] Seiichi Gohshi Department
Tree based ensemble models regularization by convex optimization
Tree based ensemble models regularization by convex optimization Bertrand Cornélusse, Pierre Geurts and Louis Wehenkel Department of Electrical Engineering and Computer Science University of Liège B-4000
DYNAMIC RANGE IMPROVEMENT THROUGH MULTIPLE EXPOSURES. Mark A. Robertson, Sean Borman, and Robert L. Stevenson
c 1999 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or
Recognizing Cats and Dogs with Shape and Appearance based Models. Group Member: Chu Wang, Landu Jiang
Recognizing Cats and Dogs with Shape and Appearance based Models Group Member: Chu Wang, Landu Jiang Abstract Recognizing cats and dogs from images is a challenging competition raised by Kaggle platform
Image Super-Resolution Using Deep Convolutional Networks
1 Image Super-Resolution Using Deep Convolutional Networks Chao Dong, Chen Change Loy, Member, IEEE, Kaiming He, Member, IEEE, and Xiaoou Tang, Fellow, IEEE arxiv:1501.00092v3 [cs.cv] 31 Jul 2015 Abstract
Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data
CMPE 59H Comparison of Non-linear Dimensionality Reduction Techniques for Classification with Gene Expression Microarray Data Term Project Report Fatma Güney, Kübra Kalkan 1/15/2013 Keywords: Non-linear
Regularized Logistic Regression for Mind Reading with Parallel Validation
Regularized Logistic Regression for Mind Reading with Parallel Validation Heikki Huttunen, Jukka-Pekka Kauppi, Jussi Tohka Tampere University of Technology Department of Signal Processing Tampere, Finland
Social Media Mining. Data Mining Essentials
Introduction Data production rate has been increased dramatically (Big Data) and we are able store much more data than before E.g., purchase data, social media data, mobile phone data Businesses and customers
CS 2750 Machine Learning. Lecture 1. Machine Learning. http://www.cs.pitt.edu/~milos/courses/cs2750/ CS 2750 Machine Learning.
Lecture Machine Learning Milos Hauskrecht [email protected] 539 Sennott Square, x5 http://www.cs.pitt.edu/~milos/courses/cs75/ Administration Instructor: Milos Hauskrecht [email protected] 539 Sennott
Mean-Shift Tracking with Random Sampling
1 Mean-Shift Tracking with Random Sampling Alex Po Leung, Shaogang Gong Department of Computer Science Queen Mary, University of London, London, E1 4NS Abstract In this work, boosting the efficiency of
Smoothing and Non-Parametric Regression
Smoothing and Non-Parametric Regression Germán Rodríguez [email protected] Spring, 2001 Objective: to estimate the effects of covariates X on a response y nonparametrically, letting the data suggest
Making Sense of the Mayhem: Machine Learning and March Madness
Making Sense of the Mayhem: Machine Learning and March Madness Alex Tran and Adam Ginzberg Stanford University [email protected] [email protected] I. Introduction III. Model The goal of our research
An Iterative Image Registration Technique with an Application to Stereo Vision
An Iterative Image Registration Technique with an Application to Stereo Vision Bruce D. Lucas Takeo Kanade Computer Science Department Carnegie-Mellon University Pittsburgh, Pennsylvania 15213 Abstract
Machine Learning in FX Carry Basket Prediction
Machine Learning in FX Carry Basket Prediction Tristan Fletcher, Fabian Redpath and Joe D Alessandro Abstract Artificial Neural Networks ANN), Support Vector Machines SVM) and Relevance Vector Machines
Cluster Analysis: Advanced Concepts
Cluster Analysis: Advanced Concepts and dalgorithms Dr. Hui Xiong Rutgers University Introduction to Data Mining 08/06/2006 1 Introduction to Data Mining 08/06/2006 1 Outline Prototype-based Fuzzy c-means
Part-Based Recognition
Part-Based Recognition Benedict Brown CS597D, Fall 2003 Princeton University CS 597D, Part-Based Recognition p. 1/32 Introduction Many objects are made up of parts It s presumably easier to identify simple
3. Interpolation. Closing the Gaps of Discretization... Beyond Polynomials
3. Interpolation Closing the Gaps of Discretization... Beyond Polynomials Closing the Gaps of Discretization... Beyond Polynomials, December 19, 2012 1 3.3. Polynomial Splines Idea of Polynomial Splines
Video stabilization for high resolution images reconstruction
Advanced Project S9 Video stabilization for high resolution images reconstruction HIMMICH Youssef, KEROUANTON Thomas, PATIES Rémi, VILCHES José. Abstract Super-resolution reconstruction produces one or
Local Gaussian Process Regression for Real Time Online Model Learning and Control
Local Gaussian Process Regression for Real Time Online Model Learning and Control Duy Nguyen-Tuong Jan Peters Matthias Seeger Max Planck Institute for Biological Cybernetics Spemannstraße 38, 776 Tübingen,
Admin stuff. 4 Image Pyramids. Spatial Domain. Projects. Fourier domain 2/26/2008. Fourier as a change of basis
Admin stuff 4 Image Pyramids Change of office hours on Wed 4 th April Mon 3 st March 9.3.3pm (right after class) Change of time/date t of last class Currently Mon 5 th May What about Thursday 8 th May?
Manifold Learning Examples PCA, LLE and ISOMAP
Manifold Learning Examples PCA, LLE and ISOMAP Dan Ventura October 14, 28 Abstract We try to give a helpful concrete example that demonstrates how to use PCA, LLE and Isomap, attempts to provide some intuition
Analecta Vol. 8, No. 2 ISSN 2064-7964
EXPERIMENTAL APPLICATIONS OF ARTIFICIAL NEURAL NETWORKS IN ENGINEERING PROCESSING SYSTEM S. Dadvandipour Institute of Information Engineering, University of Miskolc, Egyetemváros, 3515, Miskolc, Hungary,
Two Topics in Parametric Integration Applied to Stochastic Simulation in Industrial Engineering
Two Topics in Parametric Integration Applied to Stochastic Simulation in Industrial Engineering Department of Industrial Engineering and Management Sciences Northwestern University September 15th, 2014
Character Image Patterns as Big Data
22 International Conference on Frontiers in Handwriting Recognition Character Image Patterns as Big Data Seiichi Uchida, Ryosuke Ishida, Akira Yoshida, Wenjie Cai, Yaokai Feng Kyushu University, Fukuoka,
Local features and matching. Image classification & object localization
Overview Instance level search Local features and matching Efficient visual recognition Image classification & object localization Category recognition Image classification: assigning a class label to
Extracting a Good Quality Frontal Face Images from Low Resolution Video Sequences
Extracting a Good Quality Frontal Face Images from Low Resolution Video Sequences Pritam P. Patil 1, Prof. M.V. Phatak 2 1 ME.Comp, 2 Asst.Professor, MIT, Pune Abstract The face is one of the important
Class-specific Sparse Coding for Learning of Object Representations
Class-specific Sparse Coding for Learning of Object Representations Stephan Hasler, Heiko Wersing, and Edgar Körner Honda Research Institute Europe GmbH Carl-Legien-Str. 30, 63073 Offenbach am Main, Germany
Classification of Fingerprints. Sarat C. Dass Department of Statistics & Probability
Classification of Fingerprints Sarat C. Dass Department of Statistics & Probability Fingerprint Classification Fingerprint classification is a coarse level partitioning of a fingerprint database into smaller
Bildverarbeitung und Mustererkennung Image Processing and Pattern Recognition
Bildverarbeitung und Mustererkennung Image Processing and Pattern Recognition 1. Image Pre-Processing - Pixel Brightness Transformation - Geometric Transformation - Image Denoising 1 1. Image Pre-Processing
Environmental Remote Sensing GEOG 2021
Environmental Remote Sensing GEOG 2021 Lecture 4 Image classification 2 Purpose categorising data data abstraction / simplification data interpretation mapping for land cover mapping use land cover class
Sharpening through spatial filtering
Sharpening through spatial filtering Stefano Ferrari Università degli Studi di Milano [email protected] Elaborazione delle immagini (Image processing I) academic year 2011 2012 Sharpening The term
Convolution. 1D Formula: 2D Formula: Example on the web: http://www.jhu.edu/~signals/convolve/
Basic Filters (7) Convolution/correlation/Linear filtering Gaussian filters Smoothing and noise reduction First derivatives of Gaussian Second derivative of Gaussian: Laplacian Oriented Gaussian filters
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
Semi-Supervised Support Vector Machines and Application to Spam Filtering
Semi-Supervised Support Vector Machines and Application to Spam Filtering Alexander Zien Empirical Inference Department, Bernhard Schölkopf Max Planck Institute for Biological Cybernetics ECML 2006 Discovery
Least-Squares Intersection of Lines
Least-Squares Intersection of Lines Johannes Traa - UIUC 2013 This write-up derives the least-squares solution for the intersection of lines. In the general case, a set of lines will not intersect at a
COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS
COMPARISON OF OBJECT BASED AND PIXEL BASED CLASSIFICATION OF HIGH RESOLUTION SATELLITE IMAGES USING ARTIFICIAL NEURAL NETWORKS B.K. Mohan and S. N. Ladha Centre for Studies in Resources Engineering IIT
Correlation and Convolution Class Notes for CMSC 426, Fall 2005 David Jacobs
Correlation and Convolution Class otes for CMSC 46, Fall 5 David Jacobs Introduction Correlation and Convolution are basic operations that we will perform to extract information from images. They are in
Efficient online learning of a non-negative sparse autoencoder
and Machine Learning. Bruges (Belgium), 28-30 April 2010, d-side publi., ISBN 2-93030-10-2. Efficient online learning of a non-negative sparse autoencoder Andre Lemme, R. Felix Reinhart and Jochen J. Steil
Java Modules for Time Series Analysis
Java Modules for Time Series Analysis Agenda Clustering Non-normal distributions Multifactor modeling Implied ratings Time series prediction 1. Clustering + Cluster 1 Synthetic Clustering + Time series
Segmentation & Clustering
EECS 442 Computer vision Segmentation & Clustering Segmentation in human vision K-mean clustering Mean-shift Graph-cut Reading: Chapters 14 [FP] Some slides of this lectures are courtesy of prof F. Li,
Lecture 8 February 4
ICS273A: Machine Learning Winter 2008 Lecture 8 February 4 Scribe: Carlos Agell (Student) Lecturer: Deva Ramanan 8.1 Neural Nets 8.1.1 Logistic Regression Recall the logistic function: g(x) = 1 1 + e θt
Part II Redundant Dictionaries and Pursuit Algorithms
Aisenstadt Chair Course CRM September 2009 Part II Redundant Dictionaries and Pursuit Algorithms Stéphane Mallat Centre de Mathématiques Appliquées Ecole Polytechnique Sparsity in Redundant Dictionaries
Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm
1 Enhancing the SNR of the Fiber Optic Rotation Sensor using the LMS Algorithm Hani Mehrpouyan, Student Member, IEEE, Department of Electrical and Computer Engineering Queen s University, Kingston, Ontario,
2.2 Creaseness operator
2.2. Creaseness operator 31 2.2 Creaseness operator Antonio López, a member of our group, has studied for his PhD dissertation the differential operators described in this section [72]. He has compared
PLAANN as a Classification Tool for Customer Intelligence in Banking
PLAANN as a Classification Tool for Customer Intelligence in Banking EUNITE World Competition in domain of Intelligent Technologies The Research Report Ireneusz Czarnowski and Piotr Jedrzejowicz Department
203.4770: Introduction to Machine Learning Dr. Rita Osadchy
203.4770: Introduction to Machine Learning Dr. Rita Osadchy 1 Outline 1. About the Course 2. What is Machine Learning? 3. Types of problems and Situations 4. ML Example 2 About the course Course Homepage:
EECS 556 Image Processing W 09. Interpolation. Interpolation techniques B splines
EECS 556 Image Processing W 09 Interpolation Interpolation techniques B splines What is image processing? Image processing is the application of 2D signal processing methods to images Image representation
Vision based Vehicle Tracking using a high angle camera
Vision based Vehicle Tracking using a high angle camera Raúl Ignacio Ramos García Dule Shu [email protected] [email protected] Abstract A vehicle tracking and grouping algorithm is presented in this work
Bernice E. Rogowitz and Holly E. Rushmeier IBM TJ Watson Research Center, P.O. Box 704, Yorktown Heights, NY USA
Are Image Quality Metrics Adequate to Evaluate the Quality of Geometric Objects? Bernice E. Rogowitz and Holly E. Rushmeier IBM TJ Watson Research Center, P.O. Box 704, Yorktown Heights, NY USA ABSTRACT
Forecasting Trade Direction and Size of Future Contracts Using Deep Belief Network
Forecasting Trade Direction and Size of Future Contracts Using Deep Belief Network Anthony Lai (aslai), MK Li (lilemon), Foon Wang Pong (ppong) Abstract Algorithmic trading, high frequency trading (HFT)
Super-resolution method based on edge feature for high resolution imaging
Science Journal of Circuits, Systems and Signal Processing 2014; 3(6-1): 24-29 Published online December 26, 2014 (http://www.sciencepublishinggroup.com/j/cssp) doi: 10.11648/j.cssp.s.2014030601.14 ISSN:
How To Fix Out Of Focus And Blur Images With A Dynamic Template Matching Algorithm
IJSTE - International Journal of Science Technology & Engineering Volume 1 Issue 10 April 2015 ISSN (online): 2349-784X Image Estimation Algorithm for Out of Focus and Blur Images to Retrieve the Barcode
Assessment. Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall
Automatic Photo Quality Assessment Presenter: Yupu Zhang, Guoliang Jin, Tuo Wang Computer Vision 2008 Fall Estimating i the photorealism of images: Distinguishing i i paintings from photographs h Florin
