Sparse Prediction with the ksupport Norm


 Kathryn Melton
 2 years ago
 Views:
Transcription
1 Sparse Prediction with the Support Norm Andreas Argyriou École Centrale Paris Rina Foygel Department of Statistics, Stanford University Nathan Srebro Toyota Technological Institute at Chicago Abstract We derive a novel norm that corresponds to the tightest convex relaxation of sparsity combined with an l penalty. We show that this new support norm provides a tighter relaxation than the elastic net and can thus be advantageous in in sparse prediction problems. We also bound the looseness of the elastic net, thus shedding new light on it and providing justification for its use. Introduction Regularizing with the l norm, when we expect a sparse solution to a regression problem, is often justified by w being the convex envelope of w 0 (the number of nonzero coordinates of a vector w R d ). That is, w is the tightest convex lower bound on w 0. But we must be careful with this statement for sparse vectors with large entries, w 0 can be small while w is large. In order to discuss convex lower bounds on w 0, we must impose some scale constraint. A more accurate statement is that w w w 0, and so, when the magnitudes of entries in w are bounded by, then w w 0, and indeed it is the largest such convex lower bound. Viewed as a convex outer relaxation, S ( ) := { w w 0, w } { w w }. Intersecting the righthandside with the l unit ball, we get the tightest convex outer bound (convex hull) of S ( ) : { w w, w } = conv(s ( ) ). However, in our view, this relationship between w and w 0 yields disappointing learning guarantees, and does not appropriately capture the success of the l norm as a surrogate for sparsity. In particular, the sample complexity of learning a linear predictor with nonzero entries by empirical ris minimization inside this class (an NPhard optimization problem) scales as O( log d), but relaxing to the constraint w yields a sample complexity which scales as O( log d), because the sample complexity of l regularized learning scales quadratically with the l norm [, 0]. Perhaps a better reason for the l norm being a good surrogate for sparsity is that, not only do we expect the magnitude of each entry of w to be bounded, but we further expect w to be small. In a regression setting, with a vector of features x, this can be justified when E[(x w) ] is bounded (a reasonable assumption) and the features are not too correlated see, e.g. [5]. More broadly, We define this as the number of observations needed in order to ensure expected prediction error no more than ɛ worse than that of the best sparse predictor, for an arbitrary constant ɛ (that is, we suppress the dependence on ɛ and focus on the dependence on the sparsity and dimensionality d).
2 especially in the presence of correlations, we might require this as a modeling assumption to aid in robustness and generalization. In any case, we have w w w 0, and so if we are interested in predictors with bounded l norm, we can motivate the l norm through the following relaxation of sparsity, where the scale is now set by the l norm: { w w 0, w B } { w w B }. The sample complexity when using the relaxation now scales as O( log d). Sparse + l constraint. Our starting point is then that of combining sparsity and l regularization, and learning a sparse predictor with small l norm. We are thus interested in classes of the form S () := { w w 0, w }. As discussed above, the class { w } (corresponding to the standard Lasso) provides a convex relaxation of S (). But clearly we can get a tighter relaxation by eeping the l constraint: conv(s () {w ) w } {, w w w }. () Constraining (or equivalently, penalizing) both the l and l norms, as in (), is nown as the elastic net [5, ] and has indeed been advocated as a better alternative to the Lasso. In this paper, we as whether the elastic net is the tightest convex relaxation to sparsity plus l (that is, to S () ) or whether a tighter, and better, convex relaxation is possible. A new norm. We consider the convex hull (tightest convex outer bound) of S (), C := conv(s () ) = conv { w w 0, w }. () We study the gauge function associated with this convex set, that is, the norm whose unit ball is given by (), which we call the support norm. We show that, for >, this is indeed a tighter convex relaxation than the elastic net (that is, both inequalities in () are in fact strict inequalities), and is therefore a better convex constraint than the elastic net when seeing a sparse, low l norm linear predictor. We thus advocate using it as a replacement for the elastic net. However, we also show that the gap between the elastic net and the support norm is at most a factor of, corresponding to a factor of two difference in the sample complexity. Thus, our wor can also be interpreted as justifying the use of the elastic net, viewing it as a fairly good approximation to the tightest possible convex relaxation of sparsity intersected with an l constraint. Still, even a factor of two should not necessarily be ignored and, as we show in our experiments, using the tighter support norm can indeed be beneficial. To better understand the support norm, we show in Section that it can also be described as the group lasso with overlaps norm [0] corresponding to all ( d ) subsets of features. Despite the exponential number of groups in this description, we show that the support norm can be calculated efficiently in time O(d log d) and that its dual is given simply by the l norm of the largest entries. We also provide efficient firstorder optimization algorithms for learning with the support norm. Related Wor In many learning problems of interest, Lasso has been observed to shrin too many of the variables of w to zero. In particular, in many applications, when a group of variables is highly correlated, the Lasso may prefer a sparse solution, but we might gain more predictive accuracy by including all the correlated variables in our model. These drawbacs have recently motivated the use of various other regularization methods, such as the elastic net [], which penalizes the regression coefficients w with a combination of l and l norms: min Xw y + λ w + λ w : w R d, (3) More precisely, the sample complexity is O(B log d), where the dependence on B is to be expected. Note that if feature vectors are l bounded (i.e. individual features are bounded), the sample complexity when using only w B (without a sparsity or l constraint) scales as O(B d). That is, even after identifying the correct support, we still need a sample complexity that scales with B.
3 where for a sample of size n, y R n is the vector of response values, and X R n d is a matrix with column j containing the values of feature j. The elastic net can be viewed as a tradeoff between l regularization (the Lasso) and l regularization (Ridge regression [9]), depending on the relative values of λ and λ. In particular, when λ = 0, (3) is equivalent to the Lasso. This method, and the other methods discussed below, have been observed to significantly outperform Lasso in many real applications. The pairwise elastic net (PEN) [3] is a penalty function that accounts for similarity among features: w P EN R = w + w w R w, where R [0, ] p p is a matrix with R j measuring similarity between features X j and X. The trace Lasso [6] is a second method proposed to handle correlations within X, defined by w trace X = Xdiag(w), where denotes the matrix tracenorm (the sum of the singular values) and promotes a lowran solution. If the features are orthogonal, then both the PEN and the Trace Lasso are equivalent to the Lasso. If the features are all identical, then both penalties are equivalent to Ridge regression (penalizing w ). Another existing penalty is OSCAR [3], given by w OSCAR c = w + c j< max{ w j, w }. Lie the elastic net, each one of these three methods also prefers averaging similar features over selecting a single feature. The Support Norm One argument for the elastic net has been the flexibility of tuning the cardinality of the regression vector w. Thus, when groups of correlated variables are present, a larger may be learned, which corresponds to a higher λ in (3). A more natural way to obtain such an effect of tuning the cardinality is to consider the convex hull of cardinality vectors, C = conv(s () ) = conv{w Rd w 0, w }. Clearly the sets C are nested, and C and C d are the unit balls for the l and l norms, respectively. Consequently we define the support norm as the norm whose unit ball equals C (the gauge function associated with the C ball). 3 An equivalent definition is the following variational formula: Definition.. Let {,..., d}. The support norm sp is defined, for every w Rd, as := min, w sp I G v I : supp(v I ) I, I G v I = w where G denotes the set of all subsets of {,..., d} of cardinality at most. The equivalence is immediate by rewriting v I = µ I z I in the above definition, where µ I 0, z I C, I G, I G µ I =. In addition, this immediately implies that sp is indeed a norm. In fact, the support norm is equivalent to the norm used by the group lasso with overlaps [0], when the set of overlapping groups is chosen to be G (however, the group lasso has traditionally been used for applications with some specific nown group structure, unlie the case considered here). Although the variational definition. is not amenable to computation because of the exponential growth of the set of groups G, the support norm is computationally very tractable, with an O(d log d) algorithm described in Section.. As already mentioned, sp = and sp d =. The unit ball of this new norm in R 3 for = is depicted in Figure. We immediately notice several differences between this unit ball and the elastic net unit ball. For example, at points with cardinality and l norm equal to, the support norm is not differentiable, but unlie the l or elasticnet norm, it is differentiable at points with cardinality less than. Thus, the support norm is less biased towards sparse vectors than the elastic net and the l norm. 3 The gauge function γ C : R d R {+ } is defined as γ C (x) = inf{λ R + : x λc }. 3
4 . The Dual Norm Figure : Unit ball of the support norm (left) and of the elastic net (right) on R 3. It is interesting and useful to compute the dual of the support norm. For w R d, denote w for the vector of absolute values, and w i for the ith largest element of w []. We have ( ) ( ) u sp = max { w, u : w sp } = max u i : I G = ( u i ) =: u () (). i I i= This is the l norm of the largest entries in u, and is nown as the  symmetric gauge norm []. Not surprisingly, this dual norm interpolates between the l norm (when = d and all entries are taen) and the l norm (when = and only the largest entry is taen). This parallels the interpolation of the support norm between the l and l norms.. Computation of the Norm In this section, we derive an alternative formula for the support norm, which leads to computation of the value of the norm in O(d log d) steps. ( ) Proposition.. For every w R d, w sp = r d ( w i ) + w i, i= r+ i= r where, letting w 0 denote +, r is the unique integer in {0,..., } satisfying w r > d w i w r r +. (4) i= r This result shows that sp trades off between the l and l norms in a way that favors sparse vectors but allows for cardinality larger than. It combines the uniform shrinage of an l penalty for the largest components, with the sparse shrinage of an l penalty for the smallest components. Proof of Proposition.. We will use the inequality w, u w, u [7]. We have { ( w sp ) = max u, w } { d ( u () () ) : u R d = max α i w i αi : i= i= } { } d α α d 0 = max α i w i + α w i αi : α α 0. Let A r := d i= r i= i= w i for r {0,..., }. If A 0 < w then the solution α is given by α i = w i for i =,..., ( ), α i = A 0 for i =,..., d. If A 0 w then the optimal α, α lie between w and A 0, and have to be equal. So, the maximization becomes { } max α i w i αi + A α α : α α 0. i= i= 4 i=
5 If A 0 w and w > A then the solution is α i = w i for i =,..., ( ), α i = A for i = ( ),..., d. Otherwise we proceed as before and continue this process. At stage r the process terminates if A 0 w,..., Ar r w r, A r r+ < w r and all but the last two inequalities are redundant. Hence the condition can be rewritten as (4). One optimal solution is α i = w i for i =,..., r, α i = Ar r+ for i = r,..., d. This proves the claim..3 Learning with the support norm We thus propose using learning rules with support norm regularization. These are appropriate when we would lie to learn a sparse predictor that also has low l norm, and are especially relevant when features might be correlated (that is, in almost all learning tass) but the correlation structure is not nown in advance. E.g., for squared error regression problems we have: { min Xw y + λ } ( w sp ) : w R d (5) with λ > 0 a regularization parameter and {,..., d} also a parameter to be tuned. As typical in regularizationbased methods, both λ and can be selected by cross validation [8]. Despite the relationship to S (), the parameter does not necessarily correspond to the sparsity of the actual minimizer of (5), and should be chosen via crossvalidation rather than set to the desired sparsity. 3 Relation to the Elastic Net Recall that the elastic net with penalty parameters λ and λ selects a vector of coefficients given by arg min Xw y + λ w + λ w. (6) For ease of comparison with the support norm, we first show that the set of optimal solutions for the elastic net, when the parameters are varied, is the same as for the norm { := max w, w / }, w el when [, d], corresponding to the unit ball in () (note that is not necessarily an integer). To see this, let ŵ be a solution to (6), and let := ( ŵ / ŵ ) [, d]. Now for any w ŵ, if w el ŵ el, then w p ŵ p for p =,. Since ŵ is a solution to (6), therefore, Xw y Xŵ y. This proves that, for some constraint parameter B, ŵ = arg min n Xw y : w el B. Lie the support norm, the elastic net interpolates between the l and l norms. In fact, when is an integer, any sparse unit vector w R d must lie in the unit ball of el. Since the support norm gives the convex hull of all sparse unit vectors, this immediately implies that w el w sp w R d. The two norms are not equal, however. The difference between the two is illustrated in Figure, where we see that the support norm is more rounded. To see an example where the two norms are not equal, we set d = + for some large, and let w = (.5,,,..., ) R d. Then { 3 = max +,.5 + } ( =.5 + ). w el Taing u = (,,,..., ), we have u () () <, and recalling this norm is dual to the support norm: w sp > w, u =.5 + =.5. In this example, we see that the two norms can differ by as much as a factor of. We now show that this is actually the most by which they can differ. 5
6 Proposition 3.. el sp < el. Proof. We show that these bounds hold in the duals of the two norms. First, since el maximum over the l and l norms, its dual is given by u (el) Now tae any u R d. First we show u () () u u d 0. For any a R d, { := inf a + } u a a R d is a u (el). Without loss of generality, we tae u () () = u : a : + u : a : a + u a. Finally, we show that u (el) < u () (). Let a = (u u +,..., u u +, 0,..., 0). Then u (el) a + u a = (u i u + ) + u + i= (u i u + ) + u + (u i u + ) + u + = u () i= Furthermore, this yields a strict inequality, because if u > u +, the nexttolast inequality is strict, while if u = = u +, then the last inequality is strict. 4 Optimization Solving the optimization problem (5) efficiently can be done with a firstorder proximal algorithm. Proximal methods see [, 4, 4, 8, 9] and references therein are used to solve composite problems of the form min{f(x) + ω(x) : x R d }, where the loss function f(x) and the regularizer ω(x) are convex functions, and f is smooth with an LLipschitz gradient. These methods require fast computation of the gradient f and the proximity operator prox ω (x) := argmin u x + ω(u) : u R d. To obtain a proximal method for support regularization, it suffices to compute the proximity map of g = β ( sp ), for any β > 0 (in particular, for problem (5) β corresponds to L λ ). This computation can be done in O(d( + log d)) steps with Algorithm. Algorithm Computation of the proximity operator. Input v R d Output q = prox )(v) β ( sp Find r {0,..., }, l {,..., d} such that i= (). β+ z T r > r,l l +(β+)r+β+ β+ z r (7) z l > where z := v, z 0 := +, z d+ :=, T r,l := β β+ z i if i =,..., r T q i z i r,l l +(β+)r+β+ if i = r,..., l 0 if i = l +,..., d Reorder and change signs of q to conform with v T r,l l +(β+)r+β+ z l+ (8) l z i i= r 6
7 Figure : Solutions learned for the synthetic data. Left to right: support, Lasso and elastic net. Proof of Correctness of Algorithm. Since the supportnorm is sign and permutation invariant, prox g (v) has the same ordering and signs as v. Hence, without loss of generality, we may assume that v v d 0 and require that q q d 0, which follows from inequality (7) and the fact that z is ordered. Now, q = prox g (v) is equivalent to βz βq = βv βq ( sp ) (q). It suffices to show that, for w = q, βz βq is an optimal α in the proof of Proposition.. Indeed, A r corresponds to d q i = l ( ) T z i r,l l +(β+)r+β+ = T r,l (l +r+)t r,l β T l +(β+)r+β+ = (r + ) r,l l +(β+)r+β+ i= r i= r and (4) is equivalent to condition (7). For i r, we have βz i βq i = q i. For r i l, we have βz i βq i = r+ A r. For i l +, since q i = 0, we only need βz i βq i r+ A r, which is true by (8). We can now apply a standard accelerated proximal method, such as FISTA [], to (5), at each iteration using the gradient of the loss and performing a prox step using Algorithm. The FISTA guarantee ensures us that, with appropriate step sizes, after T such iterations, we have: Xw T y + λ ( w T sp ) ( Xw y + λ ( w sp ) ) + L w w (T + ). 5 Empirical Comparisons Our theoretical analysis indicates that the support norm and the elastic net differ by at most a factor of, corresponding to at most a factor of two difference in their sample complexities and generalization guarantees. We thus do not expect huge differences between their actual performances, but would still lie to see whether the tighter relaxation of the support norm does yield some gains. Synthetic Data For the first simulation we follow [, Sec. 5, example 4]. In this experimental protocol, the target (oracle) vector equals w = (3,..., 3, 0..., 0), with y = (w } { } { ) x + N (0, ). 5 5 The input data X were generated from a normal distribution such that components,..., 5 have the same random mean Z N (0, ), components 6,..., 0 have mean Z N (0, ) and components,..., 5 have mean Z 3 N (0, ). A total of 50 data sets were created in this way, each containing 50 training points, 50 validation points and 350 test points. The goal is to achieve good prediction performance on the test data. We compared the support norm with Lasso and the elastic net. We considered the ranges = {,..., d} for support norm regularization, λ = 0 i, i = { 5,..., 5}, for the regularization parameter of Lasso and support regularization and the same range for the λ, λ of the elastic net. For each method, the optimal set of parameters was selected based on mean squared error on the validation set. The error reported in Table 5 is the mean squared error with respect to the oracle w, namely MSE = (ŵ w ) V (ŵ w ), where V is the population covariance matrix of X test. To further illustrate the effect of the support norm, in Figure 5 we show the coefficients learned by each method, in absolute value. For each image, one row corresponds to the w learned for one of the 50 data sets. Whereas all three methods distinguish the 5 relevant variables, the elastic net result varies less within these variables. South African Heart Data This is a classification tas which has been used in [8]. There are 9 variables and 46 examples, and the response is presence/absence of coronary heart disease. We 7
8 Table : Mean squared errors and classification accuracy for the synthetic data (median over 50 repetition), SA heart data (median over 50 replications) and for the 0 newsgroups data set. (SE = standard error) Synthetic Heart Newsgroups Method MSE (SE) MSE (SE) Accuracy (SE) MSE Accuracy Lasso (0.0) 0.8 (0.005) 66.4 (0.53) Elastic net 0.74 (0.0) 0.8 (0.005) 66.4 (0.53) support 0.43 (0.0) 0.8 (0.005) 66.4 (0.53) normalized the data so that each predictor variable has zero mean and unit variance. We then split the data 50 times randomly into training, validation, and test sets of sizes 400, 30, and 3 respectively. For each method, parameters were selected using the validation data. In Tables 5, we report the MSE and accuracy of each method on the test data. We observe that all three methods have identical performance. 0 Newsgroups This is a binary classification version of 0 newsgroups created in [] which can be found in the LIBSVM data repository. 4 The positive class consists of the 0 groups with names of form sci.*, comp.*, or misc.forsale and the negative class consists of the other 0 groups. To reduce the number of features, we removed the words which appear in less than 3 documents. We randomly split the data into a training, a validation and a test set of sizes 4000,000 and 4996, respectively. We report MSE and accuracy on the test data in Table 5. We found that support regularization gave improved prediction accuracy over both other methods. 5 6 Summary We introduced the support norm as the tightest convex relaxation of sparsity plus l regularization, and showed that it is tighter than the elastic net by exactly a factor of. In our view, this sheds light on the elastic net as a close approximation to this tightest possible convex relaxation, and motivates using the support norm when a tighter relaxation is sought. This is also demonstrated in our empirical results. We note that the support norm has better prediction properties, but not necessarily better sparsityinducing properties, as evident from its more rounded unit ball. It is well understood that there is often a tradeoff between sparsity and good prediction, and that even if the population optimal predictor is sparse, a denser predictor often yields better predictive performance [3, 0, ]. For example, in the presence of correlated features, it is often beneficial to include several highly correlated features rather than a single representative feature. This is exactly the behavior encouraged by l norm regularization, and the elastic net is already nown to yield less sparse (but more predictive) solutions. The support norm goes a step further in this direction, often yielding solutions that are even less sparse (but more predictive) compared to the elastic net. Nevertheless, it is interesting to consider whether compressed sensing results, where l regularization is of course central, can be refined by using the support norm, which might be able to handle more correlation structure within the set of features. Acnowledgements The construction showing that the gap between the elastic net and the  overlap norm can be as large as is due to joint wor with Ohad Shamir. Rina Foygel was supported by NSF grant DMS References [] A. Bec and M. Teboulle. A fast iterative shrinagethresholding algorithm for linear inverse problems. SIAM Journal of Imaging Sciences, ():83 0, 009. [] R. Bhatia. Matrix Analysis. Graduate Texts in Mathematics. Springer, cjlin/libsvmtools/datasets/ 5 Regarding other sparse prediction methods, we did not manage to compare with OSCAR, due to memory limitations, or to PEN or trace Lasso, which do not have code available online. 8
9 [3] H.D. Bondell and B.J. Reich. Simultaneous regression shrinage, variable selection, and supervised clustering of predictors with OSCAR. Biometrics, 64():5 3, 008. [4] P.L. Combettes and V.R. Wajs. Signal recovery by proximal forwardbacward splitting. Multiscale Modeling and Simulation, 4(4):68 00, 006. [5] C. De Mol, E. De Vito, and L. Rosasco. Elasticnet regularization in learning theory. Journal of Complexity, 5():0 30, 009. [6] E. Grave, G. R. Obozinsi, and F. Bach. Trace lasso: a trace norm regularization for correlated designs. In J. ShaweTaylor, R.S. Zemel, P. Bartlett, F.C.N. Pereira, and K.Q. Weinberger, editors, Advances in Neural Information Processing Systems 4, 0. [7] G. H. Hardy, J. E. Littlewood, and G. Pólya. Inequalities. Cambridge University Press, 934. [8] T. Hastie, R. Tibshirani, and J. Friedman. The Elements of Statistical Learning: Data Mining, Inference and Prediction. Springer Verlag Series in Statistics, 00. [9] A.E. Hoerl and R.W. Kennard. Ridge regression: Biased estimation for nonorthogonal problems. Technometrics, pages 55 67, 970. [0] L. Jacob, G. Obozinsi, and J.P. Vert. Group Lasso with overlap and graph Lasso. In Proceedings of the 6th Annual International Conference on Machine Learning, pages ACM, 009. [] S.M. Kaade, K. Sridharan, and A. Tewari. On the complexity of linear prediction: Ris bounds, margin bounds, and regularization. In Advances in Neural Information Processing Systems, volume, 008. [] S. S. Keerthi and D. DeCoste. A modified finite Newton method for fast solution of large scale linear SVMs. Journal of Machine Learning Research, 6:34 36, 005. [3] A. Lorbert, D. Eis, V. Kostina, D.M. Blei, and P.J. Ramadge. Exploiting covariate similarity in sparse regression via the pairwise elastic net. In Proceedings of the 3th International Conference on Artificial Intelligence and Statistics, 00. [4] Y. Nesterov. Gradient methods for minimizing composite objective function. CORE, 007. [5] N. Srebro, K. Sridharan, and A. Tewari. Smoothness, lownoise and fast rates. In Advances in Neural Information Processing Systems 3, 00. [6] T. Suzui and R. Tomioa. SpicyMKL: a fast algorithm for multiple ernel learning with thousands of ernels. Machine learning, pages 3, 0. [7] R. Tibshirani. Regression shrinage and selection via the lasso. Journal of the Royal Statistical Society, Series B (Statistical Methodology), 58():67 88, 996. [8] P. Tseng. On accelerated proximal gradient methods for convexconcave optimization. Preprint, 008. [9] P. Tseng. Approximation accuracy, gradient methods, and error bound for structured convex optimization. Mathematical Programming, 5():63 95, 00. [0] T. Zhang. Covering number bounds of certain regularized linear function classes. The Journal of Machine Learning Research, :57 550, 00. [] H. Zou and T. Hastie. Regularization and variable selection via the elastic net. Journal of the Royal Statistical Society: Series B (Statistical Methodology), 67():30 30,
Adaptive Online Gradient Descent
Adaptive Online Gradient Descent Peter L Bartlett Division of Computer Science Department of Statistics UC Berkeley Berkeley, CA 94709 bartlett@csberkeleyedu Elad Hazan IBM Almaden Research Center 650
More informationLeast Squares Estimation
Least Squares Estimation SARA A VAN DE GEER Volume 2, pp 1041 1045 in Encyclopedia of Statistics in Behavioral Science ISBN13: 9780470860809 ISBN10: 0470860804 Editors Brian S Everitt & David
More information10. Proximal point method
L. Vandenberghe EE236C Spring 201314) 10. Proximal point method proximal point method augmented Lagrangian method MoreauYosida smoothing 101 Proximal point method a conceptual algorithm for minimizing
More informationChapter 15 Introduction to Linear Programming
Chapter 15 Introduction to Linear Programming An Introduction to Optimization Spring, 2014 WeiTa Chu 1 Brief History of Linear Programming The goal of linear programming is to determine the values of
More informationDoptimal plans in observational studies
Doptimal plans in observational studies Constanze Pumplün Stefan Rüping Katharina Morik Claus Weihs October 11, 2005 Abstract This paper investigates the use of Design of Experiments in observational
More informationLasso on Categorical Data
Lasso on Categorical Data Yunjin Choi, Rina Park, Michael Seo December 14, 2012 1 Introduction In social science studies, the variables of interest are often categorical, such as race, gender, and nationality.
More informationData analysis in supersaturated designs
Statistics & Probability Letters 59 (2002) 35 44 Data analysis in supersaturated designs Runze Li a;b;, Dennis K.J. Lin a;b a Department of Statistics, The Pennsylvania State University, University Park,
More informationRegression Using Support Vector Machines: Basic Foundations
Regression Using Support Vector Machines: Basic Foundations Technical Report December 2004 Aly Farag and Refaat M Mohamed Computer Vision and Image Processing Laboratory Electrical and Computer Engineering
More informationGI01/M055 Supervised Learning Proximal Methods
GI01/M055 Supervised Learning Proximal Methods Massimiliano Pontil (based on notes by Luca Baldassarre) (UCL) Proximal Methods 1 / 20 Today s Plan Problem setting Convex analysis concepts Proximal operators
More informationBig Data  Lecture 1 Optimization reminders
Big Data  Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Big Data  Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Schedule Introduction Major issues Examples Mathematics
More informationDegrees of Freedom and Model Search
Degrees of Freedom and Model Search Ryan J. Tibshirani Abstract Degrees of freedom is a fundamental concept in statistical modeling, as it provides a quantitative description of the amount of fitting performed
More informationThe fastclime Package for Linear Programming and LargeScale Precision Matrix Estimation in R
Journal of Machine Learning Research 15 (2014) 489493 Submitted 3/13; Revised 8/13; Published 2/14 The fastclime Package for Linear Programming and LargeScale Precision Matrix Estimation in R Haotian
More informationRegularized Logistic Regression for Mind Reading with Parallel Validation
Regularized Logistic Regression for Mind Reading with Parallel Validation Heikki Huttunen, JukkaPekka Kauppi, Jussi Tohka Tampere University of Technology Department of Signal Processing Tampere, Finland
More informationCHARACTERISTICS IN FLIGHT DATA ESTIMATION WITH LOGISTIC REGRESSION AND SUPPORT VECTOR MACHINES
CHARACTERISTICS IN FLIGHT DATA ESTIMATION WITH LOGISTIC REGRESSION AND SUPPORT VECTOR MACHINES Claus Gwiggner, Ecole Polytechnique, LIX, Palaiseau, France Gert Lanckriet, University of Berkeley, EECS,
More informationStatistical machine learning, high dimension and big data
Statistical machine learning, high dimension and big data S. Gaïffas 1 14 mars 2014 1 CMAP  Ecole Polytechnique Agenda for today Divide and Conquer principle for collaborative filtering Graphical modelling,
More informationComponent Ordering in Independent Component Analysis Based on Data Power
Component Ordering in Independent Component Analysis Based on Data Power Anne Hendrikse Raymond Veldhuis University of Twente University of Twente Fac. EEMCS, Signals and Systems Group Fac. EEMCS, Signals
More informationMultiple Linear Regression in Data Mining
Multiple Linear Regression in Data Mining Contents 2.1. A Review of Multiple Linear Regression 2.2. Illustration of the Regression Process 2.3. Subset Selection in Linear Regression 1 2 Chap. 2 Multiple
More informationStatistical Machine Learning
Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 4: classical linear and quadratic discriminants. 1 / 25 Linear separation For two classes in R d : simple idea: separate the classes
More informationThe primary goal of this thesis was to understand how the spatial dependence of
5 General discussion 5.1 Introduction The primary goal of this thesis was to understand how the spatial dependence of consumer attitudes can be modeled, what additional benefits the recovering of spatial
More informationBig Data Analytics: Optimization and Randomization
Big Data Analytics: Optimization and Randomization Tianbao Yang, Qihang Lin, Rong Jin Tutorial@SIGKDD 2015 Sydney, Australia Department of Computer Science, The University of Iowa, IA, USA Department of
More informationLecture 3: Linear methods for classification
Lecture 3: Linear methods for classification Rafael A. Irizarry and Hector Corrada Bravo February, 2010 Today we describe four specific algorithms useful for classification problems: linear regression,
More informationDefinition of a Linear Program
Definition of a Linear Program Definition: A function f(x 1, x,..., x n ) of x 1, x,..., x n is a linear function if and only if for some set of constants c 1, c,..., c n, f(x 1, x,..., x n ) = c 1 x 1
More information3. Proximal gradient method
Algorithms for largescale convex optimization DTU 2010 3. Proximal gradient method introduction proximal mapping proximal gradient method convergence analysis accelerated proximal gradient method forwardbackward
More informationOverview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model
Overview of Violations of the Basic Assumptions in the Classical Normal Linear Regression Model 1 September 004 A. Introduction and assumptions The classical normal linear regression model can be written
More informationPATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION
PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 4: LINEAR MODELS FOR CLASSIFICATION Introduction In the previous chapter, we explored a class of regression models having particularly simple analytical
More informationModern regression 2: The lasso
Modern regression 2: The lasso Ryan Tibshirani Data Mining: 36462/36662 March 21 2013 Optional reading: ISL 6.2.2, ESL 3.4.2, 3.4.3 1 Reminder: ridge regression and variable selection Recall our setup:
More informationProximal mapping via network optimization
L. Vandenberghe EE236C (Spring 234) Proximal mapping via network optimization minimum cut and maximum flow problems parametric minimum cut problem application to proximal mapping Introduction this lecture:
More informationPredicting Health Care Costs by Twopart Model with Sparse Regularization
Predicting Health Care Costs by Twopart Model with Sparse Regularization Atsuyuki Kogure Keio University, Japan July, 2015 Abstract We consider the problem of predicting health care costs using the twopart
More informationCORRELATED TO THE SOUTH CAROLINA COLLEGE AND CAREERREADY FOUNDATIONS IN ALGEBRA
We Can Early Learning Curriculum PreK Grades 8 12 INSIDE ALGEBRA, GRADES 8 12 CORRELATED TO THE SOUTH CAROLINA COLLEGE AND CAREERREADY FOUNDATIONS IN ALGEBRA April 2016 www.voyagersopris.com Mathematical
More informationBag of Pursuits and Neural Gas for Improved Sparse Coding
Bag of Pursuits and Neural Gas for Improved Sparse Coding Kai Labusch, Erhardt Barth, and Thomas Martinetz University of Lübec Institute for Neuro and Bioinformatics Ratzeburger Allee 6 23562 Lübec, Germany
More informationSolving Sets of Equations. 150 B.C.E., 九章算術 Carl Friedrich Gauss,
Solving Sets of Equations 5 B.C.E., 九章算術 Carl Friedrich Gauss, 777855 GaussianJordan Elimination In GaussJordan elimination, matrix is reduced to diagonal rather than triangular form Row combinations
More informationChapter 6. Cuboids. and. vol(conv(p ))
Chapter 6 Cuboids We have already seen that we can efficiently find the bounding box Q(P ) and an arbitrarily good approximation to the smallest enclosing ball B(P ) of a set P R d. Unfortunately, both
More informationME128 ComputerAided Mechanical Design Course Notes Introduction to Design Optimization
ME128 Computerided Mechanical Design Course Notes Introduction to Design Optimization 2. OPTIMIZTION Design optimization is rooted as a basic problem for design engineers. It is, of course, a rare situation
More informationCHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES. From Exploratory Factor Analysis Ledyard R Tucker and Robert C.
CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES From Exploratory Factor Analysis Ledyard R Tucker and Robert C MacCallum 1997 180 CHAPTER 8 FACTOR EXTRACTION BY MATRIX FACTORING TECHNIQUES In
More informationSeveral Views of Support Vector Machines
Several Views of Support Vector Machines Ryan M. Rifkin Honda Research Institute USA, Inc. Human Intention Understanding Group 2007 Tikhonov Regularization We are considering algorithms of the form min
More informationAbsolute Value Programming
Computational Optimization and Aplications,, 1 11 (2006) c 2006 Springer Verlag, Boston. Manufactured in The Netherlands. Absolute Value Programming O. L. MANGASARIAN olvi@cs.wisc.edu Computer Sciences
More informationMarketing Mix Modelling and Big Data P. M Cain
1) Introduction Marketing Mix Modelling and Big Data P. M Cain Big data is generally defined in terms of the volume and variety of structured and unstructured information. Whereas structured data is stored
More informationModern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
More informationThe pnorm generalization of the LMS algorithm for adaptive filtering
The pnorm generalization of the LMS algorithm for adaptive filtering Jyrki Kivinen University of Helsinki Manfred Warmuth University of California, Santa Cruz Babak Hassibi California Institute of Technology
More informationHow to assess the risk of a large portfolio? How to estimate a large covariance matrix?
Chapter 3 Sparse Portfolio Allocation This chapter touches some practical aspects of portfolio allocation and risk assessment from a large pool of financial assets (e.g. stocks) How to assess the risk
More informationOnline Learning, Stability, and Stochastic Gradient Descent
Online Learning, Stability, and Stochastic Gradient Descent arxiv:1105.4701v3 [cs.lg] 8 Sep 2011 September 9, 2011 Tomaso Poggio, Stephen Voinea, Lorenzo Rosasco CBCL, McGovern Institute, CSAIL, Brain
More informationApplications to Data Smoothing and Image Processing I
Applications to Data Smoothing and Image Processing I MA 348 Kurt Bryan Signals and Images Let t denote time and consider a signal a(t) on some time interval, say t. We ll assume that the signal a(t) is
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 6 Three Approaches to Classification Construct
More informationQuiz 1 Sample Questions IE406 Introduction to Mathematical Programming Dr. Ralphs
Quiz 1 Sample Questions IE406 Introduction to Mathematical Programming Dr. Ralphs These questions are from previous years and should you give you some idea of what to expect on Quiz 1. 1. Consider the
More informationEffective Linear Discriminant Analysis for High Dimensional, Low Sample Size Data
Effective Linear Discriant Analysis for High Dimensional, Low Sample Size Data Zhihua Qiao, Lan Zhou and Jianhua Z. Huang Abstract In the socalled high dimensional, low sample size (HDLSS) settings, LDA
More informationIncreasing for all. Convex for all. ( ) Increasing for all (remember that the log function is only defined for ). ( ) Concave for all.
1. Differentiation The first derivative of a function measures by how much changes in reaction to an infinitesimal shift in its argument. The largest the derivative (in absolute value), the faster is evolving.
More informationModel selection in R featuring the lasso. Chris Franck LISA Short Course March 26, 2013
Model selection in R featuring the lasso Chris Franck LISA Short Course March 26, 2013 Goals Overview of LISA Classic data example: prostate data (Stamey et. al) Brief review of regression and model selection.
More informationOn the ksupport and Related Norms
On the ksupport and Related Norms Massimiliano Pontil Department of Computer Science Centre for Computational Statistics and Machine Learning University College London (Joint work with Andrew McDonald
More informationAuxiliary Variables in Mixture Modeling: 3Step Approaches Using Mplus
Auxiliary Variables in Mixture Modeling: 3Step Approaches Using Mplus Tihomir Asparouhov and Bengt Muthén Mplus Web Notes: No. 15 Version 8, August 5, 2014 1 Abstract This paper discusses alternatives
More informationFederated Optimization: Distributed Optimization Beyond the Datacenter
Federated Optimization: Distributed Optimization Beyond the Datacenter Jakub Konečný School of Mathematics University of Edinburgh J.Konecny@sms.ed.ac.uk H. Brendan McMahan Google, Inc. Seattle, WA 98103
More informationMachine Learning Big Data using Map Reduce
Machine Learning Big Data using Map Reduce By Michael Bowles, PhD Where Does Big Data Come From? Web data (web logs, click histories) ecommerce applications (purchase histories) Retail purchase histories
More informationBy W.E. Diewert. July, Linear programming problems are important for a number of reasons:
APPLIED ECONOMICS By W.E. Diewert. July, 3. Chapter : Linear Programming. Introduction The theory of linear programming provides a good introduction to the study of constrained maximization (and minimization)
More information2.3 Convex Constrained Optimization Problems
42 CHAPTER 2. FUNDAMENTAL CONCEPTS IN CONVEX OPTIMIZATION Theorem 15 Let f : R n R and h : R R. Consider g(x) = h(f(x)) for all x R n. The function g is convex if either of the following two conditions
More informationANALYSIS, THEORY AND DESIGN OF LOGISTIC REGRESSION CLASSIFIERS USED FOR VERY LARGE SCALE DATA MINING
ANALYSIS, THEORY AND DESIGN OF LOGISTIC REGRESSION CLASSIFIERS USED FOR VERY LARGE SCALE DATA MINING BY OMID ROUHANIKALLEH THESIS Submitted as partial fulfillment of the requirements for the degree of
More informationCheng Soon Ong & Christfried Webers. Canberra February June 2016
c Cheng Soon Ong & Christfried Webers Research Group and College of Engineering and Computer Science Canberra February June (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 31 c Part I
More informationNonlinear Optimization: Algorithms 3: Interiorpoint methods
Nonlinear Optimization: Algorithms 3: Interiorpoint methods INSEAD, Spring 2006 JeanPhilippe Vert Ecole des Mines de Paris JeanPhilippe.Vert@mines.org Nonlinear optimization c 2006 JeanPhilippe Vert,
More informationELECE8104 Stochastics models and estimation, Lecture 3b: Linear Estimation in Static Systems
Stochastics models and estimation, Lecture 3b: Linear Estimation in Static Systems Minimum Mean Square Error (MMSE) MMSE estimation of Gaussian random vectors Linear MMSE estimator for arbitrarily distributed
More informationPractical Guide to the Simplex Method of Linear Programming
Practical Guide to the Simplex Method of Linear Programming Marcel Oliver Revised: April, 0 The basic steps of the simplex algorithm Step : Write the linear programming problem in standard form Linear
More informationprinceton univ. F 13 cos 521: Advanced Algorithm Design Lecture 6: Provable Approximation via Linear Programming Lecturer: Sanjeev Arora
princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 6: Provable Approximation via Linear Programming Lecturer: Sanjeev Arora Scribe: One of the running themes in this course is the notion of
More informationFacebook Friend Suggestion Eytan Daniyalzade and Tim Lipus
Facebook Friend Suggestion Eytan Daniyalzade and Tim Lipus 1. Introduction Facebook is a social networking website with an open platform that enables developers to extract and utilize user information
More informationPenalized regression: Introduction
Penalized regression: Introduction Patrick Breheny August 30 Patrick Breheny BST 764: Applied Statistical Modeling 1/19 Maximum likelihood Much of 20thcentury statistics dealt with maximum likelihood
More informationCSE 494 CSE/CBS 598 (Fall 2007): Numerical Linear Algebra for Data Exploration Clustering Instructor: Jieping Ye
CSE 494 CSE/CBS 598 Fall 2007: Numerical Linear Algebra for Data Exploration Clustering Instructor: Jieping Ye 1 Introduction One important method for data compression and classification is to organize
More informationReducing multiclass to binary by coupling probability estimates
Reducing multiclass to inary y coupling proaility estimates Bianca Zadrozny Department of Computer Science and Engineering University of California, San Diego La Jolla, CA 920930114 zadrozny@cs.ucsd.edu
More informationDuality in General Programs. Ryan Tibshirani Convex Optimization 10725/36725
Duality in General Programs Ryan Tibshirani Convex Optimization 10725/36725 1 Last time: duality in linear programs Given c R n, A R m n, b R m, G R r n, h R r : min x R n c T x max u R m, v R r b T
More informationA Semiparametric Approach for Decomposition of Absorption Spectra in the Presence of Unknown Components
A Semiparametric Approach for Decomposition of Absorption Spectra in the Presence of Unknown Components Payman Sadegh 1,2, Henrik Aalborg Nielsen 1, and Henrik Madsen 1 Abstract Decomposition of absorption
More informationBootstrapping Big Data
Bootstrapping Big Data Ariel Kleiner Ameet Talwalkar Purnamrita Sarkar Michael I. Jordan Computer Science Division University of California, Berkeley {akleiner, ameet, psarkar, jordan}@eecs.berkeley.edu
More informationPenalized Logistic Regression and Classification of Microarray Data
Penalized Logistic Regression and Classification of Microarray Data Milan, May 2003 Anestis Antoniadis Laboratoire IMAGLMC University Joseph Fourier Grenoble, France Penalized Logistic Regression andclassification
More informationMoral Hazard. Itay Goldstein. Wharton School, University of Pennsylvania
Moral Hazard Itay Goldstein Wharton School, University of Pennsylvania 1 PrincipalAgent Problem Basic problem in corporate finance: separation of ownership and control: o The owners of the firm are typically
More informationVariational approach to restore pointlike and curvelike singularities in imaging
Variational approach to restore pointlike and curvelike singularities in imaging Daniele Graziani joint work with Gilles Aubert and Laure BlancFéraud Roma 12/06/2012 Daniele Graziani (Roma) 12/06/2012
More informationLoad Balancing and Switch Scheduling
EE384Y Project Final Report Load Balancing and Switch Scheduling Xiangheng Liu Department of Electrical Engineering Stanford University, Stanford CA 94305 Email: liuxh@systems.stanford.edu Abstract Load
More informationSMOOTHING APPROXIMATIONS FOR TWO CLASSES OF CONVEX EIGENVALUE OPTIMIZATION PROBLEMS YU QI. (B.Sc.(Hons.), BUAA)
SMOOTHING APPROXIMATIONS FOR TWO CLASSES OF CONVEX EIGENVALUE OPTIMIZATION PROBLEMS YU QI (B.Sc.(Hons.), BUAA) A THESIS SUBMITTED FOR THE DEGREE OF MASTER OF SCIENCE DEPARTMENT OF MATHEMATICS NATIONAL
More informationOverview of Math Standards
Algebra 2 Welcome to math curriculum design maps for Manhattan Ogden USD 383, striving to produce learners who are: Effective Communicators who clearly express ideas and effectively communicate with diverse
More informationAlgebra Unpacked Content For the new Common Core standards that will be effective in all North Carolina schools in the 201213 school year.
This document is designed to help North Carolina educators teach the Common Core (Standard Course of Study). NCDPI staff are continually updating and improving these tools to better serve teachers. Algebra
More informationOPRE 6201 : 2. Simplex Method
OPRE 6201 : 2. Simplex Method 1 The Graphical Method: An Example Consider the following linear program: Max 4x 1 +3x 2 Subject to: 2x 1 +3x 2 6 (1) 3x 1 +2x 2 3 (2) 2x 2 5 (3) 2x 1 +x 2 4 (4) x 1, x 2
More informationWeek 5 Integral Polyhedra
Week 5 Integral Polyhedra We have seen some examples 1 of linear programming formulation that are integral, meaning that every basic feasible solution is an integral vector. This week we develop a theory
More informationData Mining  Evaluation of Classifiers
Data Mining  Evaluation of Classifiers Lecturer: JERZY STEFANOWSKI Institute of Computing Sciences Poznan University of Technology Poznan, Poland Lecture 4 SE Master Course 2008/2009 revised for 2010
More informationInformation Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay
Information Theory and Coding Prof. S. N. Merchant Department of Electrical Engineering Indian Institute of Technology, Bombay Lecture  17 ShannonFanoElias Coding and Introduction to Arithmetic Coding
More informationExponential time algorithms for graph coloring
Exponential time algorithms for graph coloring Uriel Feige Lecture notes, March 14, 2011 1 Introduction Let [n] denote the set {1,..., k}. A klabeling of vertices of a graph G(V, E) is a function V [k].
More informationCrossValidation. Synonyms Rotation estimation
Comp. by: BVijayalakshmiGalleys0000875816 Date:6/11/08 Time:19:52:53 Stage:First Proof C PAYAM REFAEILZADEH, LEI TANG, HUAN LIU Arizona State University Synonyms Rotation estimation Definition is a statistical
More informationBindel, Fall 2012 Matrix Computations (CS 6210) Week 8: Friday, Oct 12
Why eigenvalues? Week 8: Friday, Oct 12 I spend a lot of time thinking about eigenvalue problems. In part, this is because I look for problems that can be solved via eigenvalues. But I might have fewer
More informationLINEAR PROGRAMMING P V Ram B. Sc., ACA, ACMA Hyderabad
LINEAR PROGRAMMING P V Ram B. Sc., ACA, ACMA 98481 85073 Hyderabad Page 1 of 19 Question: Explain LPP. Answer: Linear programming is a mathematical technique for determining the optimal allocation of resources
More informationA Simple Introduction to Support Vector Machines
A Simple Introduction to Support Vector Machines Martin Law Lecture for CSE 802 Department of Computer Science and Engineering Michigan State University Outline A brief history of SVM Largemargin linear
More informationBig Data Analytics. Lucas Rego Drumond
Big Data Analytics Lucas Rego Drumond Information Systems and Machine Learning Lab (ISMLL) Institute of Computer Science University of Hildesheim, Germany Going For Large Scale Going For Large Scale 1
More informationPartial Least Squares (PLS) Regression.
Partial Least Squares (PLS) Regression. Hervé Abdi 1 The University of Texas at Dallas Introduction Pls regression is a recent technique that generalizes and combines features from principal component
More informationINDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS
INDISTINGUISHABILITY OF ABSOLUTELY CONTINUOUS AND SINGULAR DISTRIBUTIONS STEVEN P. LALLEY AND ANDREW NOBEL Abstract. It is shown that there are no consistent decision rules for the hypothesis testing problem
More informationJoint models for classification and comparison of mortality in different countries.
Joint models for classification and comparison of mortality in different countries. Viani D. Biatat 1 and Iain D. Currie 1 1 Department of Actuarial Mathematics and Statistics, and the Maxwell Institute
More informationSupport Vector Machines Explained
March 1, 2009 Support Vector Machines Explained Tristan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introduction This document has been written in an attempt to make the Support Vector Machines (SVM),
More informationApplied Algorithm Design Lecture 5
Applied Algorithm Design Lecture 5 Pietro Michiardi Eurecom Pietro Michiardi (Eurecom) Applied Algorithm Design Lecture 5 1 / 86 Approximation Algorithms Pietro Michiardi (Eurecom) Applied Algorithm Design
More informationSupport Vector Machines with Clustering for Training with Very Large Datasets
Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France theodoros.evgeniou@insead.fr Massimiliano
More informationALMOST COMMON PRIORS 1. INTRODUCTION
ALMOST COMMON PRIORS ZIV HELLMAN ABSTRACT. What happens when priors are not common? We introduce a measure for how far a type space is from having a common prior, which we term prior distance. If a type
More informationPredict Influencers in the Social Network
Predict Influencers in the Social Network Ruishan Liu, Yang Zhao and Liuyu Zhou Email: rliu2, yzhao2, lyzhou@stanford.edu Department of Electrical Engineering, Stanford University Abstract Given two persons
More informationPredict the Popularity of YouTube Videos Using Early View Data
000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050
More informationarxiv:1112.0829v1 [math.pr] 5 Dec 2011
How Not to Win a Million Dollars: A Counterexample to a Conjecture of L. Breiman Thomas P. Hayes arxiv:1112.0829v1 [math.pr] 5 Dec 2011 Abstract Consider a gambling game in which we are allowed to repeatedly
More informationStudying Auto Insurance Data
Studying Auto Insurance Data Ashutosh Nandeshwar February 23, 2010 1 Introduction To study auto insurance data using traditional and nontraditional tools, I downloaded a wellstudied data from http://www.statsci.org/data/general/motorins.
More informationA Stochastic 3MG Algorithm with Application to 2D Filter Identification
A Stochastic 3MG Algorithm with Application to 2D Filter Identification Emilie Chouzenoux 1, JeanChristophe Pesquet 1, and Anisia Florescu 2 1 Laboratoire d Informatique Gaspard Monge  CNRS Univ. ParisEst,
More informationA network flow algorithm for reconstructing. binary images from discrete Xrays
A network flow algorithm for reconstructing binary images from discrete Xrays Kees Joost Batenburg Leiden University and CWI, The Netherlands kbatenbu@math.leidenuniv.nl Abstract We present a new algorithm
More informationClass #6: Nonlinear classification. ML4Bio 2012 February 17 th, 2012 Quaid Morris
Class #6: Nonlinear classification ML4Bio 2012 February 17 th, 2012 Quaid Morris 1 Module #: Title of Module 2 Review Overview Linear separability Nonlinear classification Linear Support Vector Machines
More informationOn the Degrees of Freedom of the Lasso
On the Degrees of Freedom of the Lasso Hui Zou Trevor Hastie Robert Tibshirani Abstract We study the degrees of freedom of the Lasso in the framewor of Stein s unbiased ris estimation (SURE). We show that
More informationNotes from Week 1: Algorithms for sequential prediction
CS 683 Learning, Games, and Electronic Markets Spring 2007 Notes from Week 1: Algorithms for sequential prediction Instructor: Robert Kleinberg 2226 Jan 2007 1 Introduction In this course we will be looking
More informationFactoring & Primality
Factoring & Primality Lecturer: Dimitris Papadopoulos In this lecture we will discuss the problem of integer factorization and primality testing, two problems that have been the focus of a great amount
More information