Solving Large Scale Optimization Problems via Grid and Cluster Computing
|
|
|
- Barry Park
- 10 years ago
- Views:
Transcription
1 Solving Large Scale Optimization Problems via Grid and Cluster Computing Katsuki Fujisawa, Masakazu Kojima Akiko Takeda and Makoto Yamashita Abstract. Solving large scale optimization problems requires a huge amount of computational power. The size of optimization problems that can be solved on a few CPUs has been limited due to a lack of computational power. Grid and cluster computing has received much attention as a powerful and inexpensive way of solving large scale optimization problems that an existing single CPU cannot process. The aim of this paper is to show that grid and cluster computing provides tremendous power to optimization methods. The methods that this article picks up are a successive convex relaxation method for quadratic optimization problems, a polyhedral homotopy method for polynomial systems of equations and a primal-dual interior-point method for semidefinite programs. Their parallel implementations on grids and clusters together with numerical results are reported. The article also mentions a grid portal system for optimization problems briefly. Key words. Optimization Problem, Wide Area Network, Parallel Computing, Cluster Computing, Grid Computing, Semidefinite Program, Semidefinite Program Relaxation, Polynomial Equations, Polyhedral Homotopy Method This paper is a revised version of [4], which is presented at SAINT 2004 (The 2004 International Symposium on Applications and the Internet), Tokyo, Japan, January 26 30, Department of Mathematical Sciences, Tokyo Denki University. Grid Technology Research Center, National Institute of Advanced Industrial Science and Technology, Japan. [email protected] Department of Mathematical and Computing Sciences, Tokyo Institute of Technology, Oh-Okayama, Meguro-ku, Tokyo Japan. [email protected] Department of Mathematical and Computing Sciences, Tokyo Institute of Technology, Oh-Okayama, Meguro-ku, Tokyo Japan. [email protected] Department of Mathematical and Computing Sciences, Tokyo Institute of Technology, Oh-Okayama, Meguro-ku, Tokyo Japan. [email protected]
2 1 Introduction Optimization problems have a variety of practical applications in various fields such as operations research, engineering, science, biology and economics. Many combinatorial and nonconvex optimization problems are known to be N P-hard, which means that there exists no deterministic algorithm that finds an optimal solution in polynomial time unless P = N P. Hence, solving large scale combinatorial and nonconvex optimization problems requires a huge amount of computational time and resources, and the size of such problems that we can solve has been limited. As computing resources continue to improve, however, optimal solutions of larger scale optimization problems become more achievable. In particular, grid and cluster computing technology has recently received much attention as a powerful and inexpensive methodology for solving large scale optimization problems that an existing single CPU cannot process. To solve large scale optimization problems, we need new computing infrastructure which enables us to easily access to computational resources including hardware and software library distributed across a wide area network like the Internet. For example, Applegate et al. [2] implemented the Danzig, Fulkerson and Johnson s cutting plane method for the large scale TSP (traveling salesman problem). They obtained an optimal solution of the TSP that has 15,112 cities (nodes) in Germany. This problem is the largest scale TSPLIB 1 instance that has been solved to date. The computation was executed on a network of 110 processors located at Rice and Princeton Universities. They estimated the total computational time was 22.6 years, scaled to a Compaq EV6(21264) Alpha processor running at 500MHz. A group of researchers at the University of Iowa and Argonne National Laboratory solved the QAP (quadratic assignment problem) instance called NUG30 in QAPLIB 2 using the Condor 3 developed by the University of Wisconsin. The Condor is a system of daemons and tools to utilize commodity computing resources including idle desktop machines for high-throughput computing. NUG30 is known as a huge scale QAP that would require more than 10 years of computational time by a single CPU. They obtained an optimal solution in just seven days using the Condor system and 653 CPUs on average. See the paper [1] for more details. In this paper, we present grid and cluster computing for some optimization problems with numerical results. In Sections 2 and 3, we focus our attention on GridRPC, which is an RPC system redesigned for grid applications. Some GridRPC systems such as Netsolve and Ninf are widely used. The Ninf (Network based Information library for global world-wide computing Infrastructure) system [11] developed by AIST (National Institute of Advanced Industrial Science and Technology, Japan) employs a client-server model, where server and client machines are connected via a local area network or the Internet. We implemented highly parallel algorithms for some optimization problems and polynomial systems of equations with the use of the Ninf system on several PC clusters connected via a high speed local area network and/or the Internet. In Section 4, we briefly explain the SDPARA (SemiDefinite Programming Algorithm parallel version) [16], a parallel implementation of the SDPA [15], and present numerical results on PC clusters. Section 5 presents a grid portal system which enables users with few computational resources to receive benefits of parallel optimization
3 Figure 1: Ninf client-server model 2 Parallel Successive Convex Relaxation Method for Nonconvex Quadratic Optimization Problems Let R n and S n denote the n-dimensional Euclidean space and the space of n n real symmetric matrices, respectively. A general QOP is described in the the following form: } minimize c T x subject to γ i + 2q T i x + x T, (1) Q i x 0 (i = 1,... m) where c R n, γ i R, q i R n and Q i S n (i = 1,..., m). When a given QOP has a quadratic objective function such as γ 0 + 2q T 0 x + x T Q 0 x, we can transform it into QOP (1) by replacing the quadratic objective function with a new variable t and adding γ 0 2q T 0 x x T Q 0 x + t 0 to the set of inequality constraints. A general QOP is known as one of the most significant nonlinear programming problems. QOPs cover not only economics and engineering applications but also various important nonconvex mathematical programs, such as 0-1 integer programs, linear complementarity problems, bilevel quadratic programs, linear and quadratic fractional programs, and so on. Because of their theoretical and computational difficulties, however, solvable QOPs had been limited to convex QOPs where all Q i s are assumed to be positive semidefinite or general QOPs with small size. The SCRM (Successive Convex Relaxation Method) proposed by Kojima-Tuncel [7] is a powerful numerical method to compute upper bounds of general QOPs by repeated applications of SDP (semidefinite programming) relaxations. The SCRM generates and solves a large number of SDP problems at each iteration. Takeda et al. [12] reported that the SCRM can deal with some larger scale QOPs through numerical experiments on a PC cluster. To get more accurate upper bounds and/or to process larger scale of QOPs, we need more and more computing resources. We implemented a highly parallel SCRM on the Ninf system. Figure 1 shows a Ninf client-server model. The Ninf client controls the SCRM applied to a QOP and generates a large number of subproblems each of which forms an SDP problem at each iteration. All generated SDP problems are sent to the Ninf server machines. Then each Ninf server solves an SDP problem using the software 2
4 SDPA [15]. Note that each Ninf server solves only one SDP problem at a time. After finishing the execution of the SDPA, the result is sent back to the Ninf client machine. Takeda et al. [12] also showed computational efficiency of the SCRM by varying the number of Ninf servers. They reported that SDP problems were allocated to each Ninf server in balance and the total computational time consumed by each Ninf server was almost the same. Therefore good performance and high scalability were attained. Figure 2: Grid Environment(Kyoto Tokyo) We now show some numerical experiments of the SCRM on a Grid environment. Figure 2 illustrates that we have two PC clusters which are located in Kyoto and Tokyo. The Kyoto cluster has 4 nodes which are connected to a Ninf client PC through LAN (100BASE-TX). The Tokyo cluster has 4 nodes which are connected to the Ninf client PC through the Internet called SINET. We estimate the average network speed of the SINET between Kyoto and Tokyo to be about 2 to 4 Mbps. Table 1 shows numerical results of the SCRM on the grid environment illustrated in Figure 2. The first row # subproblems denotes the total number of SDP problems which the Ninf servers in the Kyoto cluster and the Tokyo cluster solved, respectively. The second row denotes the total execution time of the SDPA in each cluster. The total execution time of the Kyoto cluster is almost the same as that of the Tokyo cluster. Note also that the total transmitting time between the Ninf client and Ninf servers is extremely small. From Table 1, we can say that the SCRM is well paralleled in the grid environment. 3 Parallel Polyhedral Homotopy Method Polynomial systems have various applications in many fields of science and engineering. The software package PHoM [6] is designed to find all isolated (real and complex) 3
5 Table 1: Numerical results on grid environments Prob. name LC BLevel20-3 Ninf Server Kyoto Tokyo Kyoto Tokyo # subproblems total execution time(sec.) total trans. time C S(sec.) total trans. time C S(sec.) Table 2: Numerical results of the homotopy method (noon-9 problem (3)) StartSystem Trace Total # CPUs time(s.) sp-up-raito time(s.) sp-up-ratio time(s.) sp-up-ratio , , , , , , , , , , , , solutions of a polynomial system of equations f(x) = 0 such as the cyclic n problem: x 1 + x x n = 0 x 1 x 2 + x 2 x x n x 1 = 0 x 1 x 2 x x n x 1 x 2 = 0, (2)... x 1 x 2 x 3 x n 1 = 0, using a polyhedral homotopy method. The package consists of three modules. The first module which we call StartSystem constructs a family of polyhedral-linear homotopy functions. Takeda et al. [13] implemented a highly paralleled StartSystem using the Ninf. The second module traces homotopy paths to compute all isolated solutions of a polynomial system of equations. The Ninf servers trace the homotopy paths by applying the predictor-corrector method. The third module verifies whether all isolated solutions of the polynomial system of equations have been computed correctly. We also employ the Ninf client-server system, illustrated in Figure 1, for the parallel implementation of the first and second modules of the PHoM. Table 2 shows numerical results of the first and second modules for the noon-9 problem: x 1 x x 1 x x 1 x x = 0 x 2 x x 2 x x 2 x x = 0... x 9 x x 9 x x 9 x x = 0. (3) This problem has 19,665 isolated solutions. All numerical experiments were executed on the Presto I cluster which has 64 nodes (each CPU is Celeron 500MHz). The sp-up-ratio stands for the computational time (# of CPUs is 1) divided by the computational time (# of CPUs is k). If sp-up-ratio is sufficiently close to k, we can regard the software as well paralleled on the Ninf client-server system. Takeda et al. [13] reported that 4
6 the computation of mixed cells by StartSystem (the first module) is suitable for parallel computation through some numerical experiments. We furthermore observe from Table 2 that tracing all the homotopy paths (the second module) is also suitable for parallel computation on the Ninf client-server system. Table 3 also shows numerical results of the second module for the cyclic problems (2) with dimensions n = 11, 12 and 13. There are many homotopy paths, which can be traced independently in parallel; specifically the cyclic 13 problem has 208,012 homotopy paths, and the above noon-9 problem has 19,665 homotopy paths. We use two PC clusters which are located in Tokyo Institute of Technology. All homotopy paths of the cyclic 11 problem were computed in the Presto I cluster, while all homotopy paths of the cyclic 12 and 13 problems were computed on the Presto III cluster which has 256 nodes and 512 CPUs (each CPU is Athlon 1900+). From Table 3, we observe that the second module of the PHoM is quite suitable for parallel computation on PC clusters. Note that more than 5,000 seconds were required to compute all homotopy paths of the cyclic 13 problem even when we used 256 CPUs simultaneously. Therefore parallel computing is indispensable to process larger scale polynomial systems of equations. Table 3: Numerical results of tracing homotopy paths (cyclic n problem) cyclic-11 cyclic-12 cyclic-13 # CPUs time(s.) sp-up-raito time(s.) sp-up-ratio time(s.) sp-up-ratio 2 47, , , , , , , , , , # paths traced 16,796 41, ,012 4 Parallel Implementation of Semidefinite Programming In the last decade, SDP problems have been intensively studied in theoretical, numerical and practical aspects in various fields such as interior-point methods, combinatorial optimization, control and systems, robust optimization and quantum chemistry. Let S n denote the vector space of n n symmetric matrices. For a pair of matrices X, Y S n, the inner product is defined as X Y = n n i=1 j=1 X ijy ij. We use the notation X S n + (S n ++) to indicate that X S n is positive semidefinite (or positive definite). Given A i S n (i = 0, 1,..., m) and b R m, the standard form SDP problem is written as follows: minimize A 0 X subject to A i X = b i (i = 1, 2,..., m), X S n + }. (4) 5
7 The corresponding dual problem is as follows: m maximize b i z i subject to i=1 m A i z i + Y = A 0, Y S n + i=1. (5) It is known that the primal-dual interior point method (PDIPM) is capable of solving these problems in polynomial time. The SDPA (SemiDefinite Programming Algorithm) [15] is an optimization software, written by the C++ language, of the PDIPM for solving the standard form SDP problem. The SDPA incorporates a special data structure for handling block diagonal data matrices and an efficient method proposed by Fujisawa et al. [3] for computing search directions when problems to be solved are large scale and sparse. In many applications, however, SDP problems become too large for SDP software packages including the SDPA to solve on a single processor. It is well-known that the PDIPM for SDP problems has two major time consuming parts at each iteration even if we exploit the sparsity of the date matrices. The first part is the computation of the so-called Schur complement matrix. The second part is the Cholesky factorization of the Schur complement matrix. The SDPARA (SemiDefinite Programming Algorithm parallel version) [16] is a parallel version of the SDPA on multiple processors and distributed memory, which replaces these two bottleneck parts mentioned above by their parallel implementation using MPI and ScaLAPACK (Scalable Linear Algebra PACKage). The SDPARA reads input data m, n, b, A 0, A 1,..., A m and each processor keeps the memory space for the input data and variables X, Y, z independent of other processors, while the Schur complement matrix is divided and stored on each processor. The SDPARA can compute each row of the Schur complement matrix independently in parallel and applies a parallel Cholesky factorization provided by ScaLAPACK to the Schur complement matrix. Figure 3 shows that the SDPARA on a PC cluster (Presto III) attains high scalability for large scale SDP problems. ELEMENTS in Figure 3 corresponds to the computation of the Schur complement matrix, and CHOLESKY does to the Cholesky factorization of the Schur complement matrix. The largest scale problem arising from the quantum chemistry [8] which we can solve involves a 24, , 503 Schur complement matrix. See Yamashita et al. [16] for more details. Zhao et al. [17] proposed another SDP formulation of the quantum chemistry. The size of matrices X and Y in their formulation are very large compared with the SDP formulation proposed in [8], though the size of their Schur complement matrix is not so large (See Table 4). We solved these two problems on the PC cluster which we call ACT-JST cluster. The ACT-JST cluster in Tokyo Denki University has 40 nodes (each node has 1 or 2 GB memory) and 80 CPUs (Athlon 1.2GHz). As we described above, each processor must maintain all input data and variables, whereas the Schur complement matrix is stored on distributed memory. Therefore, the current version of the SDPARA is suitable for a large scale SDP problem which involves a large Schur complement matrix and not so large number of variables like the SDP problem (1) in Table 4. The SDP problem (2) in Table 4 requires that each processor needs more than 1 GB for input data, variables and the distributed Schur complement matrix. When an SDP problem to be solved gets larger, we need to distribute input data and variables among the processors to save memory in each processor. 6
8 Figure 3: High scalability of the SDPARA on Presto III Table 4: Large scale problems arising from the quantum chemistry problem (1):Nakata et al. [8] (2):Zhao et al. [17] size of the Schur complement matrix (= m)) 24,503 7,230 size of X and Y (= n) 630 5,990 time(s.) (64CPUs) (32CPUs) We also developed a new software package called SDPA-C to exploit the sparsity structure of large scale SDP problems based on some fundamental results about positive semidefinite matrix completion [5, 9]. The SDPARA-C [10] is a parallel implementation of SDPA-C using the parallel computing techniques including the ones used for the SD- PARA. We confirmed that the SDPARA-C attains high scalability and works efficiently to solve large scale and sparse SDP problems through some numerical experiments on PC clusters. The point we have to notice is that the SDPARA and the SDPARA-C are designed to be executed on PC clusters in which all nodes are connected by high-speed Local Area Network, since the two software require much amount of network communication between the nodes for the parallel Cholesky factorization. We discuss how users from outside of the PC cluster will access to the SDPARA through Wide Are Network in the next section. 5 Future Works We are planning to apply grid and cluster computing to some other optimization problems such as the vehicle routing problem with time windows and more general mixed integer programming problems. So far we have mainly employed the Ninf GridRPC system (which we call Ninf-1). Tanaka et al. [14] have recently redesigned the Ninf and implemented a new GridRPC system called Ninf-G. The Ninf-G is a full re-implementation of the Ninf-1 using Globus Toolkit. This takes advantage of high interoperability with some other Globus-based Grid systems. We have plans to reconstruct our software packages 7
9 Figure 4: A Grid Portal System for the SDPARA such as the SCRM and the PHoM using the Ninf-G and Globus Toolkit. We are also developing a grid portal system for some optimization problem softwares including the SDPARA. The system will enable users to easily perform parallel computation through the Internet. GridLib in Figure 4 developed by AIST (National Institute of Advanced Industrial Science and Technology, Japan) provides a development framework to construct a virtual computer center as ASP (Application Service Provider). For a GridLib user, no particular knowledge about the Web securities and the Web programming is needed. We have already finished making a grid portal system for the SDPARA with a little deal of trouble. We consider that a few modifications to the SDPARA are necessary to portalize it. The user first accesses the Web portal site, then selects an application and a problem to be solved. GridLib starts up a Ninf-G client program associated with the application selected, and sends the problem to a Ninf-G server over the Internet. Now the Ninf-G server plays the role of a Ninf-1 client, which calls Ninf-1 servers among a PC cluster to execute the SDPARA. After finishing the execution of the SDPARA, the result is back to the Web portal site following the reverse route. References [1] K. Anstreicher, N. Brixius, J-P. Goux and J. Linderoth: Solving Large Quadratic Assignment Problems on Computational Grids. Mathematical Programming, 91 (2002) [2] D. Applegate, R. Bixby, V. Chvátal and W. Cook: Implementing the Dantzig- Fulkerson-Johnson Algorithm for Large Traveling Salesman Problems. Mathematical Programming, 97 (2003) [3] K. Fujisawa, M. Kojima and K. Nakata: Exploiting Sparsity in Primal-Dual Interior- Point Methods for Semidefinite Programming. Mathematical Programming, 79 (1997)
10 [4] K. Fujisawa, M. Kojima, A. Takeda and M. Yamashita: High Performance Grid and Cluster Computing for Some Optimization Problems. Research Report B-400, Department of Mathematical and Computing Sciences, Tokyo Institute of Technology, Oh-Okayama, Meguro, Tokyo, Japan, December [5] M. Fukuda, M. Kojima, K. Murota and K. Nakata: Exploiting Sparsity in Semidefinite Programming via Matrix Completion I: General Framework. SIAM Journal on Optimization, 11 (2000) [6] T. Gunji, S. Kim, M. Kojima, A. Takeda, K. Fujisawa and T. Mizutani: PHoM a Polyhedral Homotopy Continuation Method. December 2002, revised January To appear in Computing. [7] M. Kojima and L. Tuncel: Cones of Matrices and Successive Convex Relaxations of Nonconvex Sets. SIAM Journal on Optimization, 10 (2000) [8] M. Nakata, H. Nakatsuji, M. Ehara, M. Fukuda, K. Nakata and K. Fujisawa: Variational Calculations of Fermion Second-Order Deduced Density Matrices by Semidefinite Programming Algorithm. Journal of Chemical Physics, 114 (2001) [9] K. Nakata, K. Fujisawa, M. Fukuda, M. Kojima and K. Murota: Exploit Sparsity in Semidefinite Programming via Matrix Completion II: Implementation and Numerical Results. Mathematical Programming, 95 (2003) [10] K. Nakata, M. Yamashita, K. Fujisawa and M. Kojima: A Parallel Primal-Dual Interior-Point Method for Semidefinite Programs Using Positive Definite Matrix Completion. Research Report B-398, Department of Mathematical and Computing Sciences, Tokyo Institute of Technology, Oh-Okayama, Meguro, Tokyo, Japan, November [11] M. Sato, H. Nakada, S. Sekiguchi, S. Matsuoka, U. Nagashima and H. Takagi: Ninf: A Network based Information Library for a Global World-Wide Computing Infrastructure. HPCN 97, LNCS-1225 (1997) [12] A. Takeda, K. Fujisawa, Y. Fukaya and M. Kojima: Parallel Implementation of Successive Convex Relaxation Methods for Quadratic Optimization Problems. Journal of Global Optimization, 24 (2002) [13] A. Takeda, M. Kojima and K. Fujisawa: Enumeration of All Solutions of a Combinatorial Linear Inequality System Arising from the Polyhedral Homotopy Continuation Method. Journal of the Operations Research Society of Japan, 45 (2002) [14] Y. Tanaka, H. Nakada, S. Sekiguchi, T. Suzumura and S. Matsuoka: Ninf-G : A Reference Implementation of RPC-based Programming Middleware for Grid Computing. Journal of Grid Computing, 1 (2003) [15] M. Yamashita, K. Fujisawa and M. Kojima: Implementation and Evaluation of SDPA 6.0 (SemiDefinite Programming Algorithm 6.0). Journal of Optimization Methods and Software, 18 (2003) [16] M. Yamashita, K. Fujisawa and M. Kojima: SDPARA : SemiDefinite Programming Algorithm PARAllel Version. Journal of Parallel Computing, 29 (2003)
11 [17] Z. Zhao, B. J. Braams, M. Fukuda, M. L. Overton and J. K. Percus: The Reduced Density Matrix Method for Electronic Structure Calculations and the Role of Three- Index Representability. To appear in Journal of Chemical Physics,
Solving polynomial least squares problems via semidefinite programming relaxations
Solving polynomial least squares problems via semidefinite programming relaxations Sunyoung Kim and Masakazu Kojima August 2007, revised in November, 2007 Abstract. A polynomial optimization problem whose
A high-performance software package for semidefinite programs: SDPA 7
A high-performance software package for semidefinite programs: SDPA 7 Makoto Yamashita, Katsuki Fujisawa, Kazuhide Nakata, Maho Nakata, Mituhiro Fukuda, Kazuhiro Kobayashi, and Kazushige Goto + January,
An Overview Of Software For Convex Optimization. Brian Borchers Department of Mathematics New Mexico Tech Socorro, NM 87801 borchers@nmt.
An Overview Of Software For Convex Optimization Brian Borchers Department of Mathematics New Mexico Tech Socorro, NM 87801 [email protected] In fact, the great watershed in optimization isn t between linearity
An Introduction on SemiDefinite Program
An Introduction on SemiDefinite Program from the viewpoint of computation Hayato Waki Institute of Mathematics for Industry, Kyushu University 2015-10-08 Combinatorial Optimization at Work, Berlin, 2015
Big Data Optimization at SAS
Big Data Optimization at SAS Imre Pólik et al. SAS Institute Cary, NC, USA Edinburgh, 2013 Outline 1 Optimization at SAS 2 Big Data Optimization at SAS The SAS HPA architecture Support vector machines
CHAPTER 9. Integer Programming
CHAPTER 9 Integer Programming An integer linear program (ILP) is, by definition, a linear program with the additional constraint that all variables take integer values: (9.1) max c T x s t Ax b and x integral
Nonlinear Programming Methods.S2 Quadratic Programming
Nonlinear Programming Methods.S2 Quadratic Programming Operations Research Models and Methods Paul A. Jensen and Jonathan F. Bard A linearly constrained optimization problem with a quadratic objective
Optimal Scheduling for Dependent Details Processing Using MS Excel Solver
BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 8, No 2 Sofia 2008 Optimal Scheduling for Dependent Details Processing Using MS Excel Solver Daniela Borissova Institute of
Applied Algorithm Design Lecture 5
Applied Algorithm Design Lecture 5 Pietro Michiardi Eurecom Pietro Michiardi (Eurecom) Applied Algorithm Design Lecture 5 1 / 86 Approximation Algorithms Pietro Michiardi (Eurecom) Applied Algorithm Design
6. Cholesky factorization
6. Cholesky factorization EE103 (Fall 2011-12) triangular matrices forward and backward substitution the Cholesky factorization solving Ax = b with A positive definite inverse of a positive definite matrix
The Assessment of Benchmarks Executed on Bare-Metal and Using Para-Virtualisation
The Assessment of Benchmarks Executed on Bare-Metal and Using Para-Virtualisation Mark Baker, Garry Smith and Ahmad Hasaan SSE, University of Reading Paravirtualization A full assessment of paravirtualization
Completely Positive Cone and its Dual
On the Computational Complexity of Membership Problems for the Completely Positive Cone and its Dual Peter J.C. Dickinson Luuk Gijben July 3, 2012 Abstract Copositive programming has become a useful tool
2.3 Convex Constrained Optimization Problems
42 CHAPTER 2. FUNDAMENTAL CONCEPTS IN CONVEX OPTIMIZATION Theorem 15 Let f : R n R and h : R R. Consider g(x) = h(f(x)) for all x R n. The function g is convex if either of the following two conditions
24. The Branch and Bound Method
24. The Branch and Bound Method It has serious practical consequences if it is known that a combinatorial problem is NP-complete. Then one can conclude according to the present state of science that no
Advances in Convex Optimization: Interior-point Methods, Cone Programming, and Applications
Advances in Convex Optimization: Interior-point Methods, Cone Programming, and Applications Stephen Boyd Electrical Engineering Department Stanford University (joint work with Lieven Vandenberghe, UCLA)
Interior-Point Algorithms for Quadratic Programming
Interior-Point Algorithms for Quadratic Programming Thomas Reslow Krüth Kongens Lyngby 2008 IMM-M.Sc-2008-19 Technical University of Denmark Informatics and Mathematical Modelling Building 321, DK-2800
Support Vector Machines with Clustering for Training with Very Large Datasets
Support Vector Machines with Clustering for Training with Very Large Datasets Theodoros Evgeniou Technology Management INSEAD Bd de Constance, Fontainebleau 77300, France [email protected] Massimiliano
SOLVING LINEAR SYSTEMS
SOLVING LINEAR SYSTEMS Linear systems Ax = b occur widely in applied mathematics They occur as direct formulations of real world problems; but more often, they occur as a part of the numerical analysis
Massive Data Classification via Unconstrained Support Vector Machines
Massive Data Classification via Unconstrained Support Vector Machines Olvi L. Mangasarian and Michael E. Thompson Computer Sciences Department University of Wisconsin 1210 West Dayton Street Madison, WI
Advanced Lecture on Mathematical Science and Information Science I. Optimization in Finance
Advanced Lecture on Mathematical Science and Information Science I Optimization in Finance Reha H. Tütüncü Visiting Associate Professor Dept. of Mathematical and Computing Sciences Tokyo Institute of Technology
Minimizing costs for transport buyers using integer programming and column generation. Eser Esirgen
MASTER STHESIS Minimizing costs for transport buyers using integer programming and column generation Eser Esirgen DepartmentofMathematicalSciences CHALMERS UNIVERSITY OF TECHNOLOGY UNIVERSITY OF GOTHENBURG
The Lattice Project: A Multi-Model Grid Computing System. Center for Bioinformatics and Computational Biology University of Maryland
The Lattice Project: A Multi-Model Grid Computing System Center for Bioinformatics and Computational Biology University of Maryland Parallel Computing PARALLEL COMPUTING a form of computation in which
MEng, BSc Applied Computer Science
School of Computing FACULTY OF ENGINEERING MEng, BSc Applied Computer Science Year 1 COMP1212 Computer Processor Effective programming depends on understanding not only how to give a machine instructions
Lecture 11: 0-1 Quadratic Program and Lower Bounds
Lecture : - Quadratic Program and Lower Bounds (3 units) Outline Problem formulations Reformulation: Linearization & continuous relaxation Branch & Bound Method framework Simple bounds, LP bound and semidefinite
GridSolve: : A Seamless Bridge Between the Standard Programming Interfaces and Remote Resources
GridSolve: : A Seamless Bridge Between the Standard Programming Interfaces and Remote Resources Jack Dongarra University of Tennessee and Oak Ridge National Laboratory 2/25/2006 1 Overview Grid/NetSolve
Duality of linear conic problems
Duality of linear conic problems Alexander Shapiro and Arkadi Nemirovski Abstract It is well known that the optimal values of a linear programming problem and its dual are equal to each other if at least
A QCQP Approach to Triangulation. Chris Aholt, Sameer Agarwal, and Rekha Thomas University of Washington 2 Google, Inc.
A QCQP Approach to Triangulation 1 Chris Aholt, Sameer Agarwal, and Rekha Thomas 1 University of Washington 2 Google, Inc. 2 1 The Triangulation Problem X Given: -n camera matrices P i R 3 4 -n noisy observations
Recall that two vectors in are perpendicular or orthogonal provided that their dot
Orthogonal Complements and Projections Recall that two vectors in are perpendicular or orthogonal provided that their dot product vanishes That is, if and only if Example 1 The vectors in are orthogonal
Introduction to Support Vector Machines. Colin Campbell, Bristol University
Introduction to Support Vector Machines Colin Campbell, Bristol University 1 Outline of talk. Part 1. An Introduction to SVMs 1.1. SVMs for binary classification. 1.2. Soft margins and multi-class classification.
APPM4720/5720: Fast algorithms for big data. Gunnar Martinsson The University of Colorado at Boulder
APPM4720/5720: Fast algorithms for big data Gunnar Martinsson The University of Colorado at Boulder Course objectives: The purpose of this course is to teach efficient algorithms for processing very large
2014-2015 The Master s Degree with Thesis Course Descriptions in Industrial Engineering
2014-2015 The Master s Degree with Thesis Course Descriptions in Industrial Engineering Compulsory Courses IENG540 Optimization Models and Algorithms In the course important deterministic optimization
MEng, BSc Computer Science with Artificial Intelligence
School of Computing FACULTY OF ENGINEERING MEng, BSc Computer Science with Artificial Intelligence Year 1 COMP1212 Computer Processor Effective programming depends on understanding not only how to give
DATA ANALYSIS II. Matrix Algorithms
DATA ANALYSIS II Matrix Algorithms Similarity Matrix Given a dataset D = {x i }, i=1,..,n consisting of n points in R d, let A denote the n n symmetric similarity matrix between the points, given as where
Optimization Modeling for Mining Engineers
Optimization Modeling for Mining Engineers Alexandra M. Newman Division of Economics and Business Slide 1 Colorado School of Mines Seminar Outline Linear Programming Integer Linear Programming Slide 2
Mixed Precision Iterative Refinement Methods Energy Efficiency on Hybrid Hardware Platforms
Mixed Precision Iterative Refinement Methods Energy Efficiency on Hybrid Hardware Platforms Björn Rocker Hamburg, June 17th 2010 Engineering Mathematics and Computing Lab (EMCL) KIT University of the State
Optimization Methods in Finance
Optimization Methods in Finance Gerard Cornuejols Reha Tütüncü Carnegie Mellon University, Pittsburgh, PA 15213 USA January 2006 2 Foreword Optimization models play an increasingly important role in financial
Minimizing the Number of Machines in a Unit-Time Scheduling Problem
Minimizing the Number of Machines in a Unit-Time Scheduling Problem Svetlana A. Kravchenko 1 United Institute of Informatics Problems, Surganova St. 6, 220012 Minsk, Belarus [email protected] Frank
Nonlinear Optimization: Algorithms 3: Interior-point methods
Nonlinear Optimization: Algorithms 3: Interior-point methods INSEAD, Spring 2006 Jean-Philippe Vert Ecole des Mines de Paris [email protected] Nonlinear optimization c 2006 Jean-Philippe Vert,
An interval linear programming contractor
An interval linear programming contractor Introduction Milan Hladík Abstract. We consider linear programming with interval data. One of the most challenging problems in this topic is to determine or tight
ACCELERATING COMMERCIAL LINEAR DYNAMIC AND NONLINEAR IMPLICIT FEA SOFTWARE THROUGH HIGH- PERFORMANCE COMPUTING
ACCELERATING COMMERCIAL LINEAR DYNAMIC AND Vladimir Belsky Director of Solver Development* Luis Crivelli Director of Solver Development* Matt Dunbar Chief Architect* Mikhail Belyi Development Group Manager*
Discrete Optimization
Discrete Optimization [Chen, Batson, Dang: Applied integer Programming] Chapter 3 and 4.1-4.3 by Johan Högdahl and Victoria Svedberg Seminar 2, 2015-03-31 Todays presentation Chapter 3 Transforms using
15.062 Data Mining: Algorithms and Applications Matrix Math Review
.6 Data Mining: Algorithms and Applications Matrix Math Review The purpose of this document is to give a brief review of selected linear algebra concepts that will be useful for the course and to develop
Report: Declarative Machine Learning on MapReduce (SystemML)
Report: Declarative Machine Learning on MapReduce (SystemML) Jessica Falk ETH-ID 11-947-512 May 28, 2014 1 Introduction SystemML is a system used to execute machine learning (ML) algorithms in HaDoop,
FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT MINING SYSTEM
International Journal of Innovative Computing, Information and Control ICIC International c 0 ISSN 34-48 Volume 8, Number 8, August 0 pp. 4 FUZZY CLUSTERING ANALYSIS OF DATA MINING: APPLICATION TO AN ACCIDENT
High Performance Computing. Course Notes 2007-2008. HPC Fundamentals
High Performance Computing Course Notes 2007-2008 2008 HPC Fundamentals Introduction What is High Performance Computing (HPC)? Difficult to define - it s a moving target. Later 1980s, a supercomputer performs
Performance Characteristics of a Cost-Effective Medium-Sized Beowulf Cluster Supercomputer
Res. Lett. Inf. Math. Sci., 2003, Vol.5, pp 1-10 Available online at http://iims.massey.ac.nz/research/letters/ 1 Performance Characteristics of a Cost-Effective Medium-Sized Beowulf Cluster Supercomputer
Branch and Cut for TSP
Branch and Cut for TSP jla,[email protected] Informatics and Mathematical Modelling Technical University of Denmark 1 Branch-and-Cut for TSP Branch-and-Cut is a general technique applicable e.g. to solve symmetric
Proximal mapping via network optimization
L. Vandenberghe EE236C (Spring 23-4) Proximal mapping via network optimization minimum cut and maximum flow problems parametric minimum cut problem application to proximal mapping Introduction this lecture:
5 INTEGER LINEAR PROGRAMMING (ILP) E. Amaldi Fondamenti di R.O. Politecnico di Milano 1
5 INTEGER LINEAR PROGRAMMING (ILP) E. Amaldi Fondamenti di R.O. Politecnico di Milano 1 General Integer Linear Program: (ILP) min c T x Ax b x 0 integer Assumption: A, b integer The integrality condition
Appendix: Simple Methods for Shift Scheduling in Multi-Skill Call Centers
MSOM.1070.0172 Appendix: Simple Methods for Shift Scheduling in Multi-Skill Call Centers In Bhulai et al. (2006) we presented a method for computing optimal schedules, separately, after the optimal staffing
Inner Product Spaces and Orthogonality
Inner Product Spaces and Orthogonality week 3-4 Fall 2006 Dot product of R n The inner product or dot product of R n is a function, defined by u, v a b + a 2 b 2 + + a n b n for u a, a 2,, a n T, v b,
Discuss the size of the instance for the minimum spanning tree problem.
3.1 Algorithm complexity The algorithms A, B are given. The former has complexity O(n 2 ), the latter O(2 n ), where n is the size of the instance. Let n A 0 be the size of the largest instance that can
Polytope Examples (PolyComp Fukuda) Matching Polytope 1
Polytope Examples (PolyComp Fukuda) Matching Polytope 1 Matching Polytope Let G = (V,E) be a graph. A matching in G is a subset of edges M E such that every vertex meets at most one member of M. A matching
A Reference Point Method to Triple-Objective Assignment of Supporting Services in a Healthcare Institution. Bartosz Sawik
Decision Making in Manufacturing and Services Vol. 4 2010 No. 1 2 pp. 37 46 A Reference Point Method to Triple-Objective Assignment of Supporting Services in a Healthcare Institution Bartosz Sawik Abstract.
Big Data - Lecture 1 Optimization reminders
Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Big Data - Lecture 1 Optimization reminders S. Gadat Toulouse, Octobre 2014 Schedule Introduction Major issues Examples Mathematics
Analysis of a Production/Inventory System with Multiple Retailers
Analysis of a Production/Inventory System with Multiple Retailers Ann M. Noblesse 1, Robert N. Boute 1,2, Marc R. Lambrecht 1, Benny Van Houdt 3 1 Research Center for Operations Management, University
Parallel Visualization for GIS Applications
Parallel Visualization for GIS Applications Alexandre Sorokine, Jamison Daniel, Cheng Liu Oak Ridge National Laboratory, Geographic Information Science & Technology, PO Box 2008 MS 6017, Oak Ridge National
Journal of Theoretical and Applied Information Technology 20 th July 2015. Vol.77. No.2 2005-2015 JATIT & LLS. All rights reserved.
EFFICIENT LOAD BALANCING USING ANT COLONY OPTIMIZATION MOHAMMAD H. NADIMI-SHAHRAKI, ELNAZ SHAFIGH FARD, FARAMARZ SAFI Department of Computer Engineering, Najafabad branch, Islamic Azad University, Najafabad,
Scaling from 1 PC to a super computer using Mascot
Scaling from 1 PC to a super computer using Mascot 1 Mascot - options for high throughput When is more processing power required? Options: Multiple Mascot installations Using a large Unix server Clusters
RevoScaleR Speed and Scalability
EXECUTIVE WHITE PAPER RevoScaleR Speed and Scalability By Lee Edlefsen Ph.D., Chief Scientist, Revolution Analytics Abstract RevoScaleR, the Big Data predictive analytics library included with Revolution
Parallel Data Selection Based on Neurodynamic Optimization in the Era of Big Data
Parallel Data Selection Based on Neurodynamic Optimization in the Era of Big Data Jun Wang Department of Mechanical and Automation Engineering The Chinese University of Hong Kong Shatin, New Territories,
Parallelism and Cloud Computing
Parallelism and Cloud Computing Kai Shen Parallel Computing Parallel computing: Process sub tasks simultaneously so that work can be completed faster. For instances: divide the work of matrix multiplication
GAMS, Condor and the Grid: Solving Hard Optimization Models in Parallel. Michael C. Ferris University of Wisconsin
GAMS, Condor and the Grid: Solving Hard Optimization Models in Parallel Michael C. Ferris University of Wisconsin Parallel Optimization Aid search for global solutions (typically in non-convex or discrete)
The Gurobi Optimizer
The Gurobi Optimizer Gurobi History Gurobi Optimization, founded July 2008 Zonghao Gu, Ed Rothberg, Bob Bixby Started code development March 2008 Gurobi Version 1.0 released May 2009 History of rapid,
Approximation Algorithms
Approximation Algorithms or: How I Learned to Stop Worrying and Deal with NP-Completeness Ong Jit Sheng, Jonathan (A0073924B) March, 2012 Overview Key Results (I) General techniques: Greedy algorithms
Mathematics Course 111: Algebra I Part IV: Vector Spaces
Mathematics Course 111: Algebra I Part IV: Vector Spaces D. R. Wilkins Academic Year 1996-7 9 Vector Spaces A vector space over some field K is an algebraic structure consisting of a set V on which are
A Simultaneous Solution for General Linear Equations on a Ring or Hierarchical Cluster
Acta Technica Jaurinensis Vol. 3. No. 1. 010 A Simultaneous Solution for General Linear Equations on a Ring or Hierarchical Cluster G. Molnárka, N. Varjasi Széchenyi István University Győr, Hungary, H-906
Conic optimization: examples and software
Conic optimization: examples and software Etienne de Klerk Tilburg University, The Netherlands Etienne de Klerk (Tilburg University) Conic optimization: examples and software 1 / 16 Outline Conic optimization
USING EXCEL SOLVER IN OPTIMIZATION PROBLEMS
USING EXCEL SOLVER IN OPTIMIZATION PROBLEMS Leslie Chandrakantha John Jay College of Criminal Justice of CUNY Mathematics and Computer Science Department 445 West 59 th Street, New York, NY 10019 [email protected]
BIG DATA PROBLEMS AND LARGE-SCALE OPTIMIZATION: A DISTRIBUTED ALGORITHM FOR MATRIX FACTORIZATION
BIG DATA PROBLEMS AND LARGE-SCALE OPTIMIZATION: A DISTRIBUTED ALGORITHM FOR MATRIX FACTORIZATION Ş. İlker Birbil Sabancı University Ali Taylan Cemgil 1, Hazal Koptagel 1, Figen Öztoprak 2, Umut Şimşekli
System Models for Distributed and Cloud Computing
System Models for Distributed and Cloud Computing Dr. Sanjay P. Ahuja, Ph.D. 2010-14 FIS Distinguished Professor of Computer Science School of Computing, UNF Classification of Distributed Computing Systems
! Solve problem to optimality. ! Solve problem in poly-time. ! Solve arbitrary instances of the problem. #-approximation algorithm.
Approximation Algorithms 11 Approximation Algorithms Q Suppose I need to solve an NP-hard problem What should I do? A Theory says you're unlikely to find a poly-time algorithm Must sacrifice one of three
P013 INTRODUCING A NEW GENERATION OF RESERVOIR SIMULATION SOFTWARE
1 P013 INTRODUCING A NEW GENERATION OF RESERVOIR SIMULATION SOFTWARE JEAN-MARC GRATIEN, JEAN-FRANÇOIS MAGRAS, PHILIPPE QUANDALLE, OLIVIER RICOIS 1&4, av. Bois-Préau. 92852 Rueil Malmaison Cedex. France
HPC enabling of OpenFOAM R for CFD applications
HPC enabling of OpenFOAM R for CFD applications Towards the exascale: OpenFOAM perspective Ivan Spisso 25-27 March 2015, Casalecchio di Reno, BOLOGNA. SuperComputing Applications and Innovation Department,
PRIMERGY server-based High Performance Computing solutions
PRIMERGY server-based High Performance Computing solutions PreSales - May 2010 - HPC Revenue OS & Processor Type Increasing standardization with shift in HPC to x86 with 70% in 2008.. HPC revenue by operating
Load balancing in a heterogeneous computer system by self-organizing Kohonen network
Bull. Nov. Comp. Center, Comp. Science, 25 (2006), 69 74 c 2006 NCC Publisher Load balancing in a heterogeneous computer system by self-organizing Kohonen network Mikhail S. Tarkov, Yakov S. Bezrukov Abstract.
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh
Modern Optimization Methods for Big Data Problems MATH11146 The University of Edinburgh Peter Richtárik Week 3 Randomized Coordinate Descent With Arbitrary Sampling January 27, 2016 1 / 30 The Problem
Chapter 1. Introduction
Chapter 1 Introduction Intermodal freight transportation describes the movement of goods in standardized loading units (e.g., containers) by at least two transportation modes (rail, maritime, and road)
COORDINATION PRODUCTION AND TRANSPORTATION SCHEDULING IN THE SUPPLY CHAIN ABSTRACT
Technical Report #98T-010, Department of Industrial & Mfg. Systems Egnieering, Lehigh Univerisity (1998) COORDINATION PRODUCTION AND TRANSPORTATION SCHEDULING IN THE SUPPLY CHAIN Kadir Ertogral, S. David
Solution of a Large-Scale Traveling-Salesman Problem
Chapter 1 Solution of a Large-Scale Traveling-Salesman Problem George B. Dantzig, Delbert R. Fulkerson, and Selmer M. Johnson Introduction by Vašek Chvátal and William Cook The birth of the cutting-plane
GENERALIZED INTEGER PROGRAMMING
Professor S. S. CHADHA, PhD University of Wisconsin, Eau Claire, USA E-mail: [email protected] Professor Veena CHADHA University of Wisconsin, Eau Claire, USA E-mail: [email protected] GENERALIZED INTEGER
A hierarchical multicriteria routing model with traffic splitting for MPLS networks
A hierarchical multicriteria routing model with traffic splitting for MPLS networks João Clímaco, José Craveirinha, Marta Pascoal jclimaco@inesccpt, jcrav@deecucpt, marta@matucpt University of Coimbra
Statistical machine learning, high dimension and big data
Statistical machine learning, high dimension and big data S. Gaïffas 1 14 mars 2014 1 CMAP - Ecole Polytechnique Agenda for today Divide and Conquer principle for collaborative filtering Graphical modelling,
Branch-and-Price Approach to the Vehicle Routing Problem with Time Windows
TECHNISCHE UNIVERSITEIT EINDHOVEN Branch-and-Price Approach to the Vehicle Routing Problem with Time Windows Lloyd A. Fasting May 2014 Supervisors: dr. M. Firat dr.ir. M.A.A. Boon J. van Twist MSc. Contents
Optimization Theory for Large Systems
Optimization Theory for Large Systems LEON S. LASDON CASE WESTERN RESERVE UNIVERSITY THE MACMILLAN COMPANY COLLIER-MACMILLAN LIMITED, LONDON Contents 1. Linear and Nonlinear Programming 1 1.1 Unconstrained
Experiences of numerical simulations on a PC cluster Antti Vanne December 11, 2002
xperiences of numerical simulations on a P cluster xperiences of numerical simulations on a P cluster ecember xperiences of numerical simulations on a P cluster Introduction eowulf concept Using commodity
