Big Data Predictive Analtyics for Proactive Semiconductor Equipment Maintenance: A Review
|
|
|
- Madison Marshall
- 10 years ago
- Views:
Transcription
1 Big Data Predictive Analtyics for Proactive Semiconductor Equipment Maintenance: A Review Sathyan Munirathinam.,M.C.A.,M.S Business Intelligence Engineer Micron Technology, Inc., Boise, USA [email protected] Abstract Manufacturing Industry generates about a third of all data today and the modern semiconductor manufacturing is one of the most contribution to this tsunami data volume. Terabytes of data is generated on a daily basis during ~500 steps in semiconductor chip processing. During this complex manufacturing process, equipment downtime may cause a significant loss of productivity and profit. In this paper, we are going to explore the predictive analytical algorithms and big data techniques in helping to achieve near-zero equipment downtime in the fabrication unit and to improve OEE (Overall Equipment Effectiveness), which is a key machine manufacturing productivity metric. Keywords Big Data, Equipment, Semiconductor, Predictive Analytics (PA), OEE I. INTRODUCTION WITH the entire big buzz about Big Data in different industries, there has been little impact about its usage in semiconductor manufacturing. The role and size of data volume in manufacturing has traditionally been understated. The data is very complicated and it is fragmented across the data sources between manufacturing process, quality & inspection, maintenance, testing and process control operations; To stay competitive, It is clearly known fact that there is a need in improving the predictive capabilities based on historical data of equipments to enable the fabrication facility to proactively allocate limited maintenance resources to the right location at the right time and thus maintain the high yield while achieving a high system uptime. The technology transition from 300mm wafers to 450mm demands higher operational efficiencies, minimizing downtime of equipment and improving yield from newer generation of fabrication units. Each fabrication tool can cost millions of dollars. A 300- mm fab manufacturing 20,000 wafers per month at the 65-nm node is expected to require ~$1.8 billion equipment investment. This investment is impelling semiconductor manufacturers to remain focused on maximizing equipment productivity. Industry leaders are realizing that on their own, they cannot afford to do the learning necessary to maximize equipment investment. Current database management systems have limited functionality in their ability to access, analyze, and intelligently extract information from the large variety of manufacturing data sources available within this industry. It is critical that the semiconductor industry rely on big-data Dr.B.Ramadoss, M.Tech.,PhD Professor, Department of Computer Applications, National Institute of Technology, Trichy, India [email protected] technologies access multiple data sources and derive useful defect and yield information to ensure valuable insight. Semiconductor fabrication is one of the most complicated manufacturing processes, in which the current prevailing maintenance practices are preventive maintenance, using either time-based or wafer-based scheduling strategies, which may lead to the tools being either over-maintained or undermaintained, may cause a significant loss of productivity and profit. Currently, the majority of maintenance operations in the semiconductor industry are still based on either historical reliability of fabrication equipment, or on diagnostic information from equipment performance signatures extracted from in-situ sensors. Such a fragmented, diagnosis-centered approach leads to mostly preventive maintenance along with reactive maintenance policies that use neither abundant product quality, equipment condition, equipment reliability information, nor the temporal dynamics inside that information in order to anticipate future events in the system and thus facilitate a more proactive maintenance policy. Production efficiency is a measure of the average time a process is producing product. It averages at 60% and there is a large consensus that 80% is a realistic goal. Maintenance accounts for a large portion of the remaining 40%. Tool availability is directly related to the production efficiency of a manufacturing process. It is the percentage of time a tool is available for producing product. Increased tool availability increases production efficiency. Maintenance causes a decrease in tool availability hence a loss in production efficiency and profitability. For this reason a properly functioning maintenance strategy is required for companies to compete for market share. The majority of manufacturing industries carry out maintenance using either a corrective or preventative maintenance strategy. Corrective maintenance is carried out after the failure of an equipment component, also referred to as a reactive based maintenance. In contrast preventative maintenance is carried out on a predetermined periodic time frame. This allows maintenance to be scheduled and carried out before the failure occurs. At Intel, a High Precision Maintenance (HPM) strategy is being utilized to perform preventative maintenance tasks with high precision. This strategy was implemented to reduce maintenance costs. Utilizing these maintenance strategies does not produce the optimal production efficiency, as fully functioning equipment components are replaced regardless of their condition. In recent years industries have begun to utilize predictive maintenance. This allows equipment to be maintained based on its condition. ASE 2014 ISBN:
2 A review of literature published in this field shows that there rarely exists condition-based maintenance (CBM) utilizing equipment condition as indicator, and almost no predictive maintenance (PdM) utilizing the prediction o This paper examines some of the analytical techniques, predictive models and control charts that can be leveraged by the semiconductor industry. This paper is aimed at to survey different big data predictive modeling methods for optimal maintenance in semiconductor manufacturing processes. Deteriorated equipment has a significant impact on the product quality and maintenance policies. II. PROBLEM DEFINITION The unscheduled breakdown of an individual piece of manufacturing equipment is a great cost burden, as it results in downtime for dependent/related equipment in the manufacturing chain. Inefficiencies of a single piece of critical "bottleneck" process equipment can have a negative economic impact for an entire production line. Since numerous competitive equipment manufacturers supply the various discrete tools needed in a single fab, service, tool monitoring, and data collection for the tools are inefficient and cumbersome processes. constitute the ingredients for designing data-driven models that enable virtual metrology and predictive maintenance to control manufacturing line efficiency and product quality. However the data from semiconductor manufacturing often suffer from inherent and difficult problems: high dimensionality due to the number of processing steps and measurements recorded during the entire manufacturing process, imbalanced data due to the nature of manufacturing line with majority of wafers passing and examples of defective wafers occurring rarely, missing data due to production line efficiency constraints that do not allow every measurement to be taken at every step of every wafer. There are various predictive algorithms is applied to various semiconductor equipments (few is listed below) Dry Etch Ion Implantation Tool Chemical Vapor Deposition (CVD) Tool Photolithography Semiconductor industry statistics show that most semiconductor capital equipment suffers at least 8% unscheduled downtime and loses another 7% to scheduled maintenance. At the January 2000 ISS, Michael Splinter, Senior Vice President and General Manager of Intel's technology and manufacturing group, estimated that each hour of downtime for a critical unit of process equipment can translate into $100,000 of lost revenue in today's chip-hungry market. In a typical fab, just reducing downtime by 1% on the 50 most critical tools can provide revenue opportunities and cost savings nearing one hundred million dollars annually. One way to reduce unscheduled downtime is to improve response time and repair time, and, eventually, to predict when problems will occur before they occur. High system complexity and data volume have prevented the III. BACKGROUND A. Semiconductor Manufacturing Semiconductor manufacturing is a complex process which consists of hundreds of manufacturing steps. Thus semiconductor manufacturing process is normally under consistent monitoring, it collects sensor signals from process tools, and it measures physical and electrical parameters from wafers through metrology. These tool data and metrology data It's likely that predictive maintenance and equipment monitoring will become more prevalent, as manufacturers try to minimize downtime and maximize productivity. Semiconductor factories will apply mathematical models/machine learning models to equipment and factory data to assess health and predict tool failure and then act on those outputs to optimize factory productivity. The application will monitor tools and predict failures. B. Big Data As Semi-conductor industry moves to sub-20nm technology nodes and 450mm wafer sizes, the sensor technology that monitors equipments becomes more complex, data volumes will continue to grow in an exponential rate. Due to the sheer size of manufacturing data being generated it is becoming increasingly difficult to analyze it using relational databases. Such huge volumes of data will become impossible for the traditional databases to capture and convert to meaningful information. The rise of big data and its usage can be beneficial to operations and maintenance staff in order to be proactive with regard to ongoing equipment maintenance and upkeep. The use of data can provide more opportunity to predict upcoming issues in a system or equipment and, therefore utilize maintenance in a predictive manner, rather than relying on the costly extremes such as random Preventative Maintenance or crisis-related Reactive Maintenance. Of course, all of this Big Data is meaningless without analysis. There are hidden patterns lurking within ASE 2014 ISBN:
3 these facts and figures. Decoding these patterns is what powers predictive maintenance and separates it from more traditional, reactionary approaches to equipment repair and replacement. Effectively utilizing Big Data can have a huge impact on improving manufacturing effectiveness. The cost of failed equipment leading to unplanned downtime is a major reason that the use of predictive analytics or big data has received a great amount of interest, especially when tied directly to preventative maintenance for maintaining motors, equipment, sensors, and the like. With big data, predictive preventative maintenance can optimize maintenance planning while minimalizing consequential costs due to faulty equipment. Traditional relational database management system (RDBMS) technologies such as Oracle, SQL Server, DB2, etc. are approaching their limits when processing massive sets of data in complex data analytics in an OLTP environment. Big data technologies provide a combination of distributed databases, map-reduce based processing, machine learning framework and memory-resident graph databases. In order to achieve an optimal maintenance decision-making, a method should be in place to integrate multiple data sources from different data domains. Production data, machine functional data, and sensor data are all aggregated for analysis and used to build models for predicting machine failure or poor product quality reducing failure times and costs. However handling this volume of data requires an engine that can integrate all the elements of the technology stack, extraction-transformation-loading (ETL), data cleansing, data storage, reporting, statistical modeling and data mining into either a single platform or a seamless stack, that can scale up to 10s of terabytes, handles multiple large tables of billions of rows, and executes x faster than conventional relational database technology. C. Predictive Maintenace(PdM) The pressure to reduce production costs forces manufactures to optimize equipment utilization. While the concept of Predictive Maintenance is discussed for several years (Mobley 2002, Dekker 1996, Liu 2008), tools in semiconductor processing are usually still controlled by Statistical Process Control (SPC) and Fault Detection and Classification (FDC), but these techniques do not allow prediction of tool failures. Predictive Maintenance is still not implemented in semi-conductor processing due to the complexity of the data volume. Maintenance Policies can be divided into four categories, with different levels of complexity, Suso et al. (2012a); Mobley (2002) Run to Failure (R2F) Maintenance Preventive Maintenance Predictive Maintenance Condition-based Maintenance Run to Failure (R2F) Maintenance: Corrective maintenance is the classic Run-to-Failure reactive maintenance that has no special maintenance plan in place. The machine is assumed to be fit unless proven otherwise. Cons: o High risk of collateral damage and secondary failure o High production downtime o Overtime labor and high cost of spare parts Pros: o Machines are not over-maintained o No overhead of condition monitoring or planning costs Preventive Maintenance: Preventive maintenance (PM) is the popular periodic maintenance strategy that is actively employed by all manufacturers and operators in the industry today. An optimal breakdown window is pre-calculated (at the time of component design or installation, based on a wide range of models describing the degradation process of equipment, cost structure and admissible maintenance etc.), and a preventive maintenance schedule is laid out. Maintenance is carried-out on those periodic intervals, assuming that the machine is going to break otherwise. Cons: o Calendar-based maintenance: Machines are repaired when there are no faults o There will still be unscheduled breakdowns ASE 2014 ISBN:
4 Pros: o Fewer catastrophic failures and lesser collateral damage o Greater control over spare-parts and inventory o Maintenance is performed in controlled manner, with a rough estimate of costs well-known ahead of time make smarter decisions to optimize the equipment maintenance. In the semiconductor industry, turning data into actionable information to support predictive analytics, which is a major challenge that involves both data management and processing. The main challenge arises from the explosive growth of data and it s processing for predictive analytics and decision tree analysis. Predictive Maintenance: Predictive Maintenance (PdM) is an alternative to the above two that employs predictive analytics over real-time data collected (streamed) from parts of the machine to a centralized processor that detects variations in the functional parameters and detects anomalies that can potentially lead to breakdowns. The real-time nature of the analytics helps identify the functional breakdowns long before they happen but soon after their potential cause arises. Pros: o Unexpected breakdown is reduced or even completely eliminated o Parts are ordered when needed and maintenance performed when convenient o Equipment life and there by its utilization is maximized Cons: o Investment costs for implementing the condition-based monitoring (CBM) system o Additional skills might be required to effectively use the CBM system effectively Condition-based Maintenance: Condition based maintenance (CBM) has many similarities to predictive maintenance in that the data gathered during CBM intervals are compared to statistical norms both averages and trends. CBM is more comprehensive than predictive maintenance since it uses both on-line and off-line test data. D. Predictive Analytics(PA) How can we detect faults or failure before the problems happen? This is where maintenance using predictive analytics comes in. Predictive Analytics is the Open Sesame for the world of Big Data. In Forrester report, Big Data Predictive Analytics Solutions, Forrester states: Big data is the fuel and predictive analytics is the engine that firms need to discover, deploy and profit from the knowledge they gain. There has been significant research in the last decade about the introduction of predictive models in Semiconductor Manufacturing. Predictive Analytics can be used to analyze to detect indicators of tool excursions before they happen. Moving from our current reactive state of manufacturing to a proactive state is what the real value of big data capture and analysis will be found. Predictive analytics, coupled with big data technologies, can help the semi-conductor industry to Despite extensive academic research in published literature, applications of data mining methods to semiconductor data are relatively recent. Unfortunately, there are practical limitations in automating the statistical techniques when complex interactions and nonlinearities are involved in the underlying models. Moreover, the large amount of data in current semiconductor databases makes it almost impractical to manually analyze them for valuable decision-making information. This difficulty is mainly due to the large amount of records, which contain hundreds of attributes that need to be simultaneously considered in order to accurately model the system s behavior. The need for understanding complex interaction effects, and automated analysis and discovery tools for extracting useful knowledge from huge amounts of raw data has led to the development of knowledge discovery in databases (KDD) and data mining methodologies [2]. E. Equipment Health Monitoring (EHM)/ Fingerprinting The basic idea was that semiconductor manufacturers could learn a great deal by collecting detailed trace and event information from the equipment to understand the behavior of low-level mechanisms, with the assumption that if the lowlevel mechanisms were exhibiting proper behaviors, then the entire machine would be operating within its specifications. Equipment Health Monitoring the process of assessing the equipment performance and condition based on equipment measurable states. Fingerprinting is a systematic approach that uses equipment component data values which are used to calculate additional parameters that determine the equipment health and performance. Data Mining/Predictive Analytics uses mathematical techniques which are applied to data stored in the factory ASE 2014 ISBN:
5 database searching for indicators of problems or potential issues with the equipment or processing. IV. DATA SOURCES Data is the fundamental basis for all other Predictive and Preventative maintenance functions and applications. Equipment utilization in semiconductor manufacturing fabrication units has always been a challenging issue to us. The overall equipment effectiveness (OEE) being in the range of 40-60% shows a lot of potential to improvement. Over the time lack of availability of data from the equipments had kept us from implementing the predictive maintenance procedures to the existing manufacturing equipment in the fabrication units [1]. Semiconductor manufacturing has always been dataintensive. But the data is simply moved into storage just in case scenario. And of the data that is actually processed, more than 90% is never accessed again. For our study, we started collecting the data from the following data sources: Fault Detection and Classification (FDC) data Trace Log data Equipment Tracking (ET) Data Metrology/Probe/Param Data Equipment Tracking (ET) Data: The Equipment Tracking System (ETS) collects and analyzes the data from various semiconductor equipments. By collecting. Logging equipment is the most important activity in the ETI system. The Equipment Tracking System captures the equipment status and the operation states in an Equipment Tracking database. Using this system, operators log the various event statuses of the equipments, the reason for the current state, and notes about the piece of equipment. The primary task of this system is to correctly reflect tool status. Event data in this system records the Start Time for each operation (i.e., process start or cleaning start), and the Processing Area ID, Wafer # and Lot # associated with each operation. Equipment State Notes Start Time IMHC4C5300 PM_SOURCE Changing 2/24/2014 the 2:27p.m filament IMHC4C5400 PM_SOURCE Changing the filament /23/ :22p.m End Time 2/25/2014 5:35a.m 2/24/2014 8:23 a.m Manufacturing Engineering System Equipment Tracking System Trace Log/Summary Data A. Data Consolidation Data consolidation is the key step in merging all these 4 data sets. Equipment Tracking Trace Summary Metrology Data (Event) Data (FDC) Data (Sigma) Parametric and Electrical data Metrology and Defect data Planning and Recipe data Fault Detection and Classification (FDC) Data: FDC has been recognized in the semiconductor industry as an integral component of advanced process control (APC) framework in improving overall equipment efficiency (OEE). Massive amount of trace or machine data is made available in today s semiconductor industry and fault detection has been one focus of recent efforts to reduce wafer scrap, increase equipment uptime and reduce the usage of test wafers [1-7]. The semiconductor manufacturing processes are usually through FDC to collect a large number of status variable identification (SVID) as data in real-time processes. FDC contains two functions: fault detection and the fault classification. Engineers focused on results of fault detection testing to take some necessary actions. Different fault requires different corrective actions, while fault classification function is classified based on statistics eigenvalues. So engineers can quickly refer to the machine error code and restore the machine to normal state within the least time [4, 6, 8]. Today, the fault-detection production databases in most state-of-theart fabs range from terabytes of data. Combine Event and Trace Data using Time Stamp Event/Trace Data with Wafer ID Add Metrology Info Using Wafer ID Consolidated Data Set of Event/Trace/Metrology ASE 2014 ISBN:
6 Metrology Data: Metrology data records wafer thickness measurement results (mean value and standard deviation) along with wafer # and lot #. In order to apply BN analysis, three separate datasets have to be consolidated and synchronized. In so doing, events data, trace data and metrology data are merged together according to wafer # and time stamps. B. Virtual Equipment Virtual Equipment serves as a testing environment for prediction algorithms prior to their implementation in a semiconductor manufacturing plant (fab). The Virtual Equipment uses input data that is based on historical fab data collected during multiple filament failure cycles. The aim of the Virtual equipment is to test predictive methods in regards to different analysis methods before implementing them in the fab. The virtual equipment is implemented in MATLAB/Simulink. V. CONCEPTUAL BIG DATA ARCHITECTURE Tool Log Data Sensor Events Equipmen t Tracking Wafer-in Equip ment Massively Parallel Processin g Storage (Hadoop) Equip ment Machine Learning Engine Metro logy Consolidate the data Wafer-out HIVE Other steps Predictive Maintenanc e Application For predictive analytics usage, we are looking at another Hadoop software component, the Apache Mahout* machine learning library to create predictive algorithms that can run directly on Hadoop. In our semi-conductor data flow, we have both Structured (RDBMS) and Un-structured (Trace log files). We collect the data from traditional RDBMS-based data warehouse, perform data integration and pre-processing of the structured data as a batch process, and then transfer the data to the Hadoop Cluster. The Hadoop cluster integrates the preprocessed data with relevant unstructured data, and applies predictive models, again as a batch job. The results of the analyses are loaded into the online component of the recommendation system, a NoSQL database. The NoSQL database supports low-latency lookups, which allows results to be queried interactively. There are number of NoSQL databases in big data environment, including Cassadra*, HBase, Pig and MongoDB. B. Not only SQL (NoSQL) databases NoSQL databases relax the constraints of a traditional RDBMS to deliver higher performance and scalability. NoSQL databases can extend the capabilities of Hadoop clusters by providing low-latency object retrieval or other DW-like functionality. More than a hundred different NoSQL databases are available, offering a wide variety of capabilities to address specific use case requirements. The results of the analyses are loaded into the online component of the recommendation system, a NoSQL database. The NoSQL database supports low-latency lookups, which allows results to be queries interactively. We uses a number of NoSQL databases in its big data environment including Cassandra*, HBase, and MongoDB*. C. Memory-Resident Graph Databases Graph databases (GDB) are now a viable alternative to Relational Database Systems, especially in the field of predictive analytics. The semiconductor databases are huge in terms of both number of entries and number of records. For recommendations in predictive analytics, it s important to look at the performance in searching records that are similar to the most recently observed situations, which is necessary to enable to information discovery and inference process on the use of SOMs and BNs. Tool Trace Data HDFS (Hadoop File System) MES MHS Plannin g A. Apache Hadoop* sofware Using Hadoop architecture, we can gather and process data from a variety of sources, including equipment sensors, maintenance logs, trace, FD and equipment tracking. Using that information, the system uses predictive algorithms to determine the likelihood of equipment failure. Engineers and mechanics can then take necessary steps to fix the equipment and prevent catastrophic hardware failure down the line. ASE 2014 ISBN:
7 Graph databases are said to be schema-less. They don t have the relatively rigid structure that we expect in relational databases. Instead, they can store a wide variety of information, from numbers to video and more, organized in a relatively flexible structure described by a changeable graph. Neo4J is only one of many such databases, others that you may encounter include MondoDB, AllegroGraph and FlockDB. The advantages of the graph structure include rapid creation of and changes to a database, and excellent performance for many routine operations. Almost all databases in manufacturing are organized in a sequential manner as lists based usually on the time of arrival of an entry into the database. There is a proposal ( VI. PREDICTIVE MODELS The goals of predictive modeling in semiconductor manufacturing can be inferring equipment condition based on equipment sensor data log, predicting typical process results such as etch bias after plasma etching or source/drain current after ion implantation, or sampling electrical measurement metrology. Different prediction methods are discussed in the below sub-sections. Typically predictive models used would be Neural Networks, Decision Trees or Regression Analysis to arrive at conclusions. Predictive analytic models mine the data and correlate past failures using multivariate analysis. The models can mine all the variables and conditions that contributed to past failures in order to predict future failures. Incoming data are then run through the model and asset health scores are generated on a real time basis. Predictive analytics technologies apply machine-learning algorithms to produce data-driven models of an asset. By definition, predictive analytics encompasses a variety of techniques from statistics, modeling, machine learning, and data mining that analyze current and historical data to predict future events. The processing cycle typically involves two phases of processing: 1. Training phase: Learn a model from training data 2. Predicting phase: Deploy the model to production and use that to predict the unknown or future outcome A. Dynamic Data Clustering Data clustering is a field of active research in machine learning and data mining. Most of the work has focused on static data sets. There has been little work on clustering of dynamic data. We define a dynamic data set as a set of elements whose parameters change over time. Predictive analytics provides a much-needed supplement to predictive maintenance technologies. It applies machine learning to cluster large volumes of multi-variable data. Through the cataloging of data clusters, predictive analytics establishes a comprehensive profile for each asset unique fingerprints left behind during all phases of operation. Those fingerprints provide detailed knowledge of the asset s performance at varying operating rates, with differing products, and under other changing conditions within the production environment. Clusters are based on numerous data inputs that respond to the changing conditions of an individual equipment, and they correspond with the various modes, operating ranges, and products to which the asset is applied. Once cataloged in a knowledge or experience database, clusters associated with asset degradation or other negative attribute trigger alerts. Similarly, the formation of new clusters prompts alerts as the predictive analytics technology identifies new conditions that have yet to be classified. Unlike static and limited input models, the clusters fully account for the asset s condition and recognize both subtle and significant changes in behavior. Dynamically determining the optimal number of clusters in a dataset is a challenging task, since a priori knowledge of the data is required and not always available. Ball and Hall[10] proposed the Iterative Self-Organizing Data Analysis Technique (ISODATA) to dynamically determine the number of clusters in a data set. A similar model to ISODATA is the In the below sections, we are going to review some of the predictive algorithms applied for semiconductor equipment maintenance. B. Random Forests CART (Classification and Regression Trees) For prediction purposes we use a method that utilizes an ensemble of CART models called Random Forests. The aggregation of a large number of different single models usually offers improved prediction accuracy. The methodology allows a transition from a time-based to a condition-based maintenance, a reduction of problem complexity and it offers high predictive performance. As the Random Forest approach ASE 2014 ISBN:
8 is free of parametric or distributional assumptions, the method can be applied to a wide range of predictive maintenance problems. This leads to a reduction of tool downtime, maintenance and manpower costs and improves competitiveness in the semiconductor industry. maintenance personnel. For modeling, MATLAB and a modified version of Kevin Murphy s BNT tool box (Murphy 2001) have been used. C. Multi-Variate Analysis (MVA) Multi-Variate Analysis (MVA) provides a model-centric approach to quickly identifying the complex-system fault source. Usually process variables will go out of correlation before an actual process fault occurs. Thus an MVA system will give early warning. Multivariate analysis of equipment parameter data has proven to be an effective technique for providing additional visibility to equipment performance trends in complicated control schemes. Translation of the multivariate output signal into a concise set of corrective actions will further improve equipment effectiveness. D. Bayesian Networks Bayesian Networks are already widely considered as suitable for the prediction of equipment condition (e.g. see Vachtsevanos et al. 2006) while methods based on decision trees are rather new for prediction purposes. Probabilistic graphical models are based on directed acyclic graphs. Probabilistic graphical models are based on directed acyclic graphs. Within the cognitive science and artificial intelligence such models are known as Bayesian Networks (BN). A BN is a directed acyclic graph whose nodes represent random variables and links define probabilistic dependences between variable. Towards Bayesian network methodology for predicting the equipment health factor of complex semiconductor systems. Due to the fact that all variables (predictors, response) are represented as nodes, and all conditional dependencies between those variables/nodes are represented in (directed) arrows between the nodes, Bayesian Networks are often referred to as a graphical modeling method (Pearl 1988). So models can be set up in a graphical way, defining all nodes and arrows with respect to the real relationships between the predictor and response variables. After model learning, a probabilistic prediction can be made based on new data. For the Predictive Maintenance system described here, a simple network structure was chosen with only few predictors (the first 4 selected variables from Table 1), and the remaining time to failure (TTF) as response variable (Figure 2). The model structure was determined manually utilizing expert knowledge of process and E. Artificial Neural Networks (ANN) Advanced semiconductor manufacturing processes are all made of very sophisticated machines. Requirements on these processes need hundreds of control parameters of the machine [1]. If a slight deviation of the key values changed, it may cause the process deviation (excursion). And then, the production of wafer may be in reduction or even scrapped. For the normal operation and maintenance of equipment to ensure production, failures of the equipments must be diagnosed correctly and timely. The semiconductor manufacturing processes are usually through FDC to collect a large number of status variable identification (SVID) as data in real-time processes. We can evaluate artificial neural networks (ANNs) on how well it s efficiently analyzes SVID data and can provide good results for further controls. Artificial neural back-propagation network has some functions, including possess learning, fault tolerance, and the parallel computing [4, 5]. Applying these functions, artificial neural back-propagation network can develop a predictive method for outliers machines and thus help the overall enhancement of yield in semiconductor manufacturing [6]. The idea of implementing ANN is to have an effective predictive model to detect abnormal values of FDC. Neural networks offer a number of advantages, including less formal statistical training, ability to implicitly detect complex nonlinear relationships between dependent and independent variables, ability to detect all possible interactions between predictor variables, and the availability of multiple training algorithms. Disadvantages include its black box nature, greater computational burden, proneness to overfitting, and the empirical nature of model development [7]. F. Self-Organizing Maps (SOM) In semiconductor industries, one is always faced with large volume of high dimensional data from in-situ sensors, maintenance records, inspections database, etc. Currently, there are a number of methods like Principal Component Analysis (PCA) that have been employed to reduce the dimensionality of the data in order to make it amenable to exploratory analysis. One class of such methods typically projects the data to a low-dimensional space, either linearly or ASE 2014 ISBN:
9 in a non-linear fashion, at the same time preserving their mutual relations as well as possible. The SOM is a set of unique methods that reduce the amount of data by clustering, and reduce data dimensionality through a nonlinear projection of the data onto a low-dimensional space [98]. The methods in this category include principal component analysis, multidimensional scaling, etc. G. Support Vector Machine SVM is a supervised learning method that can be used for classification and regression analysis. It generally involves a training phase that requires health indicators with the corresponding label or equipment condition (good/bad/fault). The objective of using SVM is to use the trained model to predict equipment condition given only the current health condition. H. Health Score Model (Binary logit Model) The Health Score model is based on the linear regression model and measures the likelihood that an asset or process will fail. The health score value, typically referred to simply as the Health Score, can also be used to predict the future health of the asset. The Health Score is presented as a number between 0 and 1. The higher, the number, the healthier the asset. A well-established statistical method for predicting binomial outcomes is required to predict the health score value, and the solution uses a binomial logistic algorithm for this purpose. In the binomial or binary logistic regression, the outcome can have only two possible types of values (e.g. Yes or No, Success or Failure ). Multinomial logistic refers to cases where the outcome can have three or more possible types of values (e.g., good vs. very good vs. best ). Generally outcome is coded as 0 and 1 in binary logistic regression. This kind of algorithm is limited to models where the target field is a flag or binary field. The algorithm provides enhanced statistical output when compared to a multinomial algorithm and is less susceptible to problems when the number of table cells (unique combinations of predictor values) is large relative to the number of records. I. Lifespan Analysis (Cox Regression Model) The Lifespan Analysis model analyzes time-to-failure event data. Lifespan analysis is an offline, back-end process and can be performed at regular intervals or on demand. The model is based on the Cox Regression Model. In many cases where the time to a certain event(such as failure) can be predicted, the Cox Regression technique is particularly well-suited. J. Time Series Models Time series models are used for predicting or forecasting the future behavior of variables. Time series models estimate difference equations containing stochastic components. Two commonly used forms of these models are autoregressive models (AR) and moving average (MA) models. The Box- Jenkins methodology (1976) developed by George Box and G.M. Jenkins combines the AR and MA models to produce the ARMA (autoregressive moving average) model which is the cornerstone of stationary time series analysis. ARIMA(autoregressive integrated moving average models) on the other hand are used to describe non-stationary time series. In recent years time series models have become more sophisticated and attempt to model conditional heteroskedasticity with models such as ARCH (autoregressive conditional heteroskedasticity) and GARCH (generalized autoregressive conditional heteroskedasticity) models frequently used for financial time series. In addition time series models are also used to understand inter-relationships among economic variables represented by systems of equations using VAR (vector autoregression) and structural VAR models. CONCLUSION Explosive data growth in the semiconductor industry will continue to be a major challenge and the ability to manage large data volumes and move the industry from a reactive to a predictive state has the potential to drive significant value for semi-conductor industry. However, little has been published about the use of machine-learning techniques in the manufacturing domain. Predictive maintenance with the advent of big data has reached a crossroads that will provide more intelligence for companies to make predictive maintenance decisions that will avoid the need to make the most costly decision of reactive maintenance. There is a continuous collaborative research efforts going on between semiconductor industries and academia in the areas of Virtual Metrology, Predictive Maintenance, Fault Detection, Run-to-Run control and modeling. There is lot of research going on turning the fault detection with fault prediction to reduce scrap and equipment downtime. Predictive Maintenance will augment preventative maintenance to further reduce downtime. The predictive approach helps prevent unplanned downtime and will help to identify failing parts before they impact wafer quality or cause the tool to fail. It can also extend the useful life of critical parts based on cumulative data. The predictive approach have potentially immense benefits to both semiconductor manufacturers and equipment suppliers, ensuring them to improve the chip quality, increase the yield and extend the useful life of semiconductor equipment. REFERENCES [1] Equipment_Implementation_Guideline.pdf [2] ASE 2014 ISBN:
10 [3] [4] [5] T. Kohonen, Self-Organizing Maps vol. 30: Springer, Heidelberg, 1995 [6] [7] [8] [9] [10] Peter Scheibelhofer,Dietmar Gleispach,Gunter Hayderer and Ernst Stadlober on A Methodology for Predictive Maintenance in Semiconductor Manufaturing, Austrian Journal of Statistics. Volume 41(2012), Number 3, (references) [11] Andreas Mattes, Ulrich Schopka, Martin Schellenberger on A Virtual Equipment for Benchmarking Predictive Maintenance Algorithms, 2012 Winter Simulation Conference [12] Yang Liu on Predictive Modeling for Intelligent Maintenance in Complex Semiconductor Manufacturing Processes [13] J. Clerk Maxwell, A Treatise on Electricity and Magnetism, 3rd ed., vol. 2. Oxford: Clarendon, 1892, pp [14] I.S. Jacobs and C.P. Bean, Fine particles, thin films and exchange anisotropy, in Magnetism, vol. III, G.T. Rado and H. Suhl, Eds. New York: Academic, 1963, pp [15] K. Elissa, Title of paper if known, unpublished. [16] R. Nicole, Title of paper with only first word capitalized, J. Name Stand. Abbrev., in press. [17] Y. Yorozu, M. Hirano, K. Oka, and Y. Tagawa, Electron spectroscopy studies on magneto-optical media and plastic substrate interface, IEEE Transl. J. Magn. Japan, vol. 2, pp , August 1987 [Digests 9th Annual Conf. Magnetics Japan, p. 301, 1982]. [18] M. Young, The Technical Writer s Handbook. Mill Valley, CA: University Science, ASE 2014 ISBN:
How to use Big Data in Industry 4.0 implementations. LAURI ILISON, PhD Head of Big Data and Machine Learning
How to use Big Data in Industry 4.0 implementations LAURI ILISON, PhD Head of Big Data and Machine Learning Big Data definition? Big Data is about structured vs unstructured data Big Data is about Volume
BIG DATA What it is and how to use?
BIG DATA What it is and how to use? Lauri Ilison, PhD Data Scientist 21.11.2014 Big Data definition? There is no clear definition for BIG DATA BIG DATA is more of a concept than precise term 1 21.11.14
Big Data & Analytics for Semiconductor Manufacturing
Big Data & Analytics for Semiconductor Manufacturing 半 導 体 生 産 におけるビッグデータ 活 用 Ryuichiro Hattori 服 部 隆 一 郎 Intelligent SCM and MFG solution Leader Global CoC (Center of Competence) Electronics team General
Big Data. Fast Forward. Putting data to productive use
Big Data Putting data to productive use Fast Forward What is big data, and why should you care? Get familiar with big data terminology, technologies, and techniques. Getting started with big data to realize
Data Refinery with Big Data Aspects
International Journal of Information and Computation Technology. ISSN 0974-2239 Volume 3, Number 7 (2013), pp. 655-662 International Research Publications House http://www. irphouse.com /ijict.htm Data
How To Handle Big Data With A Data Scientist
III Big Data Technologies Today, new technologies make it possible to realize value from Big Data. Big data technologies can replace highly customized, expensive legacy systems with a standard solution
How To Use Neural Networks In Data Mining
International Journal of Electronics and Computer Science Engineering 1449 Available Online at www.ijecse.org ISSN- 2277-1956 Neural Networks in Data Mining Priyanka Gaur Department of Information and
A Methodology for Predictive Failure Detection in Semiconductor Fabrication
A Methodology for Predictive Failure Detection in Semiconductor Fabrication Peter Scheibelhofer (TU Graz) Dietmar Gleispach, Günter Hayderer (austriamicrosystems AG) 09-09-2011 Peter Scheibelhofer (TU
The Data Mining Process
Sequence for Determining Necessary Data. Wrong: Catalog everything you have, and decide what data is important. Right: Work backward from the solution, define the problem explicitly, and map out the data
International Journal of Computer Science Trends and Technology (IJCST) Volume 2 Issue 3, May-Jun 2014
RESEARCH ARTICLE OPEN ACCESS A Survey of Data Mining: Concepts with Applications and its Future Scope Dr. Zubair Khan 1, Ashish Kumar 2, Sunny Kumar 3 M.Tech Research Scholar 2. Department of Computer
Keywords Big Data; OODBMS; RDBMS; hadoop; EDM; learning analytics, data abundance.
Volume 4, Issue 11, November 2014 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Analytics
INTELLIGENT DEFECT ANALYSIS, FRAMEWORK FOR INTEGRATED DATA MANAGEMENT
INTELLIGENT DEFECT ANALYSIS, FRAMEWORK FOR INTEGRATED DATA MANAGEMENT Website: http://www.siglaz.com Abstract Spatial signature analysis (SSA) is one of the key technologies that semiconductor manufacturers
ANALYTICS CENTER LEARNING PROGRAM
Overview of Curriculum ANALYTICS CENTER LEARNING PROGRAM The following courses are offered by Analytics Center as part of its learning program: Course Duration Prerequisites 1- Math and Theory 101 - Fundamentals
Managing Big Data with Hadoop & Vertica. A look at integration between the Cloudera distribution for Hadoop and the Vertica Analytic Database
Managing Big Data with Hadoop & Vertica A look at integration between the Cloudera distribution for Hadoop and the Vertica Analytic Database Copyright Vertica Systems, Inc. October 2009 Cloudera and Vertica
Predictive Analytics. Going from reactive to proactive. Mats Stellwall - Nordic Predictive Analytics Enterprise Architect 2012-06-14
Mats Stellwall - Nordic Predictive Analytics Enterprise Architect 2012-06-14 Predictive Analytics Going from reactive to proactive 2011 IBM Corporation Nothing exists until it is measured Niels Bohr the
Data Warehousing and Data Mining in Business Applications
133 Data Warehousing and Data Mining in Business Applications Eesha Goel CSE Deptt. GZS-PTU Campus, Bathinda. Abstract Information technology is now required in all aspect of our lives that helps in business
A Review of Data Mining Techniques
Available Online at www.ijcsmc.com International Journal of Computer Science and Mobile Computing A Monthly Journal of Computer Science and Information Technology IJCSMC, Vol. 3, Issue. 4, April 2014,
ANALYTICS IN BIG DATA ERA
ANALYTICS IN BIG DATA ERA ANALYTICS TECHNOLOGY AND ARCHITECTURE TO MANAGE VELOCITY AND VARIETY, DISCOVER RELATIONSHIPS AND CLASSIFY HUGE AMOUNT OF DATA MAURIZIO SALUSTI SAS Copyr i g ht 2012, SAS Ins titut
BIG DATA THE NEW OPPORTUNITY
Feature Biswajit Mohapatra is an IBM Certified Consultant and a global integrated delivery leader for IBM s AMS business application modernization (BAM) practice. He is IBM India s competency head for
An Overview of Knowledge Discovery Database and Data mining Techniques
An Overview of Knowledge Discovery Database and Data mining Techniques Priyadharsini.C 1, Dr. Antony Selvadoss Thanamani 2 M.Phil, Department of Computer Science, NGM College, Pollachi, Coimbatore, Tamilnadu,
EFFICIENT DATA PRE-PROCESSING FOR DATA MINING
EFFICIENT DATA PRE-PROCESSING FOR DATA MINING USING NEURAL NETWORKS JothiKumar.R 1, Sivabalan.R.V 2 1 Research scholar, Noorul Islam University, Nagercoil, India Assistant Professor, Adhiparasakthi College
Testing Big data is one of the biggest
Infosys Labs Briefings VOL 11 NO 1 2013 Big Data: Testing Approach to Overcome Quality Challenges By Mahesh Gudipati, Shanthi Rao, Naju D. Mohan and Naveen Kumar Gajja Validate data quality by employing
Transforming the Telecoms Business using Big Data and Analytics
Transforming the Telecoms Business using Big Data and Analytics Event: ICT Forum for HR Professionals Venue: Meikles Hotel, Harare, Zimbabwe Date: 19 th 21 st August 2015 AFRALTI 1 Objectives Describe
The Scientific Data Mining Process
Chapter 4 The Scientific Data Mining Process When I use a word, Humpty Dumpty said, in rather a scornful tone, it means just what I choose it to mean neither more nor less. Lewis Carroll [87, p. 214] In
Approaches for Implementation of Virtual Metrology and Predictive Maintenance into Existing Fab Systems
Workshop - Statistical methods applied in microelectronics 13. June 2011, Catholic University of Milan, Milan, Italy Approaches for Implementation of Virtual Metrology and Predictive Maintenance into Existing
Database Marketing, Business Intelligence and Knowledge Discovery
Database Marketing, Business Intelligence and Knowledge Discovery Note: Using material from Tan / Steinbach / Kumar (2005) Introduction to Data Mining,, Addison Wesley; and Cios / Pedrycz / Swiniarski
Data Management Practices for Intelligent Asset Management in a Public Water Utility
Data Management Practices for Intelligent Asset Management in a Public Water Utility Author: Rod van Buskirk, Ph.D. Introduction Concerned about potential failure of aging infrastructure, water and wastewater
Analytics in the Cloud. Peter Sirota, GM Elastic MapReduce
Analytics in the Cloud Peter Sirota, GM Elastic MapReduce Data-Driven Decision Making Data is the new raw material for any business on par with capital, people, and labor. What is Big Data? Terabytes of
Associate Professor, Department of CSE, Shri Vishnu Engineering College for Women, Andhra Pradesh, India 2
Volume 6, Issue 3, March 2016 ISSN: 2277 128X International Journal of Advanced Research in Computer Science and Software Engineering Research Paper Available online at: www.ijarcsse.com Special Issue
TestScape. On-line, test data management and root cause analysis system. On-line Visibility. Ease of Use. Modular and Scalable.
TestScape On-line, test data management and root cause analysis system On-line Visibility Minimize time to information Rapid root cause analysis Consistent view across all equipment Common view of test
The Predictive Data Mining Revolution in Scorecards:
January 13, 2013 StatSoft White Paper The Predictive Data Mining Revolution in Scorecards: Accurate Risk Scoring via Ensemble Models Summary Predictive modeling methods, based on machine learning algorithms
W H I T E P A P E R. Deriving Intelligence from Large Data Using Hadoop and Applying Analytics. Abstract
W H I T E P A P E R Deriving Intelligence from Large Data Using Hadoop and Applying Analytics Abstract This white paper is focused on discussing the challenges facing large scale data processing and the
Advanced Big Data Analytics with R and Hadoop
REVOLUTION ANALYTICS WHITE PAPER Advanced Big Data Analytics with R and Hadoop 'Big Data' Analytics as a Competitive Advantage Big Analytics delivers competitive advantage in two ways compared to the traditional
International Journal of Advanced Engineering Research and Applications (IJAERA) ISSN: 2454-2377 Vol. 1, Issue 6, October 2015. Big Data and Hadoop
ISSN: 2454-2377, October 2015 Big Data and Hadoop Simmi Bagga 1 Satinder Kaur 2 1 Assistant Professor, Sant Hira Dass Kanya MahaVidyalaya, Kala Sanghian, Distt Kpt. INDIA E-mail: [email protected]
Promises and Pitfalls of Big-Data-Predictive Analytics: Best Practices and Trends
Promises and Pitfalls of Big-Data-Predictive Analytics: Best Practices and Trends Spring 2015 Thomas Hill, Ph.D. VP Analytic Solutions Dell Statistica Overview and Agenda Dell Software overview Dell in
Volume 3, Issue 6, June 2015 International Journal of Advance Research in Computer Science and Management Studies
Volume 3, Issue 6, June 2015 International Journal of Advance Research in Computer Science and Management Studies Research Article / Survey Paper / Case Study Available online at: www.ijarcsms.com Image
Data Mining for Manufacturing: Preventive Maintenance, Failure Prediction, Quality Control
Data Mining for Manufacturing: Preventive Maintenance, Failure Prediction, Quality Control Andre BERGMANN Salzgitter Mannesmann Forschung GmbH; Duisburg, Germany Phone: +49 203 9993154, Fax: +49 203 9993234;
Data Mining Solutions for the Business Environment
Database Systems Journal vol. IV, no. 4/2013 21 Data Mining Solutions for the Business Environment Ruxandra PETRE University of Economic Studies, Bucharest, Romania [email protected] Over
Modern IT Operations Management. Why a New Approach is Required, and How Boundary Delivers
Modern IT Operations Management Why a New Approach is Required, and How Boundary Delivers TABLE OF CONTENTS EXECUTIVE SUMMARY 3 INTRODUCTION: CHANGING NATURE OF IT 3 WHY TRADITIONAL APPROACHES ARE FAILING
Nine Common Types of Data Mining Techniques Used in Predictive Analytics
1 Nine Common Types of Data Mining Techniques Used in Predictive Analytics By Laura Patterson, President, VisionEdge Marketing Predictive analytics enable you to develop mathematical models to help better
ANALYTICS STRATEGY: creating a roadmap for success
ANALYTICS STRATEGY: creating a roadmap for success Companies in the capital and commodity markets are looking at analytics for opportunities to improve revenue and cost savings. Yet, many firms are struggling
Data Isn't Everything
June 17, 2015 Innovate Forward Data Isn't Everything The Challenges of Big Data, Advanced Analytics, and Advance Computation Devices for Transportation Agencies. Using Data to Support Mission, Administration,
Master big data to optimize the oil and gas lifecycle
Viewpoint paper Master big data to optimize the oil and gas lifecycle Information management and analytics (IM&A) helps move decisions from reactive to predictive Table of contents 4 Getting a handle on
Big Data and Data Science: Behind the Buzz Words
Big Data and Data Science: Behind the Buzz Words Peggy Brinkmann, FCAS, MAAA Actuary Milliman, Inc. April 1, 2014 Contents Big data: from hype to value Deconstructing data science Managing big data Analyzing
not possible or was possible at a high cost for collecting the data.
Data Mining and Knowledge Discovery Generating knowledge from data Knowledge Discovery Data Mining White Paper Organizations collect a vast amount of data in the process of carrying out their day-to-day
A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM
A REVIEW PAPER ON THE HADOOP DISTRIBUTED FILE SYSTEM Sneha D.Borkar 1, Prof.Chaitali S.Surtakar 2 Student of B.E., Information Technology, J.D.I.E.T, [email protected] Assistant Professor, Information
Lambda Architecture. Near Real-Time Big Data Analytics Using Hadoop. January 2015. Email: [email protected] Website: www.qburst.com
Lambda Architecture Near Real-Time Big Data Analytics Using Hadoop January 2015 Contents Overview... 3 Lambda Architecture: A Quick Introduction... 4 Batch Layer... 4 Serving Layer... 4 Speed Layer...
locuz.com Big Data Services
locuz.com Big Data Services Big Data At Locuz, we help the enterprise move from being a data-limited to a data-driven one, thereby enabling smarter, faster decisions that result in better business outcome.
Customer Classification And Prediction Based On Data Mining Technique
Customer Classification And Prediction Based On Data Mining Technique Ms. Neethu Baby 1, Mrs. Priyanka L.T 2 1 M.E CSE, Sri Shakthi Institute of Engineering and Technology, Coimbatore 2 Assistant Professor
Azure Machine Learning, SQL Data Mining and R
Azure Machine Learning, SQL Data Mining and R Day-by-day Agenda Prerequisites No formal prerequisites. Basic knowledge of SQL Server Data Tools, Excel and any analytical experience helps. Best of all:
Moving Large Data at a Blinding Speed for Critical Business Intelligence. A competitive advantage
Moving Large Data at a Blinding Speed for Critical Business Intelligence A competitive advantage Intelligent Data In Real Time How do you detect and stop a Money Laundering transaction just about to take
Managing Cloud Server with Big Data for Small, Medium Enterprises: Issues and Challenges
Managing Cloud Server with Big Data for Small, Medium Enterprises: Issues and Challenges Prerita Gupta Research Scholar, DAV College, Chandigarh Dr. Harmunish Taneja Department of Computer Science and
Industrial Roadmap for Connected Machines. Sal Spada Research Director ARC Advisory Group [email protected]
Industrial Roadmap for Connected Machines Sal Spada Research Director ARC Advisory Group [email protected] Industrial Internet of Things (IoT) Based upon enhanced connectivity of this stuff Connecting
WHITE PAPER ON. Operational Analytics. HTC Global Services Inc. Do not copy or distribute. www.htcinc.com
WHITE PAPER ON Operational Analytics www.htcinc.com Contents Introduction... 2 Industry 4.0 Standard... 3 Data Streams... 3 Big Data Age... 4 Analytics... 5 Operational Analytics... 6 IT Operations Analytics...
Proactive Asset Management with IIoT and Analytics
Proactive Asset Management with IIoT and Analytics by Ralph Rio in Industrial Internet of Things, Analytics & Big Data Summary The Industrial Internet of Things (IIoT) with advanced analytics, offers new
Enterprise Asset Performance Management
Application Solution Enterprise Asset Performance Management for Power Utilities Using the comprehensive Enterprise Asset Performance Management solution offered by Schneider Electric, power utilities
DATA MINING AND WAREHOUSING CONCEPTS
CHAPTER 1 DATA MINING AND WAREHOUSING CONCEPTS 1.1 INTRODUCTION The past couple of decades have seen a dramatic increase in the amount of information or data being stored in electronic format. This accumulation
A Brief Introduction to Apache Tez
A Brief Introduction to Apache Tez Introduction It is a fact that data is basically the new currency of the modern business world. Companies that effectively maximize the value of their data (extract value
PRACTICAL DATA MINING IN A LARGE UTILITY COMPANY
QÜESTIIÓ, vol. 25, 3, p. 509-520, 2001 PRACTICAL DATA MINING IN A LARGE UTILITY COMPANY GEORGES HÉBRAIL We present in this paper the main applications of data mining techniques at Electricité de France,
Gerard Mc Nulty Systems Optimisation Ltd [email protected]/0876697867 BA.,B.A.I.,C.Eng.,F.I.E.I
Gerard Mc Nulty Systems Optimisation Ltd [email protected]/0876697867 BA.,B.A.I.,C.Eng.,F.I.E.I Data is Important because it: Helps in Corporate Aims Basis of Business Decisions Engineering Decisions Energy
Big Data Analytics and Decision Analysis for Manufacturing Intelligence to Empower Industry 3.5
ISMI2015, Oct. 16-18, 2015 KAIST, Daejeon, South Korea Big Data Analytics and Decision Analysis for Manufacturing Intelligence to Empower Industry 3.5 Tsinghua Chair Professor Chen-Fu Chien, Ph.D. Department
Advanced In-Database Analytics
Advanced In-Database Analytics Tallinn, Sept. 25th, 2012 Mikko-Pekka Bertling, BDM Greenplum EMEA 1 That sounds complicated? 2 Who can tell me how best to solve this 3 What are the main mathematical functions??
INTELLIGENT DEFECT ANALYSIS SOFTWARE
INTELLIGENT DEFECT ANALYSIS SOFTWARE Website: http://www.siglaz.com Semiconductor fabs currently use defect count or defect density as a triggering mechanism for their Statistical Process Control. However,
ANALYTICS BUILT FOR INTERNET OF THINGS
ANALYTICS BUILT FOR INTERNET OF THINGS Big Data Reporting is Out, Actionable Insights are In In recent years, it has become clear that data in itself has little relevance, it is the analysis of it that
Big Data Analytics. An Introduction. Oliver Fuchsberger University of Paderborn 2014
Big Data Analytics An Introduction Oliver Fuchsberger University of Paderborn 2014 Table of Contents I. Introduction & Motivation What is Big Data Analytics? Why is it so important? II. Techniques & Solutions
Data Mining Algorithms Part 1. Dejan Sarka
Data Mining Algorithms Part 1 Dejan Sarka Join the conversation on Twitter: @DevWeek #DW2015 Instructor Bio Dejan Sarka ([email protected]) 30 years of experience SQL Server MVP, MCT, 13 books 7+ courses
Oracle Big Data SQL Technical Update
Oracle Big Data SQL Technical Update Jean-Pierre Dijcks Oracle Redwood City, CA, USA Keywords: Big Data, Hadoop, NoSQL Databases, Relational Databases, SQL, Security, Performance Introduction This technical
itesla Project Innovative Tools for Electrical System Security within Large Areas
itesla Project Innovative Tools for Electrical System Security within Large Areas Samir ISSAD RTE France [email protected] PSCC 2014 Panel Session 22/08/2014 Advanced data-driven modeling techniques
Are You Ready for Big Data?
Are You Ready for Big Data? Jim Gallo National Director, Business Analytics February 11, 2013 Agenda What is Big Data? How do you leverage Big Data in your company? How do you prepare for a Big Data initiative?
Practical Data Science with Azure Machine Learning, SQL Data Mining, and R
Practical Data Science with Azure Machine Learning, SQL Data Mining, and R Overview This 4-day class is the first of the two data science courses taught by Rafal Lukawiecki. Some of the topics will be
Prediction of Stock Performance Using Analytical Techniques
136 JOURNAL OF EMERGING TECHNOLOGIES IN WEB INTELLIGENCE, VOL. 5, NO. 2, MAY 2013 Prediction of Stock Performance Using Analytical Techniques Carol Hargreaves Institute of Systems Science National University
The 4 Pillars of Technosoft s Big Data Practice
beyond possible Big Use End-user applications Big Analytics Visualisation tools Big Analytical tools Big management systems The 4 Pillars of Technosoft s Big Practice Overview Businesses have long managed
Sanjeev Kumar. contribute
RESEARCH ISSUES IN DATAA MINING Sanjeev Kumar I.A.S.R.I., Library Avenue, Pusa, New Delhi-110012 [email protected] 1. Introduction The field of data mining and knowledgee discovery is emerging as a
How In-Memory Data Grids Can Analyze Fast-Changing Data in Real Time
SCALEOUT SOFTWARE How In-Memory Data Grids Can Analyze Fast-Changing Data in Real Time by Dr. William Bain and Dr. Mikhail Sobolev, ScaleOut Software, Inc. 2012 ScaleOut Software, Inc. 12/27/2012 T wenty-first
A Case Study in Integrated Quality Assurance for Performance Management Systems
A Case Study in Integrated Quality Assurance for Performance Management Systems Liam Peyton, Bo Zhan, Bernard Stepien School of Information Technology and Engineering, University of Ottawa, 800 King Edward
The Rise of Industrial Big Data
GE Intelligent Platforms The Rise of Industrial Big Data Leveraging large time-series data sets to drive innovation, competitiveness and growth capitalizing on the big data opportunity The Rise of Industrial
NEURAL NETWORKS IN DATA MINING
NEURAL NETWORKS IN DATA MINING 1 DR. YASHPAL SINGH, 2 ALOK SINGH CHAUHAN 1 Reader, Bundelkhand Institute of Engineering & Technology, Jhansi, India 2 Lecturer, United Institute of Management, Allahabad,
ANALYTICS IN BIG DATA ERA
ANALYTICS IN BIG DATA ERA ANALYTICS TECHNOLOGY AND ARCHITECTURE TO MANAGE VELOCITY AND VARIETY, DISCOVER RELATIONSHIPS AND CLASSIFY HUGE AMOUNT OF DATA MAURIZIO SALUSTI SAS Copyr i g ht 2012, SAS Ins titut
BIG DATA TRENDS AND TECHNOLOGIES
BIG DATA TRENDS AND TECHNOLOGIES THE WORLD OF DATA IS CHANGING Cloud WHAT IS BIG DATA? Big data are datasets that grow so large that they become awkward to work with using onhand database management tools.
Data Virtualization A Potential Antidote for Big Data Growing Pains
perspective Data Virtualization A Potential Antidote for Big Data Growing Pains Atul Shrivastava Abstract Enterprises are already facing challenges around data consolidation, heterogeneity, quality, and
Big Data, Physics, and the Industrial Internet! How Modeling & Analytics are Making the World Work Better."
Big Data, Physics, and the Industrial Internet! How Modeling & Analytics are Making the World Work Better." Matt Denesuk! Chief Data Science Officer! GE Software! October 2014! Imagination at work. Contact:
Developing Scalable Smart Grid Infrastructure to Enable Secure Transmission System Control
Developing Scalable Smart Grid Infrastructure to Enable Secure Transmission System Control EP/K006487/1 UK PI: Prof Gareth Taylor (BU) China PI: Prof Yong-Hua Song (THU) Consortium UK Members: Brunel University
Manufacturing Analytics: Uncovering Secrets on Your Factory Floor
SIGHT MACHINE WHITE PAPER Manufacturing Analytics: Uncovering Secrets on Your Factory Floor Quick Take For manufacturers, operational insight is often masked by mountains of process and part data flowing
How To Make Data Streaming A Real Time Intelligence
REAL-TIME OPERATIONAL INTELLIGENCE Competitive advantage from unstructured, high-velocity log and machine Big Data 2 SQLstream: Our s-streaming products unlock the value of high-velocity unstructured log
BIG DATA & ANALYTICS. Transforming the business and driving revenue through big data and analytics
BIG DATA & ANALYTICS Transforming the business and driving revenue through big data and analytics Collection, storage and extraction of business value from data generated from a variety of sources are
ISSN: 2320-1363 CONTEXTUAL ADVERTISEMENT MINING BASED ON BIG DATA ANALYTICS
CONTEXTUAL ADVERTISEMENT MINING BASED ON BIG DATA ANALYTICS A.Divya *1, A.M.Saravanan *2, I. Anette Regina *3 MPhil, Research Scholar, Muthurangam Govt. Arts College, Vellore, Tamilnadu, India Assistant
Cloudera Enterprise Data Hub in Telecom:
Cloudera Enterprise Data Hub in Telecom: Three Customer Case Studies Version: 103 Table of Contents Introduction 3 Cloudera Enterprise Data Hub for Telcos 4 Cloudera Enterprise Data Hub in Telecom: Customer
The Semiconductor Industry: Out in Front, but Lagging Behind Tom Mariano Published September, 2014
As seen in The Semiconductor Industry: Out in Front, but Lagging Behind Tom Mariano Published September, 2014 Capital equipment suppliers must provide advanced analytical systems that leverage data generated
Loss Prevention Data Mining Using big data, predictive and prescriptive analytics to enpower loss prevention
White paper Loss Prevention Data Mining Using big data, predictive and prescriptive analytics to enpower loss prevention Abstract In the current economy where growth is stumpy and margins reduced, retailers
Bayesian networks - Time-series models - Apache Spark & Scala
Bayesian networks - Time-series models - Apache Spark & Scala Dr John Sandiford, CTO Bayes Server Data Science London Meetup - November 2014 1 Contents Introduction Bayesian networks Latent variables Anomaly
Introduction to Data Mining
Introduction to Data Mining Jay Urbain Credits: Nazli Goharian & David Grossman @ IIT Outline Introduction Data Pre-processing Data Mining Algorithms Naïve Bayes Decision Tree Neural Network Association
A Study Of Bagging And Boosting Approaches To Develop Meta-Classifier
A Study Of Bagging And Boosting Approaches To Develop Meta-Classifier G.T. Prasanna Kumari Associate Professor, Dept of Computer Science and Engineering, Gokula Krishna College of Engg, Sullurpet-524121,
The evolution of database technology (II) Huibert Aalbers Senior Certified Executive IT Architect
The evolution of database technology (II) Huibert Aalbers Senior Certified Executive IT Architect IT Insight podcast This podcast belongs to the IT Insight series You can subscribe to the podcast through
The Internet of Things and Big Data: Intro
The Internet of Things and Big Data: Intro John Berns, Solutions Architect, APAC - MapR Technologies April 22 nd, 2014 1 What This Is; What This Is Not It s not specific to IoT It s not about any specific
Integrating a Big Data Platform into Government:
Integrating a Big Data Platform into Government: Drive Better Decisions for Policy and Program Outcomes John Haddad, Senior Director Product Marketing, Informatica Digital Government Institute s Government
Higher Profits from Intelligent Semiconductor-Equipment Maintenance By Steven R. Montgomery President of SST, a division of Woodhead Industries
Higher Profits from Intelligent Semiconductor-Equipment Maintenance By Steven R. Montgomery President of SST, a division of Woodhead Industries Fabrication tool downtime means corporate profitability down
Introduction. Background
Predictive Operational Analytics (POA): Customized Solutions for Improving Efficiency and Productivity for Manufacturers using a Predictive Analytics Approach Introduction Preserving assets and improving
Comparison of K-means and Backpropagation Data Mining Algorithms
Comparison of K-means and Backpropagation Data Mining Algorithms Nitu Mathuriya, Dr. Ashish Bansal Abstract Data mining has got more and more mature as a field of basic research in computer science and
