Big Data Analytics for Predictive Manufacturing Control A Case Study from Process Industry



Similar documents
How To Realize The Predictive Enterprise Through Intelligent Process Predictions Based On Big Data Analytics

Data Mining for Manufacturing: Preventive Maintenance, Failure Prediction, Quality Control

Adaptive demand planning in a volatile business environment

How To Handle Big Data With A Data Scientist

GOZINTOGRAPHS FOR BY-PRODUCTS AND CYCLIC PRODUCTION: AN APPROACH FOR ERP SYSTEM APPLICATION

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

CLOUD BASED SEMANTIC EVENT PROCESSING FOR

International Journal of Engineering Research ISSN: & Management Technology November-2015 Volume 2, Issue-6

ANALYTICS STRATEGY: creating a roadmap for success

EL Program: Smart Manufacturing Systems Design and Analysis

Gerard Mc Nulty Systems Optimisation Ltd BA.,B.A.I.,C.Eng.,F.I.E.I

Social Innovation through Utilization of Big Data

Chapter 2 Why Are Enterprise Applications So Diverse?

Business Intelligence. A Presentation of the Current Lead Solutions and a Comparative Analysis of the Main Providers

Process-Family-Points

WHITE PAPER ON. Operational Analytics. HTC Global Services Inc. Do not copy or distribute.

How to use Big Data in Industry 4.0 implementations. LAURI ILISON, PhD Head of Big Data and Machine Learning

Big Data Integration: A Buyer's Guide

CONNECTING DATA WITH BUSINESS

jeti: A Tool for Remote Tool Integration

Tapping the benefits of business analytics and optimization

How To Understand Business Intelligence

The Big Data methodology in computer vision systems

A Study on Data Analysis Process Management System in MapReduce using BPM

A Big Data Analytical Framework For Portfolio Optimization Abstract. Keywords. 1. Introduction

Managing and Tracing the Traversal of Process Clouds with Templates, Agendas and Artifacts

Industry 4.0 and Big Data

Business Intelligence Systems

Integrating a Big Data Platform into Government:

A Hurwitz white paper. Inventing the Future. Judith Hurwitz President and CEO. Sponsored by Hitachi

An Integrated Quality Assurance Framework for Specifying Business Information Systems

A Software Framework for Risk-Aware Business Process Management

Advanced Big Data Analytics with R and Hadoop

2 Types of Complex Event Processing

Master big data to optimize the oil and gas lifecycle

ICT Perspectives on Big Data: Well Sorted Materials

Why big data? Lessons from a Decade+ Experiment in Big Data

Process Modelling from Insurance Event Log

International Journal of Advanced Engineering Research and Applications (IJAERA) ISSN: Vol. 1, Issue 6, October Big Data and Hadoop

An Event Processing Platform for Business Process Management

A Framework for Software Product Line Engineering

Distributed Framework for Data Mining As a Service on Private Cloud

ANALYTICS BUILT FOR INTERNET OF THINGS

From Big Data to Smart Data Thomas Hahn

Introduction. Background

This Symposium brought to you by

Steel supply chain transformation challenges Key learnings

INTERNET OF THINGS IN STEEL MANUFACTURING

Oracle Real Time Decisions

Sales and Operations Planning in Company Supply Chain Based on Heuristics and Data Warehousing Technology

Process Mining in Big Data Scenario

INTRUSION PREVENTION AND EXPERT SYSTEMS

Towards an Integration of Business Process Modeling and Object-Oriented Software Development

INTERNATIONAL JOURNAL OF PURE AND APPLIED RESEARCH IN ENGINEERING AND TECHNOLOGY

Advanced Analytics. The Way Forward for Businesses. Dr. Sujatha R Upadhyaya

Towards a Thriving Data Economy: Open Data, Big Data, and Data Ecosystems

Predictive Analytics. Going from reactive to proactive. Mats Stellwall - Nordic Predictive Analytics Enterprise Architect

Distributed Database for Environmental Data Integration

Apigee Insights Increase marketing effectiveness and customer satisfaction with API-driven adaptive apps

Find the Information That Matters. Visualize Your Data, Your Way. Scalable, Flexible, Global Enterprise Ready

Leveraging Information For Smarter Business Outcomes With IBM Information Management Software

Miracle Integrating Knowledge Management and Business Intelligence

BPM: new architecture driven by Business Process Planning and Control (BPPC)

Making Big Data Analytics Fast and Easy Big Data Not Delivering? Context is the key.

IMPROVING BUSINESS PROCESS MODELING USING RECOMMENDATION METHOD

META DATA QUALITY CONTROL ARCHITECTURE IN DATA WAREHOUSING

Delivering new insights and value to consumer products companies through big data

FUTURE RESEARCH DIRECTIONS OF SOFTWARE ENGINEERING AND KNOWLEDGE ENGINEERING *

Information Visualization WS 2013/14 11 Visual Analytics

Big Data, Physics, and the Industrial Internet! How Modeling & Analytics are Making the World Work Better."

Research Problems Associated with Big Data Utilization in Logistics and Supply Chains Design and Management

BIG DATA STRATEGY. Rama Kattunga Chair at American institute of Big Data Professionals. Building Big Data Strategy For Your Organization

A Guide Through the BPM Maze

Data Mining Analysis of a Complex Multistage Polymer Process

Improving proposal evaluation process with the help of vendor performance feedback and stochastic optimal control

Industrial Roadmap for Connected Machines. Sal Spada Research Director ARC Advisory Group

IMPLEMENTING A SPC INTEGRATED SYSTEM TO IMPROVE MANUFACTURING PROCESSES IN AUTOMOTIVE INDUSTRY

Handling Big(ger) Logs: Connecting ProM 6 to Apache Hadoop

Big Data. Fast Forward. Putting data to productive use

COMPUTER INTEGRATED MANUFACTURING

Software solutions for manufacturing operations management. Helping manufacturers optimize the Digital Enterprise and realize innovation

Investor Presentation. Second Quarter 2015

Enabling Self Organising Logistics on the Web of Things

Comparison of Request Admission Based Performance Isolation Approaches in Multi-tenant SaaS Applications

An Agent-Based Concept for Problem Management Systems to Enhance Reliability

Comparative Analysis of the Main Business Intelligence Solutions

Performance Prediction for Software Architectures

Data Isn't Everything

Using reporting and data mining techniques to improve knowledge of subscribers; applications to customer profiling and fraud management

TRENDS IN THE DEVELOPMENT OF BUSINESS INTELLIGENCE SYSTEMS

USING BIG DATA FOR INTELLIGENT BUSINESSES

Dotted Chart and Control-Flow Analysis for a Loan Application Process

Data Integration Models for Operational Data Warehousing

BPM for Structural Integrity Management in Oil and Gas Industry

Dimensions of Business Process Intelligence

Framework model on enterprise information system based on Internet of things

EVERYTHING THAT MATTERS IN ADVANCED ANALYTICS

Formal Methods for Preserving Privacy for Big Data Extraction Software

Easily Identify Your Best Customers

Data Virtualization A Potential Antidote for Big Data Growing Pains

Transcription:

2014 IEEE International Congress on Big Data Big Data Analytics for Predictive Manufacturing Control A Case Study from Process Industry Julian Krumeich, Dirk Werth, Peter Loos Institute for Information Systems German Research Center for Artificial Intelligence Saarbrücken, Germany {Julian.Krumeich, Dirk.Werth, Peter.Loos}@dfki.de Sven Jacobi Information and Communication Technology Saarstahl AG Völklingen, Germany sven.jacobi@saarstahl.com Abstract Nowadays, companies are more than ever forced to dynamically adapt their business process executions to currently existing business situations in order to keep up with increasing market demands in global competition. Companies that are able to analyze the current state of their processes, forecast its most optimal progress and proactively control them based on reliable predictions will be a decisive step ahead competitors. The paper at hand exploits potentials through predictive analytics on big data aiming at event-based predictions and thereby enabling proactive control of business processes. In doing so, the paper particularly focus production processes in analytical process manufacturing industries and outlines based on a case study at Saarstahl AG, a large German steel producing company which production-related data is currently collected forming a potential foundation for accurate forecasts. However, without dedicated approaches of big data analytics, the sample company cannot utilize the potential of already available data for a proactive process control. Hence, the article forms a working and discussion basis for further research in big data analytics by proposing a general system architecture. Keywords-business process forecast and simulation; predictive analytics; complex event processing; business process intelligence; event-driven business process management; eventbased predictions; process industry I. GENERAL INTRODUCTION A. Vision of the Predictive Enterprise Companies are increasingly facing the challenge to individualize their business processes due to growing market pressures in today s globalized world. To be ahead of competitors, every single business process execution must be adapted to currently existing process and business situations. The rising digitalization of the real world, driven by the era of Internet of Things, allows for an unprecedented insight into current business process situations [1]. In future, only those companies able to analyze their business operations based on this rapidly growing mass of data, predict the best proceeding process flow, and proactively control their processes with this knowledge will survive in highly competitive markets. Such a company sketches the vision of a "Predictive Enterprise" as the next stage in the evolution of real-time enterprises within the age of data as a decisive competitive asset [1], [2]. B. Motivation and Problem Statement Today's corporate practice is beyond the realization of this vision. The potential of massive amounts of data in companies, which can already be collected by state of the art sensor technologies, has only received scant attention in business process management (BPM) [3]. Frequently, this is due to the lack of technical capabilities to analyze such big data in a timely manner and to derive the right decisions. However, business analysts agree that companies are going to face existential difficulties, if they are not able to discover or even anticipate problems and defects in business process executions in time [4]. Thus, BPM software has to provide means to analyze process executions, forecast their further progress and consequently identify potential problems early enough to proactively enact appropriate countermeasures. Yet, current approaches and techniques in predictive analytics rarely take into account context situations, e.g. through the utilization of technologies like Complex Event Processing (CEP), as a basis to derive forecasts. Most predictions remain purely stochastic involving the risk of computing highly-likely forecasts of process progressions that turn out to be entirely unlikely in the current existing, perhaps exceptional, situation. This calls for a dedicated consideration of given situations respectively events to compute reliable forecasts. Even though available data sets will grow continuously owing to enhanced sensing capabilities and provide unprecedented insights into process situations, this entails the need for dedicated big data analytics solutions able to exploit the full potential out of such data. C. Research Contribution and Applied Research Method The paper at hand proposes and examines the concept of event-based process predictions and outlines its potentials for planning, forecasting and eventually controlling business processes (see sec. III and for preliminary work [5]). By highlighting how the application of such event-based predictions can make manufacturing processes more efficient, the paper put particular emphasis on the analytical process industry. In doing so, this specific type of industry as well as how production planning and control is currently conducted in this domain will be characterized at first (see sec. II). After proposing the general concept of event-based predictions, the paper analyze by means of a case study (see 978-1-4799-5057-7/14 $31.00 2014 IEEE DOI 10.1109/BigData.Congress.2014.83 543 530

sec. IV and for preliminary work [5]), which process and context data can a sample steel producing company currently collect by its applied sensor technology forming a potential basis for event-based predictions of their manufacturing processes. Hence, the paper follows case study research, which has been applied in Information Systems research for almost two decades [6]. In particular, the paper employs the methodology proposed by Benbasat et al. [7]: as the unit of analysis the steel bar production line at Saarstahl AG, a major German steel producer, was chosen. The two research questions applied to the case study are What type of data is currently available in industry processes using state-of-theart sensor technology to realize event-based predictions? and Why is it a Big Data challenge to analyze this data appropriately? The results of the selected case study are considered to be generalizable to other processmanufacturing enterprises. Since the chosen research design is a single-case study, it is particular appropriate to revelatory cases, which is typical for big data analytics cases as a relatively new phenomenon in information systems research. The case study data was collected and analyzed by the central department of information and communication technology of the company. As data selection methods interview techniques were applied and physical artefact sensor networks investigated. Based on the examined data, the paper concludes that with current state of the art (see sec. VI), the available data cannot be analyzed in a reasonable time horizon in order to make sufficient business value out of it. Without dedicated big data analytics, the sample company cannot exploit the potential of event-based predictions based on its data. For this, the paper proposes a general system architecture (see sec. V) intending to form a working and discussion basis for further research efforts in big data analytics (see sec. VII). II. PRODUCTION PROCESSES IN THE ANALYTICAL PROCESS MANUFACTURING INDUSTRY A. Characterization of Analytical Process Manufacturing Manufacturing companies can be differentiated by the number and characteristic of input factors they process into output factors. Two distinct types can be distinguished [8]: on the one hand, there are manufacturing companies producing with only a small number of input factors, many end products and on the other hand those ones generating only some end product with many input factors by a high level of vertical integration. Usually, mixed forms of both extremes exist. In case input and/or output factors can be quantifiable, manufacturing processes are considered as discrete ones. If gases or liquids are processed, manufacturers are assigned to the process industry. Besides this broad consideration of companies type of operation, manufacturing processes can be additionally classified into four types (cf. Fig. 1) [9]: continuous manufacturing processes are characterized by transforming one input factor into one output factor or processing it further. Analytical manufacturing processes feature 1:N relationships, i.e. one input factor is processed into several so-called co-products, which can be often found in the chemical industry. N:1 relationships are characteristic for synthetic manufacturing processes, as they are typical for the automotive assembly. Finally, there exist various mixed forms according to N:M relationships. Co-products, which incur by analytical manufacturing processes, are to be considered in terms of their economic importance. [10] distinguishes between cost and revenueneutral, cost-generating, but revenue neutral as well as revenue increasing co-products. From a production logistics point of view, co-products of the latter type should be particularly examined. If co-products of the same type, but different level of quality result in the analytical production, they are either to be homogenized by suitable processing steps (this may be a continuous manufacturing sub-process according to Fig. 1a) or they are considered as suitable products of varying quality gradations. B. Production Planning and Control in the Analytical Process Industry Analytical manufacturing is characteristic for production processes in industries such as the chemical or the steel producing one. In addition to main products, various types of co- and by-products of different quality and goodness result that need to be dynamically considered in the ongoing production planning and control [11]. Often, resulting co-products cyclically flow into the manufacturing process or are continuously required for additional production lines. These cyclical relations between output and input factors are diametrically opposed to the classic manufacturing industry, such as the automotive. Due to fluctuations in production as they typically occur in the process industry an accurate and timely planning and control of production processes is further complicated [8]. These fluctuations have different reasons: they can be the result varying quality of raw materials, external influences, such as temperature or pressure as well as internal influences, such as reaction rates, e.g. in the chemical industry [8]. Figure 1. Existing forms if manufacturing processes 544 531

A look at steel manufacturing companies as an example for process manufacturers reveals that resulting end products are tremendously customer-specific in terms of their levels of quality; standard products only rarely exist. This is in contrast to discrete manufacturing. Even if for example in the automotive industry, complexity is given due to the wide range of variations, there is no heterogeneity in terms of the actual product quality, which is to be measured and the logistical process flow controlled accordingly. The individual material properties as per specific customer needs forces process manufacturing companies to individually analyze and control each of their processes. For example, while it is not possible in the production of automobiles to figuratively build in air conditions instead of flawed engines, quality gradations of products are common practice in steelmaking, depending on the subsequent application, and serve as valid intermediate, by-, co- or end products. If the target deviation of a co-product is too high for customer order A, in which high-precision steel for aircraft or aerospace industry is needed, it can instead be used for customer order Z, in which steel for the building industry is needed. To avoid excess capacities, to reduce the demand for raw materials and to get free production capacities, a re-scheduling should be taken into account early in production planning, so that required co-products for customer order Z do not have to be scheduled separately. Thus, the production in the process manufacturing industry is neither characterized by strong linearity (as in discrete manufacturing), nor is the quality of resulting coproducts easily detectable. Manufacturing processes can thus be considered as strongly interweaved [10]. From an information systems point of view, the optimization of production processes in analytical process manufacturing is consequently a particular challenge [10]. Hence, choosing this type of manufacturing as the underlying object of investigation proves to be particularly appropriate to elucidate event-based predictions and illustrate its advantages for planning and control of manufacturing processes. III. PLANNING AND CONTROL OF ANALYTICAL MANUFACTURING PROCESSES USING EVENT-BASED PREDICTIONS In enterprise resource planning (ERP) systems, forecastbased methods have been applied for years to determine production demands [12]. While independent requirements are usually calculated based on bills of materials, stochastic calculations are applied to determine dependent requirements. A forecast-based control of manufacturing processes had been hardly applied in the past. One reason for that was based on the fact that forecasts had frequently tend to be too inaccurate, since datasets required for computing predictions were too small to cover the operational context situation sufficiently. While in principle, it was possible to expand the data base, this information gathering has proven to be too complex, too expensive and not accomplishable in a timely manner. Therefore the independent requirements planning and hence the higher-level production planning and control were deterministically handled so far. Today, companies still perform insufficient analyzes and forecasts based on their collected sensor data, although this could have an immense impact on their economic and ecological performance [3]. This is in contrast to industries such as insurance or banking that have fully implemented predictive analytics in their business models [13]. In the course of recent technological progress in the fields of "Internet of Things" and "Cyber-Physical Systems" it is now possible to accumulate production processes relatively cost-neutral by sensors. This allows to measure internal and external parameters of production processes in a previously unprecedented level of detail, even in real time. Thereby the technical foundation is created to establish and continuously enrich a data base allowing for predictions to be used even for highly-critical planning purposes. From the obtained mass of such process parameters, which can be regarded as elementary events, it is essential to identify patterns that can be utilized for process predictions (Fig. 2, 1). Here the research and technology field of CEP turns out to be promising. CEP has already been successfully used for fraud detection in banking and finance, in which financial transaction processes can be examined on a very fine-grained level by means of information technology [14]. Thus, with CEP, complex event patterns of a current process instance can be determined (see Fig. 2, 2) and by means of correlation with historical process data and contextual information, event-based prediction models can be derived (see Fig. 2, 3). These models allow to determine forecasts of substantially higher accuracy, since predictions are not purely based on stochastic, but instead, the actual current state is decisive for the computation (see Fig. 2, 4). Hence, they can even be used for a high-level production planning and control (see Fig. 2, 5). The example below demonstrates the disadvantages of stochastic forecasts, the potential of event-based predictions through utilizing CEP techniques and how they can be realized in principle. Figure 2. Basic concept for an event-based process forecast and control 545 532

A process manufacturing company receives two customer orders for which a production planning is carried out. For the first one, a final product D with a minimum quality of 90 quality units (QU) has to be manufactured; the second one only requires a quality of 70 QU of product D. Based on the statistical frequency of occurrence of certain quality characteristics after passing through a manufacturing process A, classic stochastic forecasting approaches determine an 80% probability that final products of type D, as resulting by-products, have a quality of 95 QU (D 1.1 ) or 80 QU (D 1.2 ; see Fig. 3a). Thus, both customer orders could be satisfied. In this context as the process proceeds, the forecast assumes further processing of the intermediate C, which is created with 90 QU. The timing of customer orders and the machine allocation plan would be based on this forecast. In addition to manufacturing process A, manufacturing process B allows to produce the final product D out of intermediate products C (see Fig. 3b). This manufacturing variant can be selected if the resulting C product in process A does not have the necessary quality of at least 90 QU to serve the further processing to D > 80 QU. However, this alternative processing consumes more time and a higher amount of material. Moreover, the end products of this production process meet only in 25% of all cases the required quality criteria. In the stochastically more likely case (75%) only one product with 70 QU can be manufactured and eventually be sold (see the lower process path in Fig. 3b). In the worst case, this final product must therefore be disposed separately or, alternatively, needs to be returned to manufacturing process A. Thus, the use of purely stochastic forecasts would lead to the process execution outlined above with a probability of 80% (visualized in red). However, this means at the same time that this prediction is simply wrong in almost a quarter of all cases and would lead to final products that are not available without additional expense (see the lower process path in Fig. 3a). Thus, the use of such purely stochastic forecasts proves to be insufficient, since it can only determine a forecast with minor significance and lead to an incorrect production planning. If the considered manufacturing processes are equipped with appropriate sensors, a database can be developed that maps the diverse states of production and the corresponding manufacturing context patterns. This underlying data could be correlated with a current process state determined by CEP. Hence, in the outlined case a significantly more accurate prediction could be identified by having knowledge about states x and y. This is possible, since the forecast is not purely stochastic, but event-based. This means for the underlying application example: if for instance after completion of process step A within manufacturing process A, a certain quality of the input raw material M 2, a machine variance of the machines used by va 1,..., va n, as well as the participation of employee m in process step B are detected, a complex event pattern will be identified that is correlated with the historical data base for a prediction. Based on the recognition of this complex event, the forecast would either significantly strengthen the probability of process variant A or, in contrast, predict the stochastic exception (see green process steps in Fig. 3a). Considering the latter, the intermediate C with a quality of 70 QU would result with high significance. For its further processing to product D, a final quality of 65 QU is determined (see green process steps in Fig. 3a). According to this, the computation forecasts that the customer orders cannot be satisfied as initially planned. Based on the determined values for C 2 and other process parameters, in accordance to the diagnosed state y, it will also be predicted that the further processing by manufacturing process B will significantly contribute to the stochastic unlikely final product D that is of sufficient quality to satisfy the customer orders (see green process steps in Fig. 3b). Hence, by incorporating current process and context parameters, the further course of a process can be predicted with considerably higher precision. If states x and y would recognized at a time t by CEP techniques, i.e., already recognized at the beginning of the production, the production could be planned more precisely and proactively controlled based on event-based predictions. For instance, certain setup procedures for starting manufacturing process B could already be performed in parallel to the running process steps B and C 2 in production process A. Stretched production time, increased material requirements as well as personnel and equipment utilization could be scheduled earlier or examined for the computation of possible alternatives. While a stochastic forecast detection is quite simple to compute, event-based prognoses yet turn out to be extremely complex even for the simplified example. The individual process and context parameters form a highly interwoven construct, which complicates the correlation of individual basic events. Looking at the amount of data to be analyzed, it becomes evident that this is a challenge that must be tackled with innovative methods of big data analytics. A case study at Saarstahl AG, a large steel producing company, outlined in the next section brings out this fact by analyzing actual available data in a selected part of their production processes. Figure 3. Flow and forecast of manufacturing processes 546 533

IV. CASE STUDY: STEEL BAR PRODUCTION AT SAARSTAHL AG A. Brief Description and Case Study Settings This case study analyzes a production part within the steel bar production at Saarstahl AG, a major German steel producer, and provides types and sizes of sensor data that is currently collectable throughout the production processes. The outlined data was collected and evaluated by the central department of information and communication technology at Saarstahl AG. In the considered production area, half a million tons of steel are produced annually. In order to meet customers' specific quality requirements for various existing end products, Saarstahl AG conducts comprehensive quality checks within its production line. These include diameter controls with laser, surface testing by magnetic particle testing, checks for internal steel errors by ultrasound, and a variety of temperature and vibration measurements. All of these techniques provide continuous sensor data at the lowest system level (L1). In addition, other sensor systems in production (ambient and positioning sensors) are installed to monitor the control of steel bars within the production line via a material flow tracking system (L2-system level). Based on this basic data and the available customer orders, the production planning and control system calculates a rough schedule (L3 to L4 system level). B. Current State of Production Planning and Control at Saarstahl AG Steelmaking processes are significantly influenced by internal and external events that lead to deviations in production quality. Some examples are fluctuating material properties of pig iron or production-induced deviations due to physical processes taking place. This entails a need for frequent adaptations of individual production instances within production and subsequently extensive re-planning of underlying production plans. For instance, if a standard further processing of steel provides the steps peeling, heat treatment and finally sawing of steel bars, then the crude steel cannot pass these steps if the sensor network used in the process detects a bending, as an example of a production deviation. In this case, this intermediate either needs to be postprocessed in a dressing and straightening machine to comply with quality requirements, or the processing of the intermediate can continue for another customer with lower quality criteria. Therefore the production planning for this process instance is obsolete and needs to be re-initiated in order to meet quality promises and deadlines to all customers, e.g., steel bars need to be factored into the normal process flow after conducting such an ad hoc processing step. Since the production planning systems compute an almost full capacity for the following days in batch mode, the simple insertion of this process instance into the running production flow may contradict to the planned execution of the production plan. Therefore, in case a production deviation is detected during the production process, an intelligent recalculation of the entire production plan would be necessary in real-time to ensure an optimal solution from a global planning perspective. If such production influencing events were even anticipatable, corresponding countermeasures could be initiated before development. Currently, the data provided by the sensor network, which is integrated into the production processes, exceeds the volume that is analyzable so far. The information and control systems used and the analysis techniques available on the market do not allow the producer to monitor and control the entire production process in real-time. Moreover, no future states and events, such as looming production deviations, can be predicted on time. Hence, so far the control of production processes is rather done in a reactive way; yet, the sketched vision of a predictive enterprise has not been realized. C. Data Characterization and Challenge of "Big Data" In the following, sample data obtained from the applied sensor networks at Saarstahl AG are described according to the Big Data characteristics proposed by Gartner [15]. If the entire sensor networks within the production process is considered which would be necessary for a comprehensive production planning and control on an L3-/L4-systems level, the "Big Data challenge" will be many times higher. An example from the sensor network illustrates the immense volume of data in monitoring the production process. In the rolling mills 31 and 32 there are two optical surface test sensors that can continuously provide real-time data for the detection of surface defects during the rolling process. Basically, this allows to take into account the varying customer demands for a particular surface quality. The unit can already prototypically detect errors and differentiate the types of errors. This optical testing generates several hundred terabytes of data annually. Currently, only a sporadic reactive analysis of these data is possible. Also, other context data from the sensor network and the systems settled on L-2 or L-3 level can currently not be linked due to the volume of data to be analyzed. While these systems could in principle detect production deviations in batch mode, this takes too long to allow timely reactions. While this is just one example of very large resulting data of individual sensors in a particular section of the factory, another example illustrates the high data diversity (variety), which is continuously collected by different sensors and sensor networks at various points throughout the production process. This places high demands on an analysis by big data principles. For instance, the further processing of steel bars as of now already provides half a million of sensor data records, which reflect one production area to a particular context. In the next couple of months the sensor performance will be advanced, such that over 1.5 million sensor data will be available on L1 and L2 level. According to the principles of CEP; however, only the identification of relevant events in this torrent of both homogeneous and heterogeneous data as well as their correlation allows to derive patterns and deviations. This is possible only by using highly structured, technical knowledge. At this point, the basic claim to a scalable solution becomes clear. Since such a sensor network should be flexibly expandable, but also allow analyses and forecasts within a required time frame. For instance, the 547 534

company plans to increase sensing in this subsection to an output of more than five million records, which underlines the need for scalability. Thus, in terms of the analysis of these large and diverse data, the responding time is crucial, since speed is a decisive competitive factor in the analysis (Velocity / Analytics). Classic reporting or batch processing would be significantly too slow, so that so-called high velocity technologies must be performed in near-real-time analyses. For the purposes of the outlined vision of predictive enterprises, it is also crucial to conduct accurate forecasts of the process sequences. Each day, an average of one terabyte of video data is recorded in a single subsection of the plant. However, a pure video analysis method is not sufficient for predictive analytics methodologies. In the existing system, it has been shown that only some production deviations could be detected by this classical approach. In addition, there is no feedback for process optimization. Therefore process data need to be included in the model formation and forecasting. Here, as outlined, over one million data sets are incurred in the coming months. For analyzing the dependencies among process and video data, data from a long period of time must be used for model training. In this case, the data volume may rapidly exceed 50 terabytes. For a real-time adaptive prediction, on average one-tenth of the data should be used. At present, however, such a number of data can hardly be processed in real-time. A direct compression of the data is impossible because of its variety to be considered. As of now, due to its big data characteristics, the production process is away from the envisioned optimum in terms of planning and control. Technically, Saarstahl AG can integrate further sensor technology into its production processes to achieve a better monitoring. However to increase the demanded business value, current analytical methods take too long to complete an analysis. V. GENERAL SYSTEM ARCHITECTURE TO REALIZE EVENT-BASED PREDICTION This section sketches a general system architecture for implementing the concept of event-based predictions based on big data analytics (cf. preliminary work on basic technological requirements derived from the outlined use case analysis in sec. IV [5]). The architecture comprises three different layers, which are often considered within analytical processes, and will form the cornerstone to realize a predictive enterprise: descriptive, predictive and prescriptive layers [16]. Sensor networks continuously providing data streams depicting current process and context situations form the foundation of the proposed architecture (Fig. 4, 1). The first layer provides means to store past behavior of processes, which is captured by sensed data, in descriptive models. In this regard, it is very important to collect process behavior in correlation with historical context situations to be able to derive certain event conditions in relation to different process outcomes. These historical description models are particularly challenging as they comprise big data characterized by their volume (cf. use case analysis in sec. IV.C). Yet, they are indispensable for having knowledge on the insights of the processes at hand. In addition to store process data in historical description models, it is compulsory to investigate current real-time conditions of processes through capturing their current events and context situations. Due to the wide variety of heterogeneous sensors, which are used throughout enterprises, this is particularly a challenge of big data analytics in terms of data variety (cf. use case analysis in sec. IV.C). Eventually, these streams have to be searched for patterns in real-time, e.g. by CEP technology (Fig. 4, 2). This allows to correlate current conditions with historical ones as the basis for deriving real-time event-based predictions. As a technical infrastructure, a platform must be available and implemented, which combines a batch-oriented analysis with that of a distributed stream mining analysis. Whereas batchoriented methods are important for training prediction models as the basis for the succeeding layer; stream-oriented ones are compulsory for the actual real-time analysis of incoming data streams. In particular, for the automatic analysis of image and video sensors as outlined in the use case dedicated algorithms needs to be available in order to derive structured information from unstructured data. The next layer of the outlined architecture will form the core component since it computes prediction models comprising forecasts about the future progress of business processes, which allows a proactive reaction to predicted problems (Fig. 4, 3). Figure 4. Interaction of individual system components realizing event-based prediction 548 535

Several prediction techniques and algorithms can be applied, all of which have to be capable of processing big data and to cope with real-time requirements, which is another challenge in terms of big data analytics. In this regard, sophisticated CEP techniques are required predicting likelihoods of the occurrence of future atomic events that will eventually trigger complex ones. These probability assumptions will realize a predictive complex event detection. Within the computed prediction models not only one single possible future process progress will be forecasted, but several multiple ones whose occurrence are depended both on non-influenceable events as well as on influenceable actions (see Fig. 4, 4). Thus, the prescriptive layer, as the third and last one, calculates certain recommendations or automatic decisions and actions to positively influence and control the outcome of business processes in accordance to specific process objectives. As a result of intelligent algorithms it will further be determined whether changes enacted within one process instance impact other running instances which in turn will affect recommendations and automatic actions. These automatic or manually triggered actions based on real-time event-based predictions eventually realize an intelligent proactive process planning and control and thus the vision of a predictive enterprise (see Fig. 4, 5). VI. STATE OF THE ART DISCUSSION A. Process Data and Context Data Acquisition and its Real-Time Analysis In order to detect context situations and progress of ongoing processes, data must be continuously collected. Technically, this can be done by means of physical and virtual sensors, which are often connected in a wireless network. Physical sensors are able to measure for instance pressure, temperature or vibration and recognize complex structures such as audio signals [17]. Additional data is obtained from IT systems by evaluating for example log files or exchanged messages at the lowest system level [18]. The next step is to analyze this mass of collected data. A common approach is that atomic events, which are detected by sensors, are being first cumulated to more complex ones, which is called CEP. CEP combines methods, techniques and tools to analyze mass of events in real-time and to obtain higher aggregated information from this most elementary data. Such approaches can be found for example in [19], who apply CEP on real-time data streams, and with particular respect to RFID-based data streams in [20]. Often ontologies are used to identify semi-automatically the semantic context of events [21]. Operations on data of atomic and complex events require algorithms that are well-adapted to the process context as well as on big data characteristics. A well-known approach for big data analytics is MapReduce, which decomposes a problem into independent sub-problems that are distributed and solved by so-called mappers. However, traditional forms of MapReduce follow a batch approach, which is inapplicable for data streams. In this regard, the field of research of stream mining and stream analytics has been formed recently [22]. B. Process Forecast Calculation and Simulation Predictive analytics refers to the forecast of prospectively occurring events. The necessary mathematical modelling can be done by using historical data. A characteristic example of a simple forecast calculation is the moving average. Such simple models only work if they are mostly independent of external influencing events, which is rarely the case. Especially in business processes several dependencies and influencing factors exist. Thus, modern statistical methods are required to recognize dependencies and patterns in large amounts of data, such as decision trees, univariate and multivariate statistics and data warehouse algorithms [22]. Approaches combining predictive analytics with BPM are coined as Business Process Intelligence (BPI) (cf. [23] for a discussion on existing definitions). Since it is feasibly to increase the context awareness in the era of Internet of Things, it is promising to use data collected by sensors to increase the accuracy of business process predictions. Reference [24] present a first conceptual framework for combining CEP with predictive analytics. Also [25] provide first conceptual works; however, they state that "[e]ventbased or event-driven business intelligence" approaches are only rudimentary researched and find limited integration in software tools. Even if [26] already combines BAM with BPI, the question on how findings from this connection can be used for adapting and optimizing business processes is still open. C. Business Process Adaptation and Optimization In traditional BPM approaches, processes are typically adapted and improved on type level at design time after passing a business process lifecycle. This ex post handling as it is regularly applied in business process controlling and mining however causes considerable time delay [27]. Since an aggregation of process execution data has to be done in a first place, before processes can eventually be optimized at type level, problematic process executions have already been completed and can no longer be optimized. Thus, scientists from the process mining community explicitly call for a stronger "operational support" [28]. This requires the existence of immediate feedback channels to the business process execution. Without the possibility of such feedback loops, only ex-post considerations are possible. At this level of run time adaptation and optimization of business processes, first approaches already exist [25]. In the scope of commercial BPM systems, there are some early adopters characterized by the term Intelligent Business Process Management Suites [29]. For example, IBM s Business Process Manager provides means to analyze operations on instance level in real-time, to control them and to respond with further process steps in an ad-hoc manner [30]; yet, a fully automation cannot be attested. On the market of real-time CEP, further vendors are present; e.g. Bosch Software Innovations or Software AG have implemented CEP functionality into BPM solutions [29]. However, in order to speak of intelligent BPM systems, a more powerful automated analysis of the variety of events must be implemented as a basis for automated event-based predictions and process control [1]. 549 536

VII. CONCLUSION AND OUTLOOK This paper outlined the potentials of event-based predictions in order to plan and eventually control business processes. Besides outlining these potentials, a general concept for event-based predictions has been conceived and the current state of the art was discussed. In addition, the article showed, by means of a case study, which process data a sample steel producing company can currently collect by its applied sensor technology. This data base is potentially available for event-based predictions of its manufacturing processes. Based on the description of the data, which was collected and evaluated by the department of information and communication technology at Saarstahl AG, the paper concluded that with current techniques and systems available data cannot be analyzed in a reasonable time frame to make sufficient business value out of it. Without dedicated big data analytics, the sample company will not be able to exploit full potentials of its data. The paper at hand should form a working and discussion basis for further research in big data analytics research based on a proposed system architecture. ACKNOWLEDGMENT This research was funded in part by the German Federal Ministry of Education and Research under grant number 01IS12050 (project IDENTIFY). REFERENCES [1] J. Krumeich, B. Weis, D. Werth, and P. Loos, Event-driven business process management: Where are we now? A comprehensive synthesis and analysis of literature, Business Process Management Journal, vol. 20, no. 4, 2014, in press. [2] A. Lundberg, Leverage complex event processing to improve operational performance, Business Intelligence Journal, vol. 11, no. 1, 2006, pp. 55 65. [3] K. Unni, Steel manufacturing could use more sensing and analysis, http://www.sensorsmag.com/process-industries/steel-manufacturingcould-use-more-sensing-and-analysis-10249, 2012, accessed 2014-04-25. [4] C. Pettey, and L. Goasduff, Gartner says between now and year-end 2014, overlooked but easily detectable business process defects will topple 10 global 2000 companies, http://www.gartner.com/ newsroom/id/1530114, 2011, accessed 2014-04-25. [5] J. Krumeich, S. Jacobi, D. Werth, and P. Loos, Towards planning and control of business processes based on event-based predictions, Proc. 17th Int. Conf. Business Information Systems (BIS 2014), LNBIP vol. 176, Springer, 2014, pp. 38-49. [6] L. Dubé, and G. Paré, Rigor in information systems positivist case research: current practice, trends and recommendations, MIS Quarterly, vol. 27, no. 4, 2003, pp. 597-636. [7] I. Benbasat, D. K. Goldstein, and M. Mead, The case research strategy in studies of information systems, MIS Quarterly, vol. 11, no. 3, 1987, pp. 369-386. [8] A.-W. Scheer, Wirtschaftsinformatik: Referenzmodelle für industrielle Geschäftsprozesse. Berlin, Heidelberg: Springer, 1998. [9] P. Riebel, Industrielle Erzeugungsverfahren in betriebswirtschaftlicher Sicht. Wiesbaden: Gabler, 1963. [10] P. Loos, Produktionslogistik in der chemischen Industrie Betriebstypologische Merkmale und Informationsstrukturen. Wiesbaden: Gabler, 1997. [11] D. Hahn, and G. Laßmann, Produktionswirtschaft Controlling industrieller Produktion. Heidelberg: Physica, 1999. [12] K. Kurbel, Produktionsplanung und -steuerung im Enterprise Resource Planning und Supply Chain Management. München: Oldenbourg, 2010. [13] M. Minelli, M. Chambers, and A. Dhiraj, Big data, big analytics: Emerging business intelligence and analytic trends for today s business. Hoboken, NJ: John Wiley & Sons, 2013. [14] R. von Ammon, T. Ertlmaier, O. Etzion, A. Kofman, and T. Paulus, Integration complex events for collaborating and dynamically changing business processes, Int. Workshop ICSOC/ServiceWave 2009, LNCS vol. 6275, Springer, 2010, pp. 370 384. [15] M. A. Beyer, and D. Laney, The importance of 'big data': A definition, https://www.gartner.com/doc/2057415, 2012, accessed 2014-04-25. [16] R. Akerkar, Advanced data analytics in business, in Big data computing, R. Akerkar et al., Eds. Boca Raton, FL: CRC Press, 2013, pp. 373-400. [17] M. Stäger, P. Lukowicz, and G. Tröster, Power and accuracy tradeoffs in sound-based context recognition systems, Pervasive and Mobile Computing, vol. 3, no. 3, 2007, pp. 300 327. [18] J. Choi, RFID context-aware systems, in Sustainable radio frequency identification solutions, C. Turcu et al., Eds. Rijeka: InTech, 2010, pp. 307 330. [19] E. Wu, Y. Diao, and S. Rizvi, High-performance complex event processing over streams, Proc. Int. Conf. Management of Data (SIGMOD 06), ACM, 2006, pp. 407 418. [20] F. Wang, S. Liu, P. Liu, and Y. Bai, Bridging physical and virtual worlds: Complex event processing for RFID data streams, 10th Int. Conf. Extending Database Technology (EDBT 2006), LNCS vol. 3896, Springer, 2006, pp. 588 607. [21] K. Taylor, and L. Leidinger, Ontology-driven complex event processing in heterogeneous sensor networks, 8th Extended Semantic Web Conf. (ESWC 2011), LNCS vol. 6644, Springer, 2011, pp. 285 299. [22] A. Brito et al., Scalable and low-latency data processing with stream MapReduce, Proc. IEEE Third Int. Conf. Cloud Computing Technology and Science (CLOUDCOM 11), IEEE Computer Society, 2011, pp. 48 58. [23] M. Linden, C. Felden, and P. Chamoni, Dimensions of business process intelligence, Business Process Management Workshops (BPM 2010), LNBIP vol. 66, Springer, 2011, pp. 208 213. [24] L. J. Fülop et al., Predictive complex event processing: A conceptual framework for combining complex event processing and predictive analytics, Proc. Fifth Balkan Conf. Informatics (BCI 12), ACM, 2012, pp. 26 31. [25] C. Janiesch, M. Matzner, and O. Müller, Beyond process monitoring: A proof-of-concept of event-driven business activity management, Business Process Management Journal, vol. 18, no. 4, 2012, pp. 625 643. [26] B. Schwegmann, M. Matzner, and C. Janiesch, A method and tool for predictive event-driven process analytics, Proc. 11th Int. Conf. Wirtschaftsinformatik (WI2013), University of Leipzig, pp. 721 735. [27] P. Loos, S. Balzert, and D. Werth, Controlling von Geschäftsprozessen, in Prozessmanagement, R. Jochem., K. Mertins, and T. Knothe, Eds. Düsseldorf: Symposium Publishing, 2010, pp. 443 471. [28] W. M. P. van der Aalst et al., Process mining manifesto, Business Process Management Workshops (BPM 2011), LNBIP vol. 99, Springer, 2012, pp. 169 194. [29] J. Sinur, W. R. Schulte, J. B. Hill, and T. Jones, Magic Quadrant for Intelligent Business Process Management Suites, http://www.gartner.com/doc/2179715/magic-quadrant-intelligentbusiness-process, 2012, accessed 2014-04-25. [30] IBM, Eds., IBM Business Process Manager Standard, http://www- 03.ibm.com/software/products/us/en/business-process-managerstandard, 2013, accessed 2014-04-25. 550 537