Wayside Defect Detector Data Mining to Predict Potential WILD Train Stops

Similar documents
NATIONWIDE WAYSIDE DETECTOR SYSTEM

FLEXURAL PERFORMANCE REQUIREMENTS FOR PRESTRESSED CONCRETE TIES BY FACTORING * * *

SAFETY TECHNOLOGY TOOLKIT

21ST CENTURY RAILROADS: SAFETY THROUGH TECHNOLOGY

Analysis of Causes of Major Train Derailment and Their Effect on Accident Rates

Integration of PTC and Ride Quality Data. Presented by: Wabtec Railway Electronics, I-ETMS PTC Supplier. and

The Value Proposition for GS1 ID Keys and RFID in Rail

Chapter 3 Local Marketing in Practice

Big Data in Railroad Engineering

Risk Assessment / Risk Management Protocol

UNECE Regulation No 106

Maximizing Fleet Efficiencies with Predictive Analytics

Finding soon-to-fail disks in a haystack

Framework for Short-Line Railroad Track Asset Management and Condition Reporting

Some Essential Statistics The Lure of Statistics

Efficiency in Software Development Projects

Railroad Accident Brief

STATISTICA Formula Guide: Logistic Regression. Table of Contents

Railway Freight Car Inspection and Safety Rules

ECONOMIC COMMISSION OF EUROPE (ECE) BRANDING

Amtrak s s Vehicle/Track Interaction Measuring System Integrating the Control Strategies of Engineering, Mechanical and Operations

WMATA S Automated Track Analysis Technology & Data Leveraging for Maintenance Decisions

International Statistical Institute, 56th Session, 2007: Phil Everson

Analytical Test Method Validation Report Template

Chapter 7: Simple linear regression Learning Objectives

Predictive Analytics Tools and Techniques

Software Development (PAKPAVE) for Flexible Pavement Design

Example: Credit card default, we may be more interested in predicting the probabilty of a default than classifying individuals as default or not.

The APT-WORM (Wheel flat and Out of Roundness Monitoring) system

CHURN PREDICTION IN MOBILE TELECOM SYSTEM USING DATA MINING TECHNIQUES

Risk Assessment for visits by schools and other groups including children.

SWITCHING STRATEGIES: MOVING TO CONDITION- BASED MAINTENANCE. By David Stevens, Technical Manager & Trainer, AVT Reliability

COMPARISONS OF CUSTOMER LOYALTY: PUBLIC & PRIVATE INSURANCE COMPANIES.

An Introduction to. Metrics. used during. Software Development

Simple Predictive Analytics Curtis Seare

Nuclear Energy. The ATLAS Railcar Project. Patrick R. Schwab. January 2016

Chapter 3. Track and Wheel Load Testing

Measurement Information Model

Standard the the Raising

GUIDELINES FOR UTILITY INSTALLATIONS Part 1 - Wire Lines and Communications Cables

Injured on the Job. Your Rights under FELA. Quick Facts: What To Do If Injured

Adaptive Demand-Forecasting Approach based on Principal Components Time-series an application of data-mining technique to detection of market movement

Fleet Management Procedure

Lift and Drag on an Airfoil ME 123: Mechanical Engineering Laboratory II: Fluids

II. DISTRIBUTIONS distribution normal distribution. standard scores

Simple Linear Regression Inference

6.4 Normal Distribution

Reducing Rail Surface Defect Service Failures on the CSXT Railroad

Data quality in Accounting Information Systems

Additional sources Compilation of sources:

Quality Control of National Genetic Evaluation Results Using Data-Mining Techniques; A Progress Report

Physics Lab Report Guidelines

Freight Car Maintenance Inspection

A Comparative Study of the Pickup Method and its Variations Using a Simulated Hotel Reservation Data

Railroad Safety Data Frequently Asked Questions (FAQ)

Determining Factors of a Quick Sale in Arlington's Condo Market. Team 2: Darik Gossa Roger Moncarz Jeff Robinson Chris Frohlich James Haas

Guidance for the Quality Assurance of Fire Protection Systems

IV. INDUSTRIAL TRACK DESIGN

DESIGN OF PRESTRESSED BARRIER CABLE SYSTEMS

STATS8: Introduction to Biostatistics. Data Exploration. Babak Shahbaba Department of Statistics, UCI

13. Poisson Regression Analysis

Neural Network based Vehicle Classification for Intelligent Traffic Control

Weight of Evidence Module

2010 SECURITY AWARENESS TRAINING

TREAD DESIGNS RADIAL AND BIAS

Statistiek II. John Nerbonne. October 1, Dept of Information Science

Design & Implementation about Mining Enterprise EAM (Enterprise Asset Management) System

ITT Motion Technologies Supplier Quality Manual

BEHAVIOR BASED CREDIT CARD FRAUD DETECTION USING SUPPORT VECTOR MACHINES

BNSF RAILWAY Safety and Hazmat Overview

Railway Track Design

A model to predict the probability of highway rail crossing accidents

1-3 id id no. of respondents respon 1 responsible for maintenance? 1 = no, 2 = yes, 9 = blank

E/ECE/324/Rev.2/Add.116/Rev.2/Amend.2 E/ECE/TRANS/505/Rev.2/Add.116/Rev.2/Amend.2

Bedford Group of Drainage Boards

Head Loss in Pipe Flow ME 123: Mechanical Engineering Laboratory II: Fluids

A wireless sensor network for traffic surveillance

A transverse strip of the deck is assumed to support the truck axle loads. Shear and fatigue of the reinforcement need not be investigated.

NC STATE UNIVERSITY Exploratory Analysis of Massive Data for Distribution Fault Diagnosis in Smart Grids

ARA Digital Train Radio System Functional Requirements For Australian Metropolitan Railway Operators. 19 September 2005

Webinar Series. Committee. Disc Brake Wheels Off Inspection and Reline. Presents

User-friendly Configuration of AUTOSAR ECUs with Specialized Software Tools

Power Analysis for Correlation & Multiple Regression

Regression 3: Logistic Regression

The Impact of In-Service Aircraft Mission Variation on Fleet Management and Logistic Support Analysis

Maximize Software Development ROI With Quality Assurance. Showing the value of the Quality Process

DESCRIPTIVE STATISTICS. The purpose of statistics is to condense raw data to make it easier to answer specific questions; test hypotheses.

Waveney Lower Yare & Lothingland Internal Drainage Board Risk Management Strategy and Policy

Transcription:

Wayside Defect Detector Data Mining to Predict Potential WILD Train Stops Leila Hajibabai a, M. Rapik Saat a, Yanfeng Ouyang a, Christopher P.L. Barkan a, Zengyi Yang a, Kim Bowling b, Kamalesh Somani b, Don Lauro b, and Xiaopeng Li a a Department of Civil and Environmental Engineering, University of Illinois at Urban-Champaign b CSX Transportation Abstract Advanced wayside detector technologies can be used to monitor the condition of railcar components, notify railroads of probable failures to equipment and infrastructure in advance, and predict alert of imminent mechanical-caused service failures. Using some statistical data-mining techniques, historical railcar health records from multiple Wayside Defect Detector (WDD) systems can potentially provide the essentials to recognize the patterns and develop the reliable and innovative rules to predict the failures and reduce related risks on railroads. In this paper, data from Wheel Impact Load Detector (WILD) and Wheel Profile Detector (WPD) were analyzed through comparing historical measurements for failed and non-failed wheels on the same truck to predict train stops due to high impact wheels. An exploratory data analysis was performed to identify the most critical measurements from each detector by comparing the distributions of several measurements from failed wheels to the ones from non-failed wheels. A logistic regression approach was used to predict the probability of potential high impact wheel train stops. Initial results show a 90% efficiency to predict the failure within 30 days after the most recent WILD reading. Keywords: wayside defect detector, prediction, failure, non-failure, statistical analysis, logistic regression, WILD, WPD 2012 AREMA

1 Introduction Train accidents, either track-or railcar-related causes, imposes huge costs, safety concerns, and affects train operations and efficiency. The likelihood of railcar mechanical failures may be reduced by (1) improving the mechanical components resistance to bear higher stresses and loads, and/or (2) using wayside defect detectors (WDD) to monitor the performance of equipment and components to take proactive actions before mechanical failures occur. Different types of WDD are used to monitor railcar health and/or detect immediate hazards that can result in train derailments. WDD can be broadly classified into two types (Lagnebäck 2007, Ouyang et al 2009): Reactive Systems detect railcar component conditions that have a short latency between detectability and failure of the component. Failure to react promptly to these alarms may result in damage to railroad infrastructure and/or a derailment. Dragging Equipment Detectors (DED) and Hot Bearing Detectors (HBD) are widely used examples of reactive WDD technologies. Predictive Systems are capable of measuring, recording, and trending the performance of vehicles and specific components. The information collected can be used to analyze the condition of equipment to predict possible failures and faults that may occur sometime in the future, thereby making it easier to plan maintenance activities. Wheel Impact Load Detectors (WILD) and Wheel Profile Detectors (WPD) are commonly used examples of these types of technologies on North American railroads. Some predictive technologies, such as WILD, may also detect certain imminent hazards and therefore can be classified as both predictive and reactive. Statistical data analysis and data mining approaches can offer the basis to recognize the patterns and dependencies between various factors derived from the WDD data. These sets of 2012 AREMA

analyses help generate innovative, reliable, and advanced rules to predict the occurrence of equipment failures, service disruptions, or possibly mechanical-caused derailments. Therefore, to reduce inspection costs and frequencies, and optimize railcar maintenance schedule, statistical and data mining techniques can be used to analyze historical railcar health record data from multiple WDD systems. This paper describes an efficient data mining framework to analyze data from different WDDs and predict high impact wheel train stops in advance. Data from WILD and WPD detectors were analyzed through comparing historical measurements for failed and non-failed wheels in the same truck. Comparing the distributions of failed and corresponding non-failed detector readings for wheels helped identify critical measurements from each detector. Then, to predict the probability of potential high impact wheel train stops, logistic regression models were developed. The exposition of the paper is as follows. First, in Section 2, methodology including the exploratory analysis and regression approach will be discussed. Afterwards, data analysis results and performance measures will be introduced in Section 3, and conclusions will be discussed in Section 4. 2 Methodology Detailed analysis of individual data sources enables further understanding of information availability and limitations. Since data obtained from multiple WDD is significantly large, the preprocessing and analyzing of the combined data from different sources is non-trivial and requires significant time and effort. Therefore, an efficient data mining framework is required to manage the large databases and prepare useful data for further railcar condition analysis. In this section, a data analysis approach to determine the failed and non-failed railcar wheels will first be introduced (Figure 1). Then the method to determine the potential and true failures and 2012 AREMA

corresponding non-failures will be discussed. The exploratory approach to identify the most critical variables from the data attributes will be applied. Finally, those critical variables are considered in a regression model, which was developed to predict the failures based on the identified significant variables. WILD Database Failure Data Mining Train stop alert & repair date thresholds Data Analysis Potential and True Failure WPD Non-Failure Controlled approach Non-Failure on the adjacent axle of the same truck Exploratory and Regression Analysis Decision Support Predicted Failures Predicted Non-Failures Figure 1. Data mining framework for condition monitoring to predict wheel failures 2.1 Data Analysis Approach The WILD and WPD data were analyzed on railcars with high impact wheel train stops, and dates of these car repair billing records to identify railcar wheel failures, and non-failures. If an instance of train stop was recorded for a wheel it would be categorized as failure; otherwise, with no train stop nor repair record, it was categorized as a non-failed wheel. Wheels in the failure category were further analyzed according to the time span between the most recent detector measurement and the train stop alert date (M-TS), and similarly based on the time span between each high impact train stop and repair dates (TS-R) (Figure 2).Certain thresholds for M-TS and TS-R were defined and tested to determine the railcar wheel potential 2012 AREMA

vs. true failures. Data availability was an important issue to determine the appropriate thresholds. A maximum threshold of 30 days was selected for M-TS and TS-R to determine the potential and true failures for wheels. In other words, if reading date was within 30 days before the high impact train stop date, the wheels would be categorized as potential failures; and, the wheels would be classified as true failures if wheel repair date was within 30 days after the train stop. M-TS TS-R Measurement Train Stop Repair Time Figure 2. Time span approach to predict wheel failures To reduce uncertainty and variance from different truck characteristics, a controlled approach has been developed by comparing failed and non-failed wheel historical measurements on the same truck. Therefore, the wheels with no records of high impact train stop nor wheel repair, that were located on the adjacent axle to the failed wheel would be considered in the non-failure category (Figure 3). Figure 3. A controlled approach to compare failed and non-failed wheel measurements 2.2 Exploratory Analysis After determining the failures and non-failures, a set of data attributes from each detector was selected for the exploratory analysis. From WILD measurement, vertical average weight, vertical 2012 AREMA

Vertical Average Weight (Kips) peak force, lateral average force, lateral peak force, and dynamic ratio were selected. Flange height, flange thickness, rim thickness, tread hollow, wheel diameter, groove tread, vertical flange (flange angle), tread built-up, and wheel angle were considered for WPD records. Based on each detector database, the time distribution of failures of each variable was compared to the average of non-failures of the same variable. Then, percentage of the failures above (or below) the non-failures average was determined. If a large percentage of failed-wheel measurements were higher (or lower) than the average measurements for non-failed-wheels, the variable would be considered as a critical variable. Figure 4 shows this exploratory analysis for the vertical average weight from WILD database. The variable is deemed critical since a large percentage of failed-wheel measurements were higher than the average measurements for the non-failures. Using this approach, vertical average weight, vertical peak force, lateral average force, lateral peak force, and dynamic ratio were all considered as critical variables for WILD data. Vertical flange, rim thickness, tread built-up, and wheel angle were the identified critical variables from WPD measurements. 60 Failed Left Wheels Failed Right Wheels 50 40 30 20 One standard deviation of the average value for non-failed left wheels One standard deviation of the average value for non-failed right wheels 10 Failures Above Non-Failure AVG (Left Wheel): 73.76% Failures Above Non-Failure AVG (Right Wheel): 83.33% Failures Below Non-Failure AVG (Left Wheel): 25.53% Failures Below Non-Failure AVG (Right Wheel): 16.67% 0 7/21/2010 8/10/2010 8/30/2010 9/19/2010 10/9/2010 10/29/2010 11/18/2010 12/8/2010 12/28/2010 1/17/2011 Date Figure 4. Exploratory analysis to determine the critical variables from each detector 2012 AREMA

2.3 Regression Analysis A regression model is needed to predict the failures using the significant variables identified in Section 2.2. In order to predict the outcome of a categorical variable according to a set of predictor variables, a logistic regression was developed. The probability of a possible outcome was modeled as a function of exploratory variables. In this paper, the logistic regression setting includes detector settings x 0,x 1,..., x n for a generic instance, regression coefficients as b 0,b 1,...,b n, and logit measure that is z b 0 b 1 x 1 b 2 x 2... b n x n. Since there were two main categories of failures, two regression models could be considered as alternatives: regression model for potential failures and true failures. In this paper, we focused on the logistic regression model based on potential failures. 85% of the data was considered to develop the model, and 15% of the data was used to evaluate the effectiveness of the model to predict failures. Based on the potential failures, we developed three regression models using: (1) WPD critical variables, (2) WILD critical variables, and (3) combined WPD and WILD critical variables. Unit-less values for all critical variables were used in the regression models according to the (measurement non-failure mean)/non-failure mean) 100 formulation. Data structure for data analysis based on WILD potential failures was considered in two separate regression models: (1) ignoring dynamic ratio and (2) including dynamic ratio as a critical variable, as follows: (1) Regression model for WILD data when dynamic ratio is ignored: where 1 P(z) 1 e z z 0.4002 0.0211x 1 0.0261x 2 0.0031x 3 0.6290x 4 (1) x 1 = vertical average force x 2 = vertical peak force 2012 AREMA

x 3 = lateral peak force x 4 = loaded/unloaded (0 or 1) The regression model was found to be significant with p-value of 0.002. (2) Regression model for WILD data when dynamic ratio is considered: where 1 P(z) 1 e z z 0.4103 0.0256x 1 0.6036x 2 (2) x 1 = dynamic ratio x 2 = loaded/unloaded (0 or 1) The regression model was found to be significant with p-value of 0.013. developed: The following regression model for data analysis based on WPD potential failures was where P(z) 1 1 e z z 1.1029 0.0416x 1 0.0272x 2 (3) x 1 = rim thickness x 2 = flange angle The regression model was found to be significant with p-value of 0.031. For WILD and WPD data combined, the following regression model without dynamic ratio was developed. where P(z) 1 1 e z z 2.4402 0.0300 x 1 (4) x 1 = vertical peak force 2012 AREMA

The regression model was found to be significant with p-value of 6.283e-10. When dynamic ratio was considered with combined WILD and WPD data, the following regression model was developed: where P(z) 1 1 e z z 2.3057 0.0317 x 1 0.0107 x 2 (5) x 1 = dynamic ratio x 2 = lateral average force The regression model was found to be significant with p-value of 1.305e-09. 2.4 Regression Model Test As described above, the regression models were developed based on 85% of the data. Afterwards, using the remaining 15% unseen data, a test for each regression model was conducted. Below an example of the test is presented for regression model (1) for WILD data when dynamic ratio is ignored. Table 1 and Figure 5 illustrate the number of true failures, false positives, and false negatives based on predictions of this regression model. As it is shown, predicted true failures and false positives (non-failures that are predicted as failures) decrease by increasing the probability of failure, while false negatives (failures that are predicted as nonfailures) increase. Therefore, it is useful to look at the summation of the two errors. The minimum total errors can be found at 50% failure probability (Table 1 and Figure 5). Table 1. Determining the failure probability at the minimum total errors for regression model (1) Failure Probability >=20% >=30% >=40% >=50% >=60% >=70% >=80% >=90% True Failure 101 95 94 91 85 81 76 59 False Positive 33 31 15 5 3 3 2 2 False Negative 0 6 7 10 16 20 25 42 2012 AREMA

Counts 120 100 True Failure Threshold with minimum sum of errors 80 60 40 20 0 Sum of Errors False Positive False Negative 0 0.2 0.4 0.6 0.8 1 Probability of Failure Figure 5. Threshold with minimum total errors for regression model (1) The confusion matrix for data mining model evaluation, where failure probability was above 50% is illustrated in Table 2-a and 2-b. Probability of Type I error (non-failures that are predicted as failures) was 15% and probability of Type II error (failures that are predicted as non-failures) was 10%. Table 2. Confusion matrix for data mining model evaluation for regression model (1) (a): count Failure Prediction Non-failure Actual Failure 91 10 Non-failure 5 28 (b): percentage Failure Prediction Non-failure Actual Failure 90% 10% Non-failure 15% 85% 2012 AREMA

3 Results Table 3 summarizes the performance results of all five regression models using the unseen data. Table 3. Summary of regression model performance using the 15% unseen data Regression Model True Failure Type I Error Type I Error WILD without dynamic ratio 90% 15% 10% WILD with dynamic ratio 89% 15% 11% WPD 40% 0% 60% WILD and WPD without dynamic ratio 20% 0% 80% WILD and WPD with dynamic ratio 60% 0% 40% As Table 3 illustrates, WILD data leads to an efficient prediction of failures. We found promising results with about 90% efficiency to predict high impact wheel train stops in the next 30 days based on the most recent WILD measurements. Failure prediction for WPD model is not as efficient as WILD model, since the number of records in WPD database were much fewer than WILD readings. Furthermore, due to the lack of symmetry between the total number of WILD and WPD records, development of a reliable combined model was prohibited at this time. 4 Discussion and Conclusions In this paper, data from WILD and WPD databases were analyzed through comparing historical measurements for failed and non-failed wheels on the same truck to predict train stops due to high wheel impacts. Through developing a controlled approach, we performed an exploratory data analysis to identify the most critical measurements from each detector by comparing the distributions of several measurements from failed wheels to the ones from non-failed wheels on the same truck, which reduces variance to identify critical variables and develop efficient prediction models. Then, a set of logistic regression models was developed to predict the probability of occurrence of potential high wheel impact train stops. Results show a 90% 2012 AREMA

efficiency to predict high wheel impact train stops within 30 days after the most recent WILD measurements. The lack of symmetry between the total number of WILD and WPD records prohibited development of a reliable combined model. Moreover, difficulty in analyzing combined data from different sources/detectors reflects the need to focus on data from different detectors that are co-located. Therefore, future research directions may focus on using WILD and WPD data from locations where they co-exist, validating the WILD trending model identified in this phase, and re-developing combined WILD and WPD regression models. References Lagnebäck, R. Evaluation of wayside condition monitoring technologies for condition-based maintenance of railway vehicles. Licentiate Thesis, Lulea University of Technology, Department of Civil, Mining and Environmental Engineering Division of Operation and Maintenance Engineering, Lulea, Sweden, 2007. Ouyang, Y., X. Li, C. P. L. Barkan, A. Kawprasert, and Y.-C. Lai. Optimal Locations of Railroad Wayside Defect Detection Installations. Computer-Aided Civil and Infrastructure Engineering 24 1 11, 2009. 2012 AREMA

Wayside Defect Detector Data Mining to Predict Potential WILD Train Stops Leila Hajibabai a, M. Rapik Saat a, Yanfeng Ouyang a, Christopher P.L. Barkan a, Zengyi Yang a, Kim Bowling b, Kamalesh Somani b, Don Lauro b, and Xiaopeng Li a a University of Illinois at Urbana-Champaign b CSX Transportation

Outline 2012 Annual Conference & Exposition Problem statement Data mining framework for condition monitoring Exploratory data analysis Regression approach Data analysis results and performance measures Conclusion

Problem Statement 2012 Annual Conference & Exposition Advanced wayside detector technologies can be used to: Monitor the condition of railcar components Notify railroads of probable failures to equipment and infrastructure in advance Predict alert of imminent mechanical-caused service failures More effective use of wayside mechanical inspection data can lead to improved safety, service reliability and economics through more effective preventive maintenance practices Objectives: Recognize patterns that ultimately can be used to generate new, reliable, advanced rules to predict the occurrence of service disruptions Develop the reliable and innovative rules to predict the failures (due to high impact wheels) and reduce related risks on railroads Using statistical data-mining techniques on historical railcar health records from multiple Wayside Defect Detector (WDD) systems

Problem Statement 2012 Annual Conference & Exposition Data from Wheel Impact Load Detector (WILD) and Wheel Profile Detector (WPD) are analyzed: Comparing historical measurements for failed and non-failed wheels on the same truck Exploratory data analysis to identify the most critical measurements from each detector by comparing the distributions of several measurements from failed wheels to the ones from non-failed wheels Logistic regression approach to predict the probability of potential high impact wheel train stops Value proposition of individual and multiple WDDs used in combinations are evaluated: WILD only WPD only WILD and WPD

Data Mining Framework for Condition Monitoring to Predict Wheel Failures Database Data Analysis WILD Failure Data Mining Train stop alert & repair date thresholds Potential and True Failure WPD Non- Failure Controlled approach Non- Failure on the adjacent axle of the same truck Decision Support Exploratory and Regression Analysis Predicted Failures Predicted Non-Failures

Data Analysis Approach: Time Span Method to Predict Wheel Failures Wheels in the failure category are further analyzed according to the time span between: The most recent detector measurement and the train stop alert date (M-TS) Each high impact train stop and repair dates (TS-R) Certain thresholds for M-TS and TS-R are defined and tested to determine the railcar wheel Potential vs. True Failures Data availability is an important issue to determine the appropriate thresholds M- TS TS- R Measurement Train Stop Repair Time

Data Analysis Approach: Time Span Method to Predict Wheel Failures Compared historical measurements for failed and non-failed wheels on the same truck to reduce uncertainty and variance Failed wheel (train stop due to high wheel impact) - Potential Failure: Reading date is within 30 days range before train stop date - True Failure: Wheel repair date is within 30 days range after train stop date Non-failed wheel - On the adjacent axle to the failed wheel - No train stop nor wheel repair Failed L1 R1 Non- failed L2 R2

Exploratory Data Analysis to Identify Critical Variables After determining the failures and non-failures, a set of data attributes from each detector is selected for the exploratory analysis WILD measurements: vertical average weight, vertical peak force, lateral average force, lateral peak force, and dynamic ratio WPD measurements: flange height, flange thickness, rim thickness, tread hollow, wheel diameter, groove tread, vertical flange (flange angle), tread built-up, and wheel angle Based on each detector database, the time distribution of failures of each variable was compared to the average of non-failures of the same variable The percentage of the failures above (or below) the non-failures average is determined If a large percentage of failed-wheel measurements were higher (or lower) than the average measurements for non-failed-wheels, the variable would be considered as a critical variable

Sample Exploratory Data Analysis to Identify Critical Variables 60 Failed Le Wheels Failed Right Wheels 50 Vertical Average Weight (Kips) 40 30 20 10 One standard deviation of the average value for non-failed left wheels One standard deviation of the average value for non-failed right wheels Failures Above Non-Failure AVG (Left Wheel): 73.76% Failures Above Non-Failure AVG (Right Wheel): 83.33% Failures Below Non-Failure AVG (Left Wheel): 25.53% Failures Below Non-Failure AVG (Right Wheel): 16.67% 0 7/21/2010 8/10/2010 8/30/2010 9/19/2010 10/9/2010 10/29/2010 11/18/2010 12/8/2010 12/28/2010 1/17/2011 Date Based on this result Vercal Average Weight is deemed crical because large percentage of failed- wheel measurements are higher than the average measurements for non- failed- wheel

Summary of Critical Variables Identified WILD: Vertical Average Weight Vertical Peak Force Lateral Average Force Lateral Peak Force Dynamic Ratio WPD: Vertical Flange (Flange Angle) Rim Thickness Tread Built-up Wheel Angle

Logistic Regression Approach Logistic regression setting Detector readings x 1,x 2,,x n for a generic instance Regression coefficients: b 0, b 1,, b n Logit measure z = b 0 +b 1x1 x 1 + b 2x2 x 2 + + b nxn x n Labeled failures False negatives (missed failures) True positives (good predictions) False negatives False positives Predicted failures False positives

Regression Model Alternatives Potential Failure (85% of the data for model construction and 15% for test) WPD critical variables (failure and non-failure records) WILD critical variables (failure and non-failure records) Combined WPD and WILD critical variables Sparse data to perform True Failure (verified repair) analysis Used unit-less values for all critical variables in regression models: (measurement non-failure mean)/non-failure mean 100

Regression Model: Data Structure for Data Analysis based on WILD Potential Failure (Absolute Value for Lateral Forces) Logistic Regression (if Dynamic Ratio is ignored): P(z) = 1 1+ e z z = 0.4002-0.0211x 1 + 0.0261x 2-0.0031x 3-0.6290x 4 The critical variables in the model (0.1 significance): Vertical Average Force(0.001), Vertical Peak Force(0.001), and Lateral Peak Force(0.1), respectively Loading condition(0.01) has also been considered in this model Significance Test of Logistic Regression: H 0 : coefficients=0; H 1 : O.W. The test is rejected, therefore the regression model between the dependent and independent variables is significant with p-value=0.002

Test of Logistic Regression using the 15% Unseen WILD Data: without Dynamic Ratio (Absolute Value for Lateral Forces) Failure Probability >=0.2 >=0.3 >=0.4 >=0.5 >=0.6 >=0.7 >=0.8 >=0.9 True Failure 101 95 94 91 85 81 76 59 False Positive 33 31 15 5 3 3 2 2 False Negative 0 6 7 10 16 20 25 42 Confusion Matrix for Data Mining Model Evaluaon (P>=50%): Count: Percentage: Actual Actual Failure Predicon Non-Failure Failure 91 10 Non-Failure 5 28 Failure Predicon Non-Failure Failure 90% 10% Non-Failure 15% 85% Probability of Type II Error Probability of Type I Error Counts 120 100 80 60 40 20 0 True Failure Sum of Errors False Positive False Negative Threshold with minimum sum of errors 0 0.2 0.4 0.6 0.8 1 Probability of Failure

Summary of Regression Model Performance Using the 15% Unseen Data Model True Failure Type I Error Type II Error Promising results WILD-w/out Dyn. Ratio 90% 15% 10% WILD-w/ Dyn. Ratio 89% 15% 11% WPD 40% 0% 60% WPD & WILD-w/out Dyn. Ratio WPD & WILD-w/ Dyn. Ratio 20% 0% 80% 60% 0% 40%

Conclusion 2012 Annual Conference & Exposition We developed a methodical, controlled approach to analyze data by comparing failed and non-failed cases on the same truck Reduce variance to make it easier to identify critical variables and develop good prediction models We found promising results with about 90% efficiency to predict high wheel impact train stops in the next 30 days based on the most recent WILD measurements Detailed analysis of individual data sources enable further understanding of information availability and limitations Increase value proposition by identifying potential new rules to avoid service failures and derailments

Conclusion 2012 Annual Conference & Exposition The lack of symmetry between the total number of WILD and WPD records prohibit development of a reliable combined model Difficulty in analyzing combined data from different sources/detectors reflects the need to focus on data from different detectors located at the same location Could be helpful to guide future plan to install which detectors at super sites

QUESTIONS?

Appendix

Regression Model: Data Structure for Data Analysis based on WILD Potential Failure (Absolute Value for Lateral Forces) Logistic Regression (if Dynamic Ratio is considered): P(z) = 1 1+ e z z = 0.4103 +0.0256x 1-0.6036x 2 The critical variables in the model (0.01 significance): Dynamic Ratio(0.001) Loading condition(0.01) has also been considered in this model Significance Test of Logistic Regression: H 0 : coefficients=0; H 1 : O.W. The test is rejected, therefore the regression model between the dependent and independent variables is significant with p-value=0.013

Test of Logistic Regression using the 15% Unseen WILD Data: with Dynamic Ratio (Absolute Value for Lateral Forces) Failure Probability >=0.2 >=0.3 >=0.4 >=0.5 >=0.6 >=0.7 >=0.8 >=0.9 True Failure 101 96 94 90 84 81 74 63 False Positive 33 33 13 5 3 3 3 2 False Negative 0 5 7 11 17 20 27 38 Confusion Matrix for Data Mining Model Evaluaon (P>=50%): Count: Percentage: Actual Actual Failure Predicon Non-Failure Failure 90 11 Non-Failure 5 28 Failure Predicon Non-Failure Failure 89% 11% Non-Failure 15% 85% 120 100 Probability of Type II Error Probability of Type I Error Counts 80 60 40 20 0 True Failure Sum of Errors False Positive False Negative Threshold with minimum sum of errors 0 0.2 0.4 0.6 0.8 1 Probability of Failure

Regression Model: Data Structure for Data Analysis based on WPD Potential Failure Logistic Regression: P(z) = 1 1+ e z z = 1.1029-0.0416x 1 + 0.0272x 2 The critical variables in the model (0.1 significance): Rim Thickness(0.05) and Flange Angle(0.1), respectively Significance Test of Logistic Regression: H 0 : coefficients=0; H 1 : O.W. The test is rejected, therefore the regression model between the dependent and independent variables is significant with p-value=0.031

Test of Logistic Regression using the 15% Unseen WPD Data Failure Probability >=0.2 >=0.3 >=0.4 >=0.5 >=0.6 >=0.7 >=0.8 >=0.9 True Failure 3 3 3 2 1 0 0 0 False Positive 8 5 2 0 0 0 0 0 False Negative 2 2 2 3 4 5 5 5 Confusion Matrix for Data Mining Model Evaluaon (P>=50%): Count: Percentage: Actual Actual Failure Predicon Non-Failure Failure 2 3 Non-Failure 0 10 Failure Predicon Non-Failure Failure 40% 60% Non-Failure 0% 100% Probability of Type I Error Counts 12 10 Probability of Type II Error 8 6 4 2 0 Sum of Errors True Failure False Negative 0 0.2 0.4 0.6 0.8 1 Probability of Failure Threshold with minimum sum of errors False Positive

Regression Model: Data Structure for Data Analysis based on WPD and WILD Potential Failure (Absolute Value for Lateral Forces) Logistic Regression (if Dynamic Ratio is ignored): P(z) = 1 1+ e z z = 2.4402 +0.0300x 1 The critical variables in the model (0.001 significance): Vertical Peak Force (0.001) Significance Test of Logistic Regression: H 0 : coefficients=0; H 1 : O.W. The test is rejected, therefore the regression model between the dependent and independent variables is significant with p-value=6.283e-10

Test of Logistic Regression using the 15% Unseen WPD & WILD Data: without Dynamic Ratio (Absolute Value for Lateral Forces) Failure Probability >=0.2 >=0.3 >=0.4 >=0.5 >=0.6 >=0.7 >=0.8 >=0.9 True Failure 1 1 1 1 1 1 1 1 False Positive 0 0 0 0 0 0 0 0 False Negative 4 4 4 4 4 4 4 4 Confusion Matrix for Data Mining Model Evaluaon: Predicon Count: Percentage: Actual Actual Failure Non-Failure Failure 1 4 Non-Failure 0 10 Failure Predicon Non-Failure Failure 20% 80% Non-Failure 0% 100% Probability of Type II Error Probability of Type I Error Counts 4.5 4 3.5 3 2.5 2 1.5 1 0.5 0 False Negative True Failure False Positive 0 0.2 0.4 0.6 0.8 1 Probability of Failure

Regression Model: Data Structure for Data Analysis based on WPD and WILD Potential Failure (Absolute Value for Lateral Forces) Logistic Regression (if Dynamic Ratio is considered): P(z) = 1 1+ e z z = 2.3057 +0.0317x 1 + 0.0107x 2 The critical variables in the model (0.1 significance): Dynamic Ratio (0.001), and Lateral Average Force(0.1), respectively Significance Test of Logistic Regression: H 0 : coefficients=0; H 1 : O.W. The test is rejected, therefore the regression model between the dependent and independent variables is significant with p-value=1.305e-09

Test of Logistic Regression using the 15% Unseen WPD & WILD Data: with Dynamic Ratio (Absolute Value for Lateral Forces) Failure Probability >=0.2 >=0.3 >=0.4 >=0.5 >=0.6 >=0.7 >=0.8 >=0.9 True Failure 3 3 1 1 1 1 1 1 False Positive 0 0 0 0 0 0 0 0 False Negative 2 2 4 4 4 4 4 4 Confusion Matrix for Data Mining Model Evaluaon (P>=30%): Count: Percentage: Actual Actual Failure Predicon Non-Failure Failure 3 2 Non-Failure 0 10 Failure Predicon Non-Failure Failure 60% 40% Non-Failure 0% 100% Probability of Type I Error Counts 4.5 1.5 0.5 Probability of Type II Error 4 3.5 3 2.5 2 1 0 True Failure Sum of Errors False Positive 0 0.2 0.4 0.6 0.8 1 Probability of Failure