Skip to content

A predictive maintenance framework contrasting feed-forward and sequential machine learning models to forecast industrial machinery failures, achieving near-perfect AUROC scores with GRU models.

License

Notifications You must be signed in to change notification settings

alex-w-99/ML-predictive-machinery-maintenance

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

29 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Machine Learning for Predictive Maintenance in Industrial Machinery

This project utilizes and contrasts several machine learning techniques — including logistic regression, artificial neural networks (ANNs), and sequential ANNs — to address the challenge of preemptive equipment failure detection in industrial machinery. Using the Microsoft Azure Predictive Maintenance dataset, this project develops a comprehensive predictive maintenance system. It showcases how to enhance safety and minimize economic losses in industrial settings by deriving actionable insights that promote proactive maintenance strategies.

To ensure the effectiveness of the models, this project addresses various challenges such as computational limitations, feature engineering, and potential overfitting. A structured train-validate-test framework is employed, focusing on metrics such as AUROC, accuracy, precision, and recall, with some models achieving validation and test AUROC figures exceeding 0.999. Through these efforts, the project demonstrates how advanced predictive techniques can minimize downtime, ensure operational safety, and extend the lifespan of industrial machinery.

The notebooks in this repository are organized as follows:

  1. notebooks/data_exploration/: Contains notebooks for initial data visualization and exploration to understand the data's structure, distributions, etc.
  2. notebooks/preprocess_data/: Includes notebooks that handle data cleaning, transformation, and preparation tasks to convert raw data into a format suitable for analysis.
  3. notebooks/create_features_targets/: Notebooks in this directory focus on engineering the features (e.g., summarizing a 24-hour window to later make a prediction for) and preparing target variables (e.g., predicting future failure 24-hours into the future) based on different forecasting windows.
  4. notebooks/run_experiments/: This directory houses notebooks that run various machine learning experiments, comparing model performances and documenting results.

To get started with running the code yourself, please see the Usage section below.

Data and Preprocessing

This project uses the Microsoft Azure Predictive Maintenance dataset which includes machine data over the year 2015 from 5 different sources: maintenance records, failure history, error logs, static machine attributes such as model and age, and hourly telemetry data (hourly readings on voltage, rotation, pressure, and vibration). These data sources are preprocessed to merge into a unified dataset where each machine is represented by hourly records over the year 2015.

  • Feature Data: The data is structured into discrete 24-hour windows to effectively model both static and dynamic aspects of machine performance. Experimentation with different time windows (12-, 48-, 72-, and 168-hours) confirmed that a 24-hour window was the smallest interval that achieved consistent, high-quality model performance, beyond which longer windows offered equivalent (or inferior) performance.
    • Feedforward Models: For logistic regression and MLPs, features within each window are summarized into a single vector using statistical measures such as mean, standard deviation, skewness, kurtosis, etc., resulting in 45 features per window.
    • Sequential Models: In contrast, RNN, LSTM, and GRU process each 24-hour window as a sequence of hourly snapshots, with 25 features per hour to maintain temporal relationships.
  • Target Data: Output labels are generated to indicate whether a machine will fail within the subsequent 6- or 24-hours, enabling the models to predict short-term and medium-term failures.

To ensure accurate training and evaluation, features and labels are carefully aligned to avoid data leakage.

Methodology

To ensure thorough evaluation and prevent overfitting, the dataset was partitioned randomly with 60% of data for training, 20% for validation, and 20% for testing. Each "example" corresponds to a 24-hour period for a given machine, and whether it failed in the subsequent 24-hour period. For ANNs, training was conducted over 200 epochs using binary cross entropy to optimize the models effectively without overfitting.

  • Logistic Regression served primarily for illustrative purposes to establish a baseline, demonstrating the non-linear relationship in the dataset where simpler linear models might fail.
  • MLP and Sequential Models (RNN, LSTM, GRU) were fine-tuned to optimize parameters like learning rate, batch size, and dropout rates over several experiments (not all included in notebooks). For sequential models, specific attention was paid to optimizing RNN dropout and hidden layers.

Results

The evaluation of the models yielded varied results. Logistic regression, a linear model, showed suprisingly respectable AUROC performance but struggled with precision-recall, indicating its limitations in addressing the nuanced complexities of the dataset. Meanwhile, the MLP unexpectedly outperformed both the RNN and LSTM in across many metrics, serving as a reminder of both the difficulties of sequence processing and that more complex models are not always better. Finally, the standout performer was the GRU model, which achieved nearly perfect AUROC scores in both validation and test phases.

For all results information, please see the notebook files in notebooks/run_experiments/.

Discussion

Logistic regression was quite effective despite its simplicity, suggesting that many relationships between features and targets may exhibit linear characteristics in this case. The RNN and LSTM models, specifically designed to handle sequences, did not perform as expected, likely suffering from vanishing or exploding gradients (exacerbated by the length of input sequences, 24 hours). The MLP's strong performance indicates that less complex models can be highly effective with thorough feature engineering and appropriate hyperparameter tuning. The GRU's superior results reinforce the importance of choosing the right model architecture that aligns with the specific dynamics of the dataset.

This project ultimately demonstrates that machine learning can significantly enhance predictive maintenance, adeptly handling non-linear relationships which are ubiquitous in industrial settings and exemplifying how ML can effectively tackle non-linear problems across many domains.

Model Performance Snapshots

...

Usage

To properly utilize this project, ensure that Python can import from the project's src/ and data/ directories, necessary for the Jupyter notebooks to import the necessary modules and data files.

Environment Setup

First, set up the environment using Conda:

# create environment from project directory
$ conda env create -f environment.yml
$ conda activate ml-pred-mchn
Running the Notebooks

With the environment activated, launch Jupyter to run the notebooks:

(ml-pred-mchn) $ jupyter notebook

This will start Jupyter. Open and execute the project notebooks to explore the code. When Jupyter is launched from the root of the project directory, it automatically configures the path to include the src/ and data/ directories, enabling straightforward imports within the notebooks.

Areas for Future Improvement

  • Refine models to predict individual failure codes (rather than just any failure) to enhance model robustness.
  • Investigate hybrid models that combine MLPs for "static" features (e.g., machine make) and RNNs for temporal data (e.g., rotation) to improve prediction accuracy.
  • Explore advanced ensemble methods to improve performance.

Acknowledgements

Contact Information

  • Alexander Wilcox
  • Email: alexander.w.wilcox [at] gmail [dot] com

About

A predictive maintenance framework contrasting feed-forward and sequential machine learning models to forecast industrial machinery failures, achieving near-perfect AUROC scores with GRU models.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published