Skip to content

mkwilbux/predictive-maintenance-python

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

3 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Real-time time-series anomaly detection

Programming Language Python
Skills (beg, intermediate, advanced) Intermediate
Time to complete project (in increments of 15 min) 25 min
Hardware needed (hardware used) Up Squared* board
Target Operating System

The monitoring of manufacturing equipment is vital to any industrial process. Sometimes it is critical that equipment be monitored in real-time for faults and anomalies to prevent damage and correlate equipment behavior faults to production line issues. Fault detection is the pre-cursor to predictive maintenance.

There are several methods which don't require training of a neural network to be able to detect failures, starting with the most basic (FFT), to the most complex (Gaussian Mixture Model). These have the advantage of being able to be re-used with minor modifications on different data streams, and don't require a lot of known previously classified data (unlike neural nets). In fact, some of these methods can be used to classify data in order to train DNNs.

What you'll learn

  • Basic implementation of FFT, Logistic Regression, K-Means clustering, GMM.
  • How FFT is helpful in Feature engineering of vibrational data of a machine.

Setup

Gather your materials

Get the code

Open the example in a console or any python supported IDE (Spyder). Set the working directory where your code and dataset is stored.

Run the application

There are 4 samples FFT, Logistic Regression, Kmeans, GMM. For each one of these there is one or multiple files to execute.

SAMPLE FILE NAME EXPECTED OUTPUT Note
FreqTime.py Frequency v/s Time plot
Utils.py Have all the function for all the module
Train_logistic_regression.py Saved weight of the trained module in the filename "LogisticRegression" Training for the logistic Regression is done on testset1 ,bearing4_y axis (failed ),bearing 2_x axis (passed ) and testset2 bearing1(failed),bearing2(passed)*training dataset is increased for better result.
Test_logistic_regression.py Result for the particular test set which bearing is failed and the plot of last 100 predicted labels for all the bearings. Take input path from the user , he wants to check which bearing is suspected to fail, which is in normal condition
Train_kmeans.py Saved weight of the trained module in the filename "Kmeans" Training for the Kmeans is done on testset1 ,bearing4_y axis (failed ),bearing 2_x axis (passed ) and testset2 bearing1(failed),bearing2(passed)*training dataset is increased for better result.
Test_kmeans.py Result for the particular test set which bearing is failed and the plot of last 100 predicted labels for all the bearings. Take input path from the user , he wants to check which bearing is suspected to fail, which is in normal conditionThe label 0-7 is the range(early-most critical for failure)****** its able to detect the bearing 1_y axis which is very close to failue, hence showing
Train_GMM.py Saved weight of the trained module in the filename "GMM" Training for the GMM is done on testset1 ,bearing4_y axis (failed ),bearing 2_x axis (passed ) and testset2 bearing1(failed),bearing2(passed)*training dataset is increased for better result.
Test_GMM.py Result for the particular test set which bearing is failed and the plot of last 100 predicted labels for all the bearings. Take input path from the user, he wants to check which bearing is suspected to fail, which is in normal conditionThe label 0-2 is the range(early-most critical for failure)

How it works

  1. FFT: A fast Fourier transform (FFT) is an algorithm that samples a signal over a period of time (or space) and divides it into its frequency components. These components are single sinusoidal oscillations at distinct frequencies each with their own amplitude and phase. Y = fft( X) computes the discrete Fourier transform (DFT) of X using a fast Fourier transform (FFT) algorithm. If X is a vector, then fft(X) returns the Fourier transform of the vector. More details

  2. Logistic Regression: Logistic regression is a statistical method for analyzing a dataset in which there are one or more independent variables that determine an outcome. The outcome is measured with a dichotomous variable (in which there are only two possible outcomes).

    In logistic regression, the dependent variable is binary or dichotomous, i.e. it only contains data coded as 1 (TRUE, success, pregnant, etc.) or 0 (FALSE, failure, non-pregnant, etc.).The goal of logistic regression is to find the best fitting (yet biologically reasonable) model to describe the relationship between the dichotomous characteristic of interest (dependent variable = response or outcome variable) and a set of independent (predictor or explanatory) variables. More details

  3. Kmeans Clustering: K-means clustering is a type of unsupervised learning, which is used when you have unlabeled data (i.e., data without defined categories or groups). The goal of this algorithm is to find groups in the data, with the number of groups represented by the variable K. The algorithm works iteratively to assign each data point to one of K groups based on the features that are provided. Data points are clustered based on feature similarity.

    The results of the K-means clustering algorithm are:

    • The centroids of the K clusters, which can be used to label new data
    • Labels for the training data (each data point is assigned to a single cluster)

    Each centroid of a cluster is a collection of feature values which define the resulting groups. Examining the centroid feature weights can be used to qualitatively interpret what kind of group each cluster represents. More details

  4. GMM: A Gaussian mixture model is a probabilistic model that assumes all the data points are generated from a mixture of a finite number of Gaussian distributions with unknown parameters. One can think of mixture models as generalizing k-means clustering to incorporate information about the covariance structure of the data as well as the centers of the latent Gaussians.

    The GaussianMixture object implements the expectation-maximization (EM) algorithm for fitting mixture-of-Gaussian models. It can also draw confidence ellipsoids for multivariate models, and compute the Bayesian Information Criterion to assess the number of clusters in the data. A GaussianMixture.fit method is provided that learns a Gaussian Mixture Model from train data. Given test data, it can assign to each sample the Gaussian it mostly probably belongs to using the GaussianMixture.predict method.

    The GaussianMixture comes with different options to constrain the covariance of the difference classes estimated: spherical, diagonal, tied or full covariance. more details

Code Explanation:

For all the samples the basic approach is same. Following are the steps that are basic steps:

  • Take the fft of each bearing of each file.
  • Calculate the Frequency and amplitude of it
  • Calculate the top 5 amplitude and their corresponding frequency.
  • Repeat the same for each bearing and each data file. Stored the result in the result data frame.

For FFT: Gives the Frequency vs Time Plot of each maximum frequency for each dataset.

Figure 1 Figure 1.plot for the testset2, max2 frequency for all the bearing.

For Logistic regression:

Calculate the label for the data frame (in depended variable). Assume that first 70% is in normal condition and rest 71-100% are suspected to fail. For the passed one give label '0' and for suspected to fail give label '1'. Train the logistic regression on result data frame. Stored the model using numpy.

For testing: Take the input path from the user, he wants to check which bearing is suspected to fail &which is in normal condition. Check the no of one label, and no of zero label in the last 100 predictions. If the no of one prediction is more than no oh zero label, then bearing is suspected to fail else it is in normal condition.

Figure 2 Figure 2.predicted last 100 labels for testset2, for all four bearings

Figure 3 Figure 3.result for the bearing which is failed for the testset2

For Kmeans Clustering:

Fit the result data frame into Kmeans cluster , grouping into 8 cluster(no is obtained from elbow method). Save the trained model into Kmeans.

For testing, take input path from the user, he wants to check which bearing is suspected to fail &which is in normal condition. Calculate the no of label 5, 6 ,7 for the last 100 predictions if the its greater than 25 then its suspected to fail, else in normal condition.

Figure 4 Figure 4.predicted last 100 labels for testset2, for all four bearings using the Kmeans clustering

Figure 5 Figure 5. result of testset2 which bearing has failed using the Kmeans clustering

For GMM:

Fit the result data frame into gmm model, grouping into 3 components (components depicts no of clusters). Save the trained model as GMM using numpy.

For the testing, take input path from the user, he wants to check which bearing is suspected to fail &which is in normal condition. Calculate the no of label 2 for the last 100 predictions if the its greater than 50 then its suspected to fail, else in normal condition.

Figure 6 Figure 6.predicted last 100 labels for testset2, for all four bearings using the GMM clustering

Figure 7 Figure 7. result of testset2 which bearing has failed using the GMM clustering

NOTE: TESTSET 3 of NASA the bearing dataset is discarded for the observations because of the following reasons:

1: It has 6324 file in actuality, but according to the documentation its contains 4448 data file. This makes very noisy data.

2: None of the bearing indicate symptoms of failure. However, it suddenly fails. This makes data inconsistent.

The above listed reasons describe how testset3 exhibits unpredictable behavior.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%