Home >Technology peripherals >AI >How can unsupervised machine learning benefit industrial automation?
Modern industrial environments are filled with sensors and smart components, and all of these devices together produce a wealth of data. This data, untapped in most factories today, powers a variety of exciting new applications. In fact, according to IBM, the average factory generates 1TB of production data every day. However, only about 1% of data is turned into actionable insights.
Machine learning (ML) is a fundamental technology designed to leverage this data and unlock massive amounts of value. Using training data, machine learning systems can build mathematical models that teach a system to perform specific tasks without explicit instructions.
ML uses algorithms that act on data to make decisions largely without human intervention. The most common form of machine learning in industrial automation is supervised machine learning, which uses large amounts of historical data labeled by humans to train models (i.e., the training of human-supervised algorithms).
This is useful for well-known problems such as bearing defects, lubrication failures, or product defects. Where supervised machine learning falls short is when enough historical data isn't available, labeling is too time-consuming or expensive, or users aren't sure exactly what they're looking for in the data. This is where unsupervised machine learning comes into play.
Unsupervised machine learning aims to operate on unlabeled data using algorithms that are good at recognizing patterns and pinpointing anomalies in the data. Properly applied unsupervised machine learning serves a variety of industrial automation use cases, from condition monitoring and performance testing to cybersecurity and asset management.
Supervised machine learning is easier to perform than unsupervised machine learning. With a properly trained model, it can provide very consistent, reliable results. Supervised machine learning may require large amounts of historical data - as is needed to include all relevant cases, i.e., in order to detect product defects, the data needs to contain a sufficient number of cases of defective products. Labeling these massive data sets can be time-consuming and expensive. Furthermore, training models is an art. It requires large amounts of data, properly organized, to produce good results.
Today, the process of benchmarking different ML algorithms is significantly simplified using tools like AutoML. At the same time, over-constraining the training process may result in a model that performs well on the training set but performs poorly on real data. Another key drawback is that supervised machine learning is not very effective at identifying unexpected trends in data or discovering new phenomena. For these types of applications, unsupervised machine learning can provide better results.
Compared with supervised machine learning, unsupervised machine learning only operates on unlabeled inputs . It provides powerful tools for data exploration to discover unknown patterns and correlations without human help. The ability to operate on unlabeled data saves time and money and enables unsupervised machine learning to operate on the data as soon as the input is generated.
The downside is that unsupervised machine learning is more complex than supervised machine learning. It is more expensive, requires a higher level of expertise, and often requires more data. Its output tends to be less reliable than supervised ML, and ultimately requires human supervision for optimal results.
The three important forms of unsupervised machine learning techniques are clustering, anomaly detection and data dimensionality reduction.
As the name suggests, clustering involves analyzing a data set to identify shared characteristics between the data and group similar instances together . Because clustering is an unsupervised ML technique, the algorithm (rather than a human) determines the ranking criteria. Therefore, clustering can lead to surprising discoveries and is an excellent data exploration tool.
To give a simple example: imagine three people being asked to sort fruits in a production department. One might sort by fruit type -- citrus, stone fruit, tropical fruit, etc.; another might sort by color; and a third might sort by shape. Each method highlights a different set of characteristics.
Clustering can be divided into many types. The most common are:
Mutually exclusive clustering (Exclusive Clustering): A data instance is exclusively assigned to a cluster.
Fuzzy or overlapping clustering (Fuzzy Clustering): A data instance can be assigned to multiple clusters. For example, oranges are both citrus and tropical fruits. In the case of unsupervised ML algorithms operating on unlabeled data, it is possible to assign a probability that a data block correctly belongs to group A versus group B.
Hierarchical clustering: This technique involves building a hierarchical structure of clustered data, rather than a set of clusters. Oranges are citrus fruits, but they are also included in the larger spherical fruit group and can be further absorbed by all fruit groups.
Let’s look at a set of the most popular clustering algorithms:
K-mean (K-means) algorithm classifies data into K clusters, where the value of K is preset by the user. At the beginning of the process, the algorithm randomly assigns K data points as centroids for K clusters. Next, it calculates the mean between each data point and the centroid of its cluster. This results in resorting the data to the cluster. At this point, the algorithm recalculates the centroid and repeats the mean calculation. It repeats the process of calculating centroids and reordering clusters until it reaches a constant solution (see Figure 1).
Figure 1: K-means algorithm divides the data set into K clusters, first randomly selecting K data points as centroids, and then randomly distribute the remaining instances across the cluster.
K-means algorithm is simple and efficient. It is very useful for pattern recognition and data mining. The downside is that it requires some advanced knowledge of the dataset to optimize the setup. It is also disproportionately affected by outliers.
The K-median algorithm is a close relative of K-means . It uses essentially the same process, except instead of calculating the mean of each data point, it calculates the median. Therefore, the algorithm is less sensitive to outliers.
Here are some common use cases for cluster analysis:
Anomaly detection is critical for a variety of use cases from defect detection to condition monitoring to cybersecurity. This is a key task in unsupervised machine learning. There are several anomaly detection algorithms used in unsupervised machine learning, let’s take a look at the two most popular ones:
The standard method of anomaly detection is to establish a set of normal values and then analyze each piece of data to see whether and how much it deviates from the normal value. This is a very time-consuming process when working with massive data sets of the kind used in ML. The isolation forest algorithm takes the opposite approach. It defines outliers as being neither common nor very different from other instances in the data set. Therefore, they are more easily isolated from the rest of the dataset on other instances.
The isolation forest algorithm has minimal memory requirements and the time required is linearly related to the size of the data set. They can handle high-dimensional data even if it involves irrelevant attributes.
One of the challenges of identifying outliers only by their distance from the centroid Yes, data points that are a short distance from a small cluster may be outliers, while points that appear to be far away from a large cluster may not be. The LOF algorithm is designed to make this distinction.
LOF defines an outlier as a data point with a local density deviation that is much greater than its neighboring data points (see Figure 2). Although like K-means it does require some user setup ahead of time, it can be very effective. It can also be applied to novelty detection when used as a semi-supervised algorithm and trained on normal data only.
Figure 2: Local Outlier Factor (LOF) uses the local density deviation of each data point to calculate the anomaly score , thereby distinguishing normal data points from outliers.
The following are a few use cases for anomaly detection:
Machine learning is based on large amounts of data, often very large amounts. It’s one thing to filter a data set with ten to dozens of features. Datasets with thousands of features (and they certainly exist) can be overwhelming. Therefore, the first step in ML can be dimensionality reduction to reduce the data to the most meaningful features.
A common algorithm used for dimensionality reduction, pattern recognition, and data exploration is principal component analysis (PCA). A detailed discussion of this algorithm is beyond the scope of this article. Arguably it can help identify mutually orthogonal data subsets, i.e. they can be removed from the data set without affecting the main analysis. PCA has several interesting use cases:
Unsupervised machine learning is not better or worse than supervised machine learning. For the right project, it can be very effective. That said, the best rule of thumb is to keep it simple, so unsupervised machine learning is generally only used on problems that supervised machine learning cannot solve.
Think about the following questions to determine which machine learning approach is best for your project:
Finally, here are some tips to help ensure success:
Data collected in industrial settings can be a valuable resource, but only if properly harnessed. Unsupervised machine learning can be a powerful tool for analyzing data sets to extract actionable insights. Adopting this technology can be challenging, but it can provide a significant competitive advantage in a challenging world.
The above is the detailed content of How can unsupervised machine learning benefit industrial automation?. For more information, please follow other related articles on the PHP Chinese website!