Home >Technology peripherals >AI >Ten commonly used distance measurement methods in machine learning
Distance metric is the basis of supervised and unsupervised learning algorithms, including k-nearest neighbor, support vector machine and k-means clustering.
The choice of distance metric affects our machine learning results, so it is important to consider which metric is best suited for the problem. Therefore, we should be cautious when deciding which measurement method to use. But before we make a decision, we need to understand how distance measurement works and what measurements we can choose from.
This article will briefly introduce commonly used distance measurement methods, how they work, how to calculate them in Python, and when to use them. This deepens knowledge and understanding and improves machine learning algorithms and results.
Before we delve deeper into the different distance measurements, we first need to have a general idea of how they work and how to choose the appropriate measurement.
Distance metric is used to calculate the difference between two objects in a given problem space, i.e. features in the dataset. This distance can then be used to determine the similarity between features, with the smaller the distance, the more similar the features.
For distance measurement, we can choose between geometric distance measurement and statistical distance measurement. Which distance measurement should be chosen depends on the type of data. Features may be of different data types (e.g., real values, Boolean values, categorical values), and the data may be multidimensional or consist of geospatial data.
Euclidean distance measures the shortest distance between two real-valued vectors. Due to its intuitiveness, simplicity of use and good results for many use cases, it is the most commonly used distance metric and the default distance metric for many applications.
Euclidean distance can also be called l2 norm, and its calculation method is:
The Python code is as follows
from scipy.spatial import distance distance.euclidean(vector_1, vector_2)
Euclidean distance has two main disadvantages. First, distance measurements do not work with data of higher dimensions than 2D or 3D space. Second, if we do not normalize and/or normalize the features, distances may be skewed by units.
Manhattan distance is also called taxi or city block distance because the distance between two real-valued vectors is based on the fact that a person can only move at right angles computational. This distance measure is often used for discrete and binary attributes so that true paths can be obtained.
The Manhattan distance is based on the l1 norm, and the calculation formula is:
The Python code is as follows
from scipy.spatial import distance distance.cityblock(vector_1, vector_2)
The distance of Manhattan has two major disadvantages. It is not as intuitive as Euclidean distance in high-dimensional space, nor does it show the shortest possible path. While this may not be a problem, we should be aware that this is not the shortest distance.
The Chebyshev distance is also called the checkerboard distance because it is the maximum distance in any dimension between two real-valued vectors. It is commonly used in warehouse logistics, where the longest path determines the time required to get from one point to another.
The Chebyshev distance is calculated by l-infinite norm:
The Python code is as follows
from scipy.spatial import distance distance.chebyshev(vector_1, vector_2)
Chebyshev distance has only very specific use cases and is therefore rarely used.
Minkowski distance is the generalized form of the above distance measure. It can be used for the same use cases while providing high flexibility. We can choose the p-value to find the most appropriate distance measure.
The calculation method of Minkowski distance is:
The Python code is as follows
from scipy.spatial import distance distance.minkowski(vector_1, vector_2, p)
Because Minkowski distance represents a different distance metric and has the same main drawbacks as them, such as problems in high-dimensional spaces and dependence on feature units. Additionally, the flexibility of p-values can also be a disadvantage, as it can reduce computational efficiency since multiple calculations are required to find the correct p-value.
Cosine similarity is a measure of direction. Its size is determined by the cosine between two vectors and ignores the size of the vector. Cosine similarity is often used in high dimensions where the size of the data does not matter, for example, recommender systems or text analysis.
余弦相似度可以介于-1(相反方向)和1(相同方向)之间,计算方法为:
余弦相似度常用于范围在0到1之间的正空间中。余弦距离就是用1减去余弦相似度,位于0(相似值)和1(不同值)之间。
Python代码如下
from scipy.spatial import distance distance.cosine(vector_1, vector_2)
余弦距离的主要缺点是它不考虑大小而只考虑向量的方向。因此,没有充分考虑到值的差异。
半正矢距离测量的是球面上两点之间的最短距离。因此常用于导航,其中经度和纬度和曲率对计算都有影响。
半正矢距离的公式如下:
其中r为球面半径,φ和λ为经度和纬度。
Python代码如下
from sklearn.metrics.pairwise import haversine_distances haversine_distances([vector_1, vector_2])
半正矢距离的主要缺点是假设是一个球体,而这种情况很少出现。
汉明距离衡量两个二进制向量或字符串之间的差异。
对向量按元素进行比较,并对差异的数量进行平均。如果两个向量相同,得到的距离是0之间,如果两个向量完全不同,得到的距离是1。
Python代码如下
from scipy.spatial import distance distance.hamming(vector_1, vector_2)
汉明距离有两个主要缺点。距离测量只能比较相同长度的向量,它不能给出差异的大小。所以当差异的大小很重要时,不建议使用汉明距离。
统计距离测量可用于假设检验、拟合优度检验、分类任务或异常值检测。
Jaccard指数用于确定两个样本集之间的相似性。它反映了与整个数据集相比存在多少一对一匹配。Jaccard指数通常用于二进制数据比如图像识别的深度学习模型的预测与标记数据进行比较,或者根据单词的重叠来比较文档中的文本模式。
Jaccard距离的计算方法为:
Python代码如下
from scipy.spatial import distance distance.jaccard(vector_1, vector_2)
Jaccard指数和距离的主要缺点是,它受到数据规模的强烈影响,即每个项目的权重与数据集的规模成反比。
Sörensen-Dice指数类似于Jaccard指数,它可以衡量的是样本集的相似性和多样性。该指数更直观,因为它计算重叠的百分比。Sörensen-Dice索引常用于图像分割和文本相似度分析。
计算公式如下:
Python代码如下
from scipy.spatial import distance distance.dice(vector_1, vector_2)
它的主要缺点也是受数据集大小的影响很大。
动态时间规整是测量两个不同长度时间序列之间距离的一种重要方法。可以用于所有时间序列数据的用例,如语音识别或异常检测。
为什么我们需要一个为时间序列进行距离测量的度量呢?如果时间序列长度不同或失真,则上述面说到的其他距离测量无法确定良好的相似性。比如欧几里得距离计算每个时间步长的两个时间序列之间的距离。但是如果两个时间序列的形状相同但在时间上发生了偏移,那么尽管时间序列非常相似,但欧几里得距离会表现出很大的差异。
动态时间规整通过使用多对一或一对多映射来最小化两个时间序列之间的总距离来避免这个问题。当搜索最佳对齐时,这会产生更直观的相似性度量。通过动态规划找到一条弯曲的路径最小化距离,该路径必须满足以下条件:
我们可以使用 Python 中的 fastdtw 包:
from scipy.spatial.distance import euclidean from fastdtw import fastdtw distance, path = fastdtw(timeseries_1, timeseries_2, dist=euclidean)
动态时间规整的一个主要缺点是与其他距离测量方法相比,它的计算工作量相对较高。
在这篇文章中,简要介绍了十种常用的距离测量方法。本文中已经展示了它们是如何工作的,如何在Python中实现它们,以及经常使用它们解决什么问题。如果你认为我错过了一个重要的距离测量,请留言告诉我。
The above is the detailed content of Ten commonly used distance measurement methods in machine learning. For more information, please follow other related articles on the PHP Chinese website!