Distance Metric
Defining the distance metric is very important to have a good performing model.
https://medium.com/analytics-vidhya/role-of-distance-metrics-in-machine-learning-e43391a6bf2e
Suppose two objects and both have features
From Lecture 13of Carnegie. The Minskowski metric/distance is defined by
This is a generalization of two Distance Metrics you are very familiar with in ML:
- L1 Distance (Manhattan Distance) ()
- L2 Distance (Euclidean Distance) ()
- Chebyshev Distance “sup” distance ()
Other Distance metrics
-
Edit Distance: General technique for measuring similarity, where we look at the amount of effort to transform one object into another. OH, Earth Mover’s Distance is an edit distance metric then?
Properties that every distance metric should have
- Symmetry:
- Otherwise you can claim that Alex looks like Bob, but Bob looks nothing like Alex
- Constancy of Self-Similarity ,
- Otherwise Alex looks more like Bob, that Bob does
- Positivity Separation
- Otherwise if there are objects in your world that are different, but you cannot tell apart
- Triangle Inequality
- Otherwise you could claim Alex is very like Bob, and Alex is very like Carl, but Bob is very unlike Carl
I also see non-negativity on wikipedia:
- Non-Negativity:
Examples
Taken from wikipedia
Metrics
- Total variation distance (sometimes just called “the” statistical distance)
- Hellinger distance
- Lévy–Prokhorov metric
- Wasserstein metric: also known as the Kantorovich metric, or earth mover’s distance
- Mahalanobis Distance
Divergences
- Kullback-Leibler Divergence
- Rényi’s divergence
- Jensen–Shannon divergence
- Bhattacharyya distance (despite its name it is not a distance, as it violates the triangle inequality)
- f-divergence: generalizes several distances and divergences
- Discriminability index, specifically the Bayes discriminability index is a positive-definite symmetric measure of the overlap of two distributions.