Euclidean distance
In mathematics, the Euclidean distance or Euclidean metric is the "ordinary" straightline distance between two points in Euclidean space. With this distance, Euclidean space becomes a metric space. The associated norm is called the Euclidean norm. Older literature refers to the metric as the Pythagorean metric. A generalized term for the Euclidean norm is the L^{2} norm or L^{2} distance.
Contents
Definition[edit]
The Euclidean distance between points p and q is the length of the line segment connecting them ().
In Cartesian coordinates, if p = (p_{1}, p_{2},..., p_{n}) and q = (q_{1}, q_{2},..., q_{n}) are two points in Euclidean nspace, then the distance (d) from p to q, or from q to p is given by the Pythagorean formula:^{[1]}

(1)
The position of a point in a Euclidean nspace is a Euclidean vector. So, p and q may be represented as Euclidean vectors, starting from the origin of the space (initial point) with their tips (terminal points) ending at the two points. The Euclidean norm, or Euclidean length, or magnitude of a vector measures the length of the vector:^{[1]}
where the last expression involves the dot product.
Describing a vector as a directed line segment from the origin of the Euclidean space (vector tail), to a point in that space (vector tip), its length is actually the distance from its tail to its tip. The Euclidean norm of a vector is seen to be just the Euclidean distance between its tail and its tip.
The relationship between points p and q may involve a direction (for example, from p to q), so when it does, this relationship can itself be represented by a vector, given by
In a two or threedimensional space (n = 2, 3), this can be visually represented as an arrow from p to q. In any space it can be regarded as the position of q relative to p. It may also be called a displacement vector if p and q represent two positions of some moving point.
The Euclidean distance between p and q is just the Euclidean length of this displacement vector:

(2)
which is equivalent to equation 1, and also to:
One dimension[edit]
In the context of Euclidean geometry, a metric is established in one dimension by fixing two points on a line, and choosing one to be the origin. The length of the line segment between these points defines the unit of distance and the direction from the origin to the second point is defined as the positive direction. This line segment may be translated along the line to build longer segments whose lengths correspond to multiples of the unit distance. In this manner real numbers can be associated to points on the line (as the distance from the origin to the point) and these are the Cartesian coordinates of the points on what may now be called the real line. As an alternate way to establish the metric, instead of choosing two points on the line, choose one point to be the origin, a unit of length and a direction along the line to call positive. The second point is then uniquely determined as the point on the line that is at a distance of one positive unit from the origin.
The distance between any two points on the real line is the absolute value of the numerical difference of their coordinates. It is common to identify the name of a point with its Cartesian coordinate. Thus if p and q are two points on the real line, then the distance between them is given by:
In one dimension, there is a single homogeneous, translationinvariant metric (in other words, a distance that is induced by a norm), up to a scale factor of length, which is the Euclidean distance. In higher dimensions there are other possible norms.
Two dimensions[edit]
In the Euclidean plane, if p = (p_{1}, p_{2}) and q = (q_{1}, q_{2}) then the distance is given by
This is equivalent to the Pythagorean theorem.
Alternatively, it follows from (2) that if the polar coordinates of the point p are (r_{1}, θ_{1}) and those of q are (r_{2}, θ_{2}), then the distance between the points is
Three dimensions[edit]
In threedimensional Euclidean space, the distance is
n dimensions[edit]
In general, for an ndimensional space, the distance is
Squared Euclidean distance[edit]
The square of the standard Euclidean distance, which is known as the squared Euclidean distance (SED), is also of interest; as an equation:
Squared Euclidean distance is of central importance in estimating parameters of statistical models, where it is used in the method of least squares, a standard approach to regression analysis. The corresponding loss function is the squared error loss (SEL), and places progressively greater weight on larger errors. The corresponding risk function (expected loss) is mean squared error (MSE).
Squared Euclidean distance is not a metric, as it does not satisfy the triangle inequality. However, it is a more general notion of distance, namely a divergence (specifically a Bregman divergence), and can be used as a statistical distance. The Pythagorean theorem is simpler in terms of squared distance (since there is no square root); if , then:
In information geometry, the Pythagorean identity can be generalized from SED to other Bregman divergences, including relative entropy (Kullback–Leibler divergence), allowing generalized forms of least squares to be used to solve nonlinear problems.
The SED is a smooth, strictly convex function of the two points, unlike the distance, which is not smooth when two points are equal and is not strictly convex (because it is linear). The SED is thus preferred in optimization theory, since it allows convex analysis to be used. Since squaring is a monotonic function of nonnegative values, minimizing the SED is equivalent to minimizing the Euclidean distance, so the optimization problem is equivalent in terms of either, but easier to solve using the SED.
If one of the points is fixed, the SED can be interpreted as a potential function, in which case a normalization factor of one half is used, and the sign may be switched, depending on convention. In detail, given two points , the vector points from to and has magnitude proportional to their Euclidean distance. If one fixes , one can thus define a smooth vector field "pointing at " by This is the gradient of the scalarvalued function "half SED from ", where the half cancels the two in the power rule. Writing half the squared distance from as , one has Alternatively, one can consider the vector field pointing from , and omit the minus sign.
In information geometry, the notion of a vector field of "pointing from one point to another" can be generalized to statistical manifolds – one can use an affine connection to connect tangent vectors at different points and the exponential map to flow from one point to another, and on a statistical manifold this is invertible, defining a unique "difference vector" from any given point to another. In this context, the SED (whose gradient generates the standard difference vector) generalized to a divergence that generates the information geometry of the manifold; a uniform construction of such a divergence (given the geometric structure) is called a canonical divergence.^{[2]}
In the field of rational trigonometry, the SED is referred to as quadrance.
See also[edit]
 Chebyshev distance measures distance assuming only the most significant dimension is relevant.
 Euclidean distance matrix
 Manhattan distance measures distance following only axisaligned directions.
 Minkowski distance is a generalization that unifies Euclidean distance, Manhattan distance, and Chebyshev distance.
 Pythagorean addition
 Haversine distance giving greatcircle distances between two points on a sphere from their longitudes and latitudes.
 Vincenty's formulae well known as "Vincent distance"
References[edit]
 ^ ^{a} ^{b} Anton, Howard (1994), Elementary Linear Algebra (7th ed.), John Wiley & Sons, pp. 170–171, ISBN 9780471587422
 ^ Ay & Amari 2015, 2. A New Approach to the General Inverse Problem.
 Ay, Nihat; Amari, Shunichi (2015). "A Novel Approach to Canonical Divergences within Information Geometry" (PDF). Entropy. 17 (12): 8111–8129. Bibcode:2015Entrp..17.8111A. doi:10.3390/e17127866.
Further reading[edit]
 Deza, Elena; Deza, Michel Marie (2009). Encyclopedia of Distances. Springer. p. 94.
 "Cluster analysis". March 2, 2011.