scipy.spatial.distance.jensenshannon#

scipy.spatial.distance.jensenshannon(p, q, base=None, *, axis=0, keepdims=False)[source]#

Compute the Jensen-Shannon distance (metric) between two probability arrays. This is the square root of the Jensen-Shannon divergence.

The Jensen-Shannon distance between two probability vectors p and q is defined as,

\[\sqrt{\frac{D(p \parallel m) + D(q \parallel m)}{2}}\]

where \(m\) is the pointwise mean of \(p\) and \(q\) and \(D\) is the Kullback-Leibler divergence.

This routine will normalize p and q if they don’t sum to 1.0.

Parameters
p(N,) array_like

left probability vector

q(N,) array_like

right probability vector

basedouble, optional

the base of the logarithm used to compute the output if not given, then the routine uses the default base of scipy.stats.entropy.

axisint, optional

Axis along which the Jensen-Shannon distances are computed. The default is 0.

New in version 1.7.0.

keepdimsbool, optional

If this is set to True, the reduced axes are left in the result as dimensions with size one. With this option, the result will broadcast correctly against the input array. Default is False.

New in version 1.7.0.

Returns
jsdouble or ndarray

The Jensen-Shannon distances between p and q along the axis.

Notes

New in version 1.2.0.

Examples

>>> from scipy.spatial import distance
>>> distance.jensenshannon([1.0, 0.0, 0.0], [0.0, 1.0, 0.0], 2.0)
1.0
>>> distance.jensenshannon([1.0, 0.0], [0.5, 0.5])
0.46450140402245893
>>> distance.jensenshannon([1.0, 0.0, 0.0], [1.0, 0.0, 0.0])
0.0
>>> a = np.array([[1, 2, 3, 4],
...               [5, 6, 7, 8],
...               [9, 10, 11, 12]])
>>> b = np.array([[13, 14, 15, 16],
...               [17, 18, 19, 20],
...               [21, 22, 23, 24]])
>>> distance.jensenshannon(a, b, axis=0)
array([0.1954288, 0.1447697, 0.1138377, 0.0927636])
>>> distance.jensenshannon(a, b, axis=1)
array([0.1402339, 0.0399106, 0.0201815])