The NumPy function np.std
takes an optional parameter ddof
: "Delta Degrees of Freedom". By default, this is 0
. Set it to 1
to get the MATLAB result:
>>> np.std([1,3,4,6], ddof=1)
2.0816659994661326
To add a little more context, in the calculation of the variance (of which the standard deviation is the square root) we typically divide by the number of values we have.
But if we select a random sample of N
elements from a larger distribution and calculate the variance, division by N
can lead to an underestimate of the actual variance. To fix this, we can lower the number we divide by (the degrees of freedom) to a number less than N
(usually N-1
). The ddof
parameter allows us change the divisor by the amount we specify.
Unless told otherwise, NumPy will calculate the biased estimator for the variance (ddof=0
, dividing by N
). This is what you want if you are working with the entire distribution (and not a subset of values which have been randomly picked from a larger distribution). If the ddof
parameter is given, NumPy divides by N - ddof
instead.
The default behaviour of MATLAB's std
is to correct the bias for sample variance by dividing by N-1
. This gets rid of some of (but probably not all of) of the bias in the standard deviation. This is likely to be what you want if you're using the function on a random sample of a larger distribution.
The nice answer by @hbaderts gives further mathematical details.
std([1 3 4 6],1)
is equivalent to NumPy's defaultnp.std([1,3,4,6])
. All of this is quite clearly explained in the documentation for Matlab and NumPy, so I strongly recommend that the OP be sure to read those in the future. – Linnell