check_grad — SciPy v1.15.2 Manual (original) (raw)

scipy.optimize.

scipy.optimize.check_grad(func, grad, x0, *args, epsilon=1.4901161193847656e-08, direction='all', rng=None)[source]#

Check the correctness of a gradient function by comparing it against a (forward) finite-difference approximation of the gradient.

Parameters:

funccallable func(x0, *args)

Function whose derivative is to be checked.

gradcallable grad(x0, *args)

Jacobian of func.

x0ndarray

Points to check grad against forward difference approximation of grad using func.

args\*args, optional

Extra arguments passed to func and grad.

epsilonfloat, optional

Step size used for the finite difference approximation. It defaults tosqrt(np.finfo(float).eps), which is approximately 1.49e-08.

directionstr, optional

If set to 'random', then gradients along a random vector are used to check grad against forward difference approximation using func. By default it is 'all', in which case, all the one hot direction vectors are considered to check grad. If func is a vector valued function then only 'all' can be used.

rng{None, int, numpy.random.Generator}, optional

If rng is passed by keyword, types other than numpy.random.Generator are passed to numpy.random.default_rng to instantiate a Generator. If rng is already a Generator instance, then the provided instance is used. Specify rng for repeatable function behavior.

If this argument is passed by position or seed is passed by keyword, legacy behavior for the argument seed applies:

Changed in version 1.15.0: As part of the SPEC-007transition from use of numpy.random.RandomState tonumpy.random.Generator, this keyword was changed from seed to rng. For an interim period, both keywords will continue to work, although only one may be specified at a time. After the interim period, function calls using the_seed_ keyword will emit warnings. The behavior of both seed and_rng_ are outlined above, but only the rng keyword should be used in new code.

The random numbers generated affect the random vector along which gradients are computed to check grad. Note that rng is only used when _direction_argument is set to ‘random’.

Returns:

errfloat

The square root of the sum of squares (i.e., the 2-norm) of the difference between grad(x0, *args) and the finite difference approximation of grad using func at the points x0.

Examples

import numpy as np def func(x): ... return x[0]**2 - 0.5 * x[1]**3 def grad(x): ... return [2 * x[0], -1.5 * x[1]**2] from scipy.optimize import check_grad check_grad(func, grad, [1.5, -1.5]) 2.9802322387695312e-08 # may vary rng = np.random.default_rng() check_grad(func, grad, [1.5, -1.5], ... direction='random', seed=rng) 2.9802322387695312e-08