Optimal Brain Damage is an approach to identify and remove unimportant weights from a neural network. The procedure involves training the network, computing the second derivatives for each parameter, computing the saliencies, sorting the parameters by saliency, and deleting some low-saliency parameters.
The saliency of a parameter is defined as the change in the objective function caused by deleting that parameter. Using the second derivative of the objective function with respect to the parameters avoids the prohibitive labor of directly evaluating the saliency by temporarily deleting each parameter and reevaluating the objective function. They approximate the objective function by a Taylor series where a perturbation of the parameter vector will change the objective function by
where is a component of the perturbed parameters , is the component of the gradient of with respect to , and are elements of the Hessian matrix of with respect to . Second order methods are typically difficult with neural networks due to the enormity of the Hessian matrix. Optimal Brain Damage introduces a simple diagonal approximation where the change in the objective function caused by deleting several parameters is the sum of the changes caused by deleting each parameter individually. Cross terms of the Hessian are neglected so the third term in can be discarded. An extremal approacimation assumes that parameter deletion will be performed after training has converged. The parameter vector is then at a local minimum of such that the first term of can be neglected. These simplifications reduce to:
where the diagonal terms of the second derivatives are given by:
Given a standard formula for computing network state, and , where is the state of unit , is the weighted sum of the input, is the connection from unit to , and is the activation function, the summand can be expanded to
And back propagated from layer to layer with a boundary condition at the output layer.