Fréchet Derivatives 3: Deep Linear Networks
Introduction
The Fréchet derivative makes the computation of certain matrix- and vector-valued derivatives by hand substantially less painful. The first blog post in this series defined the Fréchet derivative and applied it to some very simple functions, while the second blog post applied it to the derivation of the normal equations and gradient updates for linear least squares.
In this final blog post, we will apply the Fréchet derivative to computing the gradients for a somewhat peculiar version of linear regression: the deep linear network.
Neural Networks Without All That Pesky Nonlinear Stuff
Neural networks are currently pushing forward the state of the art in machine learning and artificial intelligence. If you’re unfamiliar with them, check out this blogpost for a gentle introduction.
The most common form of neural network, a feed-forward neural network, is a parameterized function whose parameters are matrices and which looks like
where each \(\sigma_i\) is some nonlinear function. In English, we take the input \(x\) and multiply it by a series of matrices, but in between we apply nonlinear transformations.
Neural networks can mimic just about any function you’d like, provided you carefully choose the \(W_i\)s. This genericness is part of what makes them so powerful, but it also makes them hard to understand.
If we’re interested in more deeply understanding things like how neural networks learn and why they behave strangely sometimes, we need a model that’s like a neural network in certain important ways, but which we can analyze.
Enter the linear network. I wrote about single-layer linear networks in a blog post introducing linear algebra ideas to neuroscientists. A linear network chooses each \(\sigma_i\) to be the identity function, or a “pass-through” function that doesn’t touch its inputs. The result is a network that looks like
Unlike in the nonlinear case, it’s easy to write down what function the network is computing. If we denote by \(W\) the matrix that results from multiplying all of the \(L\) matrices together, we obtain
and we know quite a bit about functions which can be represented by matrices. These linear functions, in fact, are just about the only functions we really understand.
Computing the Loss Gradients of a Deep Linear Network
This derivation is taken directly from this arXiV paper, which goes on to prove some more cool stuff about linear networks.
When we train a neural network, we’re interested in adjusting the parameters so as to minimize some negative outcome, which we call the loss, \(l\). In linear regression, this loss is often the squared prediction error.
For our purposes, we won’t need to know what it is. We just need the following fact:
that is, there exists a function \(L\) that is just a function of \(W\), our product of matrices, and which is equal to the original loss \(l\), which is a function of all of the matrices.
A side note: you might think of both of these as being functions also of the data \(x\), but we won’t be referring to the data, so I’ve chosen to suppress that in the notation.
And before we take off, I’ll introduce one last piece of notation. We’ll need to refer to abbreviated versions of our matrix product \(W\), e.g. the product of the first 5 or last 3 matrices. Inspired by Python, we’ll denote the products truncated at \(i\) from the front and from the back as
respectively. If \(i\) is 1 or \(L\), the resulting matrices are taken to be the identity.
We now recall the Fréchet definition of the derivatve \(\nabla g\), of a matrix-to-scalar function \(g\):
see the first blog post in this series for details. Most importantly, the inner product \(\langle , \rangle\) is in the Frobenius norm, meaning it applies to matrices and can also be written
where \(\mathrm{tr}\) is the trace of the matrix. This is equivalent to multiplying the matrices entrywise and summing up the results, much like the more familiar inner product of vectors.
Let’s take our derivative definition and plug in our linear network loss function, computing the derivative with respect to the \(k\)th weight matrix:
where the second line follows by multiplying out the matrix product and applying our definitions.
Next, we swap over to the trace version of the inner product and apply the cyclic property which was so crucial in the the second blog post in this series. That is, underneath the trace, we are allowed to permute matrices, letting us write
and thus, pattern-matching to the definition of the derivative, we have that
That is, we take the derivative with respect to the whole product of matrices and then propagate it through every other matrix, transposed.
Closing
This proof may feel shockingly short and direct. Where are all the indices and sums? This is the advantage, which I hope this series has made clear, of using Fréchet derivatives on polynomial and linear functions that have scalar outputs but take in vectors or matrices.
For functions, like neural networks, where there is mixture of linear and nonlinear components, the Fréchet derivative can be used to handle the linear component, which is frequently the source of troubles, from shapes to indexing. With that part handled, attention can go where it is most needed, on the nonlinear piece.