26.12.2022 Views

TheoryofDeepLearning.2022

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

30 theory of deep learning

The proof of the Claim follows in straightforward fashion from

implementing the message passing protocol as an acyclic circuit.

Next we show how to compute ∇ 2 f (z) · v where v is a given fixed

vector. Let g(z) = 〈∇ f (z), v〉 be a function from R d → R. Then by

the Claim above, g(z) can be computed by a network of size O(V + E).

Now apply the Claim again on g(z), we obtain that ∇g(z) can also be

computed by a network of size O(V + E).

Note that by construction,

∇g(z) = ∇ 2 f (z) · v.

Hence we have computed the Hessian vector product in network size

time.

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!