What are the best books to study Neural Networks from a purely mathematical perspective?

$\begingroup$

I am looking for a book that goes through the mathematical aspects of neural networks, from simple forward passage of multilayer perceptron in matrix form or differentiation of activation functions, to back propagation in CNN or RNN (to mention some of the topics).

Do you know any book that goes in depth into this theory? I've had a look at a couple (such as Pattern Recognition and Machine Learning by Bishop) but still have not found a rigorous one (with exercises would be a plus). Do you have any suggestions?

$\endgroup$

7 Answers

$\begingroup$

I'd recommend Deep Learning by Goodfellow, Bengio and Courville. I don't know if I'd call it "purely mathematical", but it covers a good amount of math background in the first few chapters. No exercises, though.

$\endgroup$ 1 $\begingroup$

For MLPs, there is a rigorous derivation in the optimization textbook by Edwin Chong and Zak. Although it is notation heavy as all things related to neural networks must be.

This book is for some reason freely available online. See page 219 of

I think there is essentially no good mathematical textbook on convolutional neural networks or RNN in existence. People essentially just base their intuition off of MLPs. But it is not hard to create a mathematically rigorous derivation of forward and backward propagation of CNN or RNN.

$\endgroup$ 1 $\begingroup$

Gilbert Strang (of MIT OCW Linear Algebra lectures and Introduction to Linear Algebra fame) has a new textbook on linear algebra for deep learning,Linear Algebra and Learning from Data.

It's got a decent course in linear algebra, some statistics & optimization, the calculus needed for stochastic gradient descent, and then applies them all to neural network models.

$\endgroup$ $\begingroup$

One of my favorite books on theoretical aspects of neural networks is Anthony and Bartlett's book: "Neural Network Learning Theoretical Foundations".

This book studies neural networks in the context of statistical learning theory. You will find loads of estimates of VC dimensions of sets of networks and all that fun stuff.

I should say that this book does not go into detail on CNNs and RNNs.

$\endgroup$ $\begingroup$

This field is in its nascent age. Not too many materials for "pure mathematical" lovers. Perhaps you would like to take a look at Stanford's STAT581 course (Theories of Deep Learning).

$\endgroup$ $\begingroup$

Not a book but maybe of some interest for a current perspective:

Backprop as Functor: A compositional perspective on supervised learning Brendan Fong David I. Spivak Remy Tuyeras (2018) gives a category theoretic structural framework based on the algorithm:

This is further discussed by David Spivak (2019) via:

$\endgroup$ $\begingroup$

I find this book usefulNeural Networks - A Systematic Introduction a book by Raul Rojas

also, this Perceptrons: an introduction to computational geometry by Minsky & Papert 1969 is useful from pure mathematical perspective

$\endgroup$

Your Answer

Sign up or log in

Sign up using Google Sign up using Facebook Sign up using Email and Password

Post as a guest

By clicking “Post Your Answer”, you agree to our terms of service, privacy policy and cookie policy

You Might Also Like