Course - Theories of Deep Learning MT25


My notes for this course are a little different from my other [[University Notes]]U, since (at least now) it is assessed by mini-project at the end of the term; this means I’m trying to optimise more for understanding1 rather than exam grades. For this reason, some of the things I take notes on here might not actually be covered in the course explicitly (e.g. [[Notes - Theories of Deep Learning MT25, Vapnik-Chervonenkis dimension]]U).

Notes

Lectures

Reading List

Each lecture above is annotated with the articles and papers that were mentioned. Once a week, we also receive amount of

See:

Problem Sheets

Questions / To-Do List

  • Implement proof that “each MNIST digit class is contained on a locally less than 15 dimensional space”
  • Not known whether you can achieve the optimal $\epsilon^{-d/n}$ width using just one activation function, although it is possible with 2
  1. Although inevitably I will probably find myself reward-mispecificiationing into optimising for mini-project results. 




Related posts