Transferring Knowledge between Neural Network Architectures
|Director of thesis||Dr. Francois Fleuret|
|Co-director of thesis|
|Summary of thesis||
Deep Neural Nets are powerful tools for statistical modelling. However, they are plagued by several problems, including, poor performance on small datasets, high cost of training and having large model sizes. Knowledge transfer methods have recently emerged as a unified framework to solve these problems. This involves transferring representations from an previously trained “teacher” model to an untrained “student” model. For our research, we propose to build algorithmic tools that improve knowledge transfer between neural networks. We intend to obtain improvements by using architecture independent quantities, which are useful for knowledge transfer between neural nets of different architectures. In our initial work we apply one such quantity, the Jacobian, and find that it indeed improves knowledge transfer.
|Administrative delay for the defence||2021|