Data Science Asked by donkey on August 25, 2021
I am looking to calculate the information contained in a neural network. I am also looking to calculate the maximum information contained by any neural network in a certain amount of bits. These two measures should be comparable (as in I can compare whether my current neural network has reached the max or is less than the max and by how much).
Information is relative, so I define it relative to the real a priori distribution of the data that the neural network is trying to estimate.
I have come across Von Neumann entropy which can be applied to a matrix, but because it is not additive I can’t apply it to a series of weight matrices (assuming the weight matrices encode all the information of a neural network).
I found three other papers Entropy-Constrained raining of Deep Neural Networks , Entropy and mutual information in models of deep neural networks and Deep Learning and the Information Bottleneck Principle. The second contains a link to this github repo, but this method requires the activation functions and weight matrices to be known which is not the case for finding the max entropy of any neural network in n bits.
How can I calculate the amount of information contain in/entropy of a neural network? And how can I calculate the same measure for any neural network in n bits?
To start, see Information Theory, Inference, and Learning Algorithms by David J.C. MacKay, starting with chapter 40 for information capacity of a single neuron (two bits per weight) through to at least chapter 42 for Hopfield Networks (fully connected feedback). The classic reference for information of a Hopfield Network is Information Capacity of the Hopfield Model by Abu-Mostafa & St. Jacques, but the textbook should have enough.
Answered by C8H10N4O2 on August 25, 2021
Get help from others!
Recent Answers
Recent Questions
© 2024 TransWikia.com. All rights reserved. Sites we Love: PCI Database, UKBizDB, Menu Kuliner, Sharing RPP