Essence Neural Networks


About

Project Details

Essence Neural Networks are a lightweight framework that distills the essential structure of the model into a smaller, interpretable model. The unsupervised Hebbian network learns a single-layer representation of input arrays and synaptic weights are strengthened toward the test array, based on co-activation of input pixels with neuronal firing

unSupervised Hebbian

How the network Specializes on 1 input

The network receives 3 inputs: {1, 0, -1}, {0, 1, 1}, {1, 1, 0} to train on and the neurons either fire in response to the test input or are strongly inhibited by the test input. Weights are initialized and the input pattern weight activations are summed and passed through an activation function.

Hebbian updates increase weights where active pixels align with the correct class and decrease the weights for activation in the wrong class.

Finally, weights are normalized and updated according to the hebbian rule, and the test outputs are calculated based on strength of activation to input

Initial input weights of 2 neurons before training on test input {1, 0, -1}. Neuron 1 fired strongly to the test input while the other neuron didn’t fire to the test input